Search Results

Search found 4263 results on 171 pages for 'so aware'.

Page 70/171 | < Previous Page | 66 67 68 69 70 71 72 73 74 75 76 77  | Next Page >

  • 'Unlimited' free trial of Windows Server 2008 by deleting and reinstalling VM? [closed]

    - by MrVimes
    I am using Virtual Machine software (VirtualBox) to learn Windows Server 2008 R2 Network Infrastructure (70-642). Trouble is - I'm learning at an extremely slow pace and so the trial periods of my virtual machines are close to running out. If I delete the VMs then install WS2008R2 from scratch on new VMs is that violating the acceptable use policy of Microsoft? I am aware that I can extend the trial, but it seems I can only do that by 10 days at a time. Also I think having to re-install from scratch is a good way to reinforce the knowledge.

    Read the article

  • All HTTPS, or is it OK to accept HTTP and redirect (secure vs. user friendly)

    - by tharrison
    Our site currently redirects requests sent to http://example.com to https://example.com -- everything beyond this is served over SSL. For now, the redirect is done with an Apache rewrite rule. Our site is dealing with money, however, so security is pretty important. Does allowing HTTP in this way pose any greater security risk than just not opening or listening on port 80? Ideally, it's a little more user-friendly to redirect. (I am aware that SSL is only one of a large set of security considerations, so please make the generous assumption that we have done at least a "very good" job of covering various security bases.)

    Read the article

  • Is there a word processor similar to MS Word which saves files as readable txt files?

    - by zenbomb
    I'm writing a paper together with my supervisor and would like to have a more sophisticated version control than *_291112_NEW_NEW_revised1.doc files. My supervisor is a non-computer person will never ever use LaTeX or git and loves MS Word. I'm therefore looking for an alternative to Word (I need commenting on text passages!) which stores the files as clean text (Markup for formating is fine), so I'm able to put them under version control on my side. I'm aware that git can also handle binary files, but I'd prefer the cleaner way of looking at the contents directly. If there's a way to automatically extract the text from word files, I'm fine with that too for now.

    Read the article

  • Setting CPU cores off-limits to all threads not specified (preferably in Windows 7)

    - by Shinrai
    I have a really specific machine configuration in the works that would really be helped out if there were any way to do this...basically what I'm looking for is the opposite of setting CPU Affinity for a process. I want to be able to tell Windows "No applications except [x] are allowed on [these cores]." Is there any mechanism whatsoever for doing this? (Yes, I am aware of some of the potential issues this could cause and I normally would never fool with processor affinities, since the OS usually does a damned good job itself, but this is a pretty odd situation involving some software that is very CPU-bound constantly having to wait on interrupts and DPCs and things from other threads.)

    Read the article

  • Redirecting output to email

    - by Alpha
    I am aware that in Windows you can use the following snippet to redirect the output of a command line tool to the clipboard: mytool | clip Also, it can be done with files: clip < myFile.txt I also know Windows has several other devices you can use as output / input for these redirections (probably the printer is the most famous of them). However, is there something to send the text to the default email program? I would love to find that there is something like: mytool | email. Is there something alike?

    Read the article

  • Linux virtual disk stripping or multi-path samba share?

    - by wachpwnski
    I am trying to build a file storage box for media. It needs to span two or more directories or partitions as one share. There are a few solutions but reasons why I want to avoid them, among these are: Using LVM2 for stripping. I don't really have the resources to back up everything on the volumes incase one HDD goes south. I would end up loosing everything. Maybe there is a better option for this to prevent data loss with hot swappable drives or some kind of raid. Using symbolic links in the share. This will get tedious every time a new sub-directory is added. Is there some kind of software raid I can use to merge two directories virtually? I am aware of the issue where /dev/hda1/media/file.1 and /dev/hdb1/media/file.1 both exist. But I'm sure there are some creative solutions for this.

    Read the article

  • Moving cpanel backup of magento site to VPS

    - by user2564024
    I was having my site in shared hosting, I took the entire backup, its structure is like addons homedir mysql resellerpackages suspendinfo bandwidth homedir_paths mysql.sql sds userconfig counters httpfiles mysql-timestamps sds2 userdata cp locale nobodyfiles shadow va cron logaholic pds shell vad digestshadow logs proftpdpasswd ssl version dnszones meta psql sslcerts vf domainkeys mm quota ssldomain fp mma resellerconfig sslkeys has_sslstorage mms resellerfeatures suspended Now I have subscribed to vps, I have copied the files inside homedir/public_html to var/www/html of my new hosting, but am seeing the following error when I view it browser, There has been an error processing your request Exception printing is disabled by default for security reasons. Error log record number: 259343920016 I have just created database with name magenhto inside mysql. Previously I had cpanel and used one click installer. Hence am not aware of how to use that data inside mysql to this new system and are there any more changes.

    Read the article

  • Replacement for NewSID when working with Windows 7?

    - by sbussinger
    Has anyone got a lead on a replacement for the old SysInternals NewSID utility that will work for Windows 7? I found out the hard way that NewSID will totally hose a Win7 setup (BSOD on reboot). Apparently the same problem occurs for Server 2008 R2. They've officially announced that NewSID will be retired in November, possibly because of this issue. I'm aware of SYSPREP, but it's not a clean replacement in my case so I was hoping that there was another utility similar to NewSID that worked with Win7. Reworking our system to use SYSPREP will be ugly and slower as well. Any thoughts on another alternative? Thanks!

    Read the article

  • Ubuntu stops auto-mounting flash drive

    - by Brian
    It seems that after being up for a few days, my Ubuntu system refuses to auto-mount hot-plugged USB disks (i.e. flash drives). The output from dmesg shows that the kernel recognizes the device correctly. The only solution I'm aware of at the moment is to reboot (logging out may work as well, but the impact is the same since I have a bunch of stuff open and it takes a few minutes to get everything situated after startup/login). I thought gvfs-fuse-daemon was the thing responsible for managing filesystems in userspace, but killing and restarting that doesn't help. Any other ideas?

    Read the article

  • How do I backup (from Vista Home Premium) to a FAT32 HDD connected to a wireless router with no user account or password set?

    - by Scubadooper
    I have a Seagate expansion 2TB HDD which I've managed to get working on my wireless router (thanks to fat32format) however I haven't managed to set my backup to work with it. Vist Home Premium requires that I input a username and password in the backup utility. As far as I'm aware that type of access control isn't set on the HDD: Can I set up the access control on the HDD? If so how Or, can I set the backup to work without it? If so how Thanks for your help, I haven't been able to find the answer anywhere on the net so far

    Read the article

  • Load balanced proxies to avoid an API request limit

    - by ClickClickClick
    There is a certain API out there which limits the number of requests per day per IP. My plan is to create a bunch of EC2 instances with elastic IPs to sidestep the limitation. I'm familiar with EC2 and am just interested in the configuration of the proxies and a software load balancer. I think I want to run a simple TCP Proxy on each instance and a software load balancer on the machine I will be requesting from. Something that allows the following to return a response from a different IP (round robin, availability, doesn't really matter..) eg. curl http://www.bbc.co.uk -x http://myproxyloadbalancer:port Could anyone recommend a combination of software or even a link to an article that details a pleasing way to pull it off? (My client won't be curl but is proxy aware.. I'll be making the requests from a Ruby script..)

    Read the article

  • Iptables to lock down compromised server to a single ip

    - by ollybee
    I have a Linux server which is compromised, I can see nasty looking perl scripts executing with root privileges. I want to get some data off it before I wipe it. How can I block all inbound and outbound traffic except for my ip? It's a Centos server I assume i can do this with iptables? I'm aware a the server is rooted there is a possibility that attackers could have made changes on the server that would prevent this from working. Ill be testing to make sure and only have the server online for a couple of hours before it is nuked.

    Read the article

  • Is there a limit on CNAME records for subdomains?

    - by Vilx-
    I've bought a domain, say mydomain.com. I'd like to set up google apps to work with it. As such, I'd like to create subdomains like mail.mydomain.com, docs.mydomain.com, etc. To integrate it with google, I must add to all of them CNAME records to ghs.google.com. However my domain registrar says that they can create only one CNAME in my domain, for which they chose mail.mydomain.com. The rest of them, they say, I must create as A type records. Are they being lazy, or is there some limitation here that I'm not aware of?

    Read the article

  • How to batch rename files based on file header/metadata in Windows?

    - by Infraded
    I have a directory full of randomly named files of different types, all with no file extensions. Most are images, with some videos, and some plaintext. I've used one of the Windows versions of file to confirm the files can all be identified by their headers/metadata, but would like to automate the naming as there are roughly 2400 files. I don't care so much about the filename as much as just having the appropriate extension for it's type. Is anyone aware of a program or script that can do this?

    Read the article

  • Open table cache in MySQL

    - by vvanscherpenseel
    I have my open table cache set to 1800 and I have a total of 1112 tables. MySQL Tuning Primer reports that 100% of my table cache is used yet my table cache hit rate is 5%. I understand that this happens due to concurrent connections all opening tables. I think I should raise the cache limit. I understand that the cache size is limited by the file descriptor limit of my operating system, but are there any other practical limitations I should be aware of? Searching Google or this very website yields mostly posts explaining the connection-factor or come up with indecisive answers. My question: can I safely increase the open table cache limit? Is there a maximum?

    Read the article

  • Kernel compiling with -j2+ parameter ends prematurely with no error message or output bzImage

    - by Minix
    I've noticed quite a while ago that compiling a kernel with the parameter -j set to 1 or more doesn't produce a bzImage. Instead, it ends prematurely without any advice. I have reproduced the same behavior in both my netbook and home server. As far as I'm aware, the point where the compilation stops is random - Compiling twice with the same parameters will probably stop at different files. However, when I run make with no -j* parameter the compilation ends just fine and outputs a working bzImage. Both machines run Intel Atom (N270 on the netbook and 330 on the server) and I've compiled for these processors. If I recall correctly, I've tried compiling both with Atom and with generic x86_64 options. The kernel version I'm building is 2.6.34.1 I've always compiled normally with those options in my Core2Duo and Pentium Dual Core machines. Has anyone experienced this issue? Any ideas why does this happens? Is there a fix or workaround?

    Read the article

  • What happened to NewsGator and FeedDemon?

    - by user1413
    It's been a while since I used NewsGator and FeedDemon. I tried to log on last week and today and both seem broken. NewsGator no longer has a login -- it has been replaced by DataWire on the www.newsgator.com home page and my login no longer works. FeedDemon is completely broken -- it is no longer sucking in RSS feeds. What happened? And what do I do now? Do I need to get a completely new RSS feeder? Or is there something else that NewsGator is doing that I'm not aware of?

    Read the article

  • How to monitor RAM usage for Hyper-V VMs ?

    - by Mac
    A bit of context first : on Windows 2008 Standard x64 with 8Gb RAM, I have 5 VMs running which should take up 1664Mb RAM (3*256Mb+384Mb+512Mb). There is nothing else running on this server except the basic OS components (this not a Core installation). I know that each VM will use more RAM on the host than what has been configured in Hyper-V. But when I run the task manager, it says 6.7Gb used ! If I sum up the RAM used by each process in the task manager (showing all users processes), I get to something around 1Gb... So : how can I check how much RAM each VM is really using on the host (it does not seem to be available via task manager) ? Note that I am aware of the fact that my problem could be unrelated to VM RAM usage, but I would still very much like to know how to do this.

    Read the article

  • Latexmk on OS X

    - by Barum Rho
    Does anyone use latexmk on OS X with either Preview or another PDF, dvi, or PS reader? I've recently switched from Linux and installed MacTex to use Latex. I am used to using Vim to edit the source file and having latex -pvc running so that it will update the preview on xdvi upon save. I am aware that I can run xdvi using XQuartz, but I have it setup to run fullscreen, so it is not very useful when I don't want to go into X11 mode altogether. The problem with using Preview is that it will only update from the PDF file when Preview receives focus. I could not figure out if there is a command to trigger Preview to reload the PDF file.

    Read the article

  • Apache routing vhosts to /var/www

    - by FHannes
    One user at my site has reported that he reaches the content at /var/www when browsing to any of the vhosts at my server. As far as I’m aware, my Apache server does not contain a document root that references this folder. On top of that, this user seems to be the only one experiencing the issue. According to his ISP, the issue isn’t caused by them, yet, on his mobile connection, he can access the site. When browsing to my server’s IP, he also receives the correct content from the default vhost. What could be the possible causes of this issue and how can I get it to stop? I’ve explored pretty much every option I could think of.

    Read the article

  • Free firewall for Windows Server 2003

    - by n00b_Admin
    We have a Windows Server 2003 R2 Machine here at work that I need to ensure is accessible only from specific set of IP Addresses. The Server is on a LAN and I am aware that one of the possible solutions maybe limit access to the Server's VLAN. However that is not feasible at the moment. I've been searching for a firewall that could do this, Windows' Firewall is pretty limited and I don't see an option in it to limit incoming data based on Source IP addresses. Most free firewalls like Comodo and ZoneAlarm don't work on Windows Server. Any ideas on how to accomplish this would be of great help. Thanks in advance.

    Read the article

  • Is there a way to create a copy-on-write copy of a directory?

    - by BCS
    I'm thinking of a situation where I would have something that creates a copy of a directory, tweaks a few files, and then does some processing on the result. This wold be done fairly often, maybe a few dozen times a day. (The exact use case is testing patch submissions; dupe the code, patch it, build/test/report/etc.) What I'm looking for could be done by creating a new directory structure and populating it with hard links from the origonal. However this only works if all the tools you use delete and recreate files rather than edit them in place. Is there a way to have the file system do copy-on-write for a file? Note: I'm aware that many FSs use COW at a block level (all updates are done via writes to new blocks) but this is not what I want.

    Read the article

  • Are there web search engines with exact match or regex capabilities (for related terms)?

    - by naxa
    Every once in a while I come upon a situation when google's way of searching results in too broad results even if I enclose my search terms in quotation marks. For example now I've tried to find pages that contain both "py.path" and "path.py" without much success. I'm currently aware of engines|sites like Google Code Search for searching actual code and (apparently) Stackoverflow for searching QAs, symbolhound that lets me find symbols, and also wikipedia is often a good place to find lists of symbols. But none of these seems to perform very good in matching exactly on search term pairs (or tuples) and use a broad-enough segment of the web. Is there a website that is good with exactly finding search term pairs? (Why not?)

    Read the article

  • Cloud storage services offering one-time download links? [closed]

    - by TARehman
    Is anyone aware of consumer-targeted cloud storage services that allow users to generate a one-time download link for hosted files? Case in point: I have an encrypted container with some documents I need to send to a vendor. I would prefer to give them a one-time download link, so that I know when they have accessed the file, and then inform them of the passphrase by phone. I have heard that MediaFire offers 1-time links, but that they are buried in tons of advertising. At the moment, I'm not sure that I consider MediaFire fully legitimate; I'm more interested in solutions with Google Drive, Box.net, DropBox, etc.

    Read the article

  • What's the best way to copy deduplicated files onto a new Server 2012 drive?

    - by Screndib
    We have a deduplicated volume on a Windows Server 2012 machine that is approaching it's limits. It is a 1.3TB drive with ~10TB of duplicated data. We want to copy all of this data onto a larger 4TB drive. What is the best way to perform this copy such that we only copy the 1.3TB of deduplicated data instead of unpacking the entire 10TB and repacking it on the other end? edit: I attempted a standard explorer file copy and a Copy-Item but neither appeared to be dedup-aware. I didn't run either to completion however so I can't say this is the case for sure.

    Read the article

< Previous Page | 66 67 68 69 70 71 72 73 74 75 76 77  | Next Page >