Search Results

Search found 19975 results on 799 pages for 'disk queue length'.

Page 504/799 | < Previous Page | 500 501 502 503 504 505 506 507 508 509 510 511  | Next Page >

  • Image archiving on network folders.

    - by Steve
    Our company uses Symantec Enterprise Vault to archive network folders and files, presumably to save on disk space. I can't see any other benefits. Our company is an architectural firm, and the problem our users face is locating particular images in network folders. Because all the files are archived, Windows Explorer is unable to generate image thumbnails. Each image needs to be individually restored from archive by double-click before it can be viewed. This is a big time-waster for our architects. Symantec say there is no workaround for this. Does anyone know of an alternative we could use for archiving images? Alternatively, some batch utility to create and maintain a thumbs.db file in each folder? Thanks.

    Read the article

  • Is it possible to run a "Windows Server 2008 R2 SP1 Hyper-V Virtual Machine" on Windows 7x64 Pro machine?

    - by tbone
    Specifically, I want to run this: http://www.microsoft.com/en-us/download/details.aspx?displaylang=en&id=27417 This download contains a three Windows Server 2008 R2 SP1 Hyper-V Virtual Machine set for evaluating and demonstrating Office 2010, SharePoint 2010 and Project Server 2010. System requirements Supported operating systems: Windows Server 2008 R2 Additionally you will need: •Windows Server 2008 R2 (SP1 recommended) with the Hyper-V role enabled. •Drive Formatting: NTFS •Processor: Intel VT or AMD-V capable •RAM: 8 GB or more recommended •Hard disk space required for install: 100 GB ------------------------- So, the above text seems to indicate that you need a physical Windows 2008 Server (r2) server running, but from the googling I've done on the subject I've yet to come across a discussion that definitively answers the question. Many posts I've read seem to indicate it might be possible to run it on Windows 7 using one of the following: VirtualBox, Windows Virtual PC, VMWare but I'm not entirely sure.

    Read the article

  • How to view dirty page count in Windows Server 2003

    - by Mark Wilkins
    Is there a way to view the number of dirty pages (cached file pages that need to still be written to disk) in Windows Server 2003? In Windows 7, for example, I can use Performance Monitor and use the "Dirty Pages" counter (one of the cache counters). This counter does not seem to be available in Server 2003. Also on Windows 7 (and other later systems), I can use Sysinternals RAMMap and effectively see the dirty pages on a file-by-file basis. Is there something similar for Server 2003?

    Read the article

  • Would upgrading memory from 4GB to 8GB on my laptop solve swapping issues?

    - by Tom
    I have a laptop with 4GB of memory with Windows 7 on it and I often experience with Eclipse that it is swapped out to disk. On the net they usually write 4GB of RAM is more than enough for average use and aside from Eclipse+Android Emulator I don't really use other extra apps, yet Eclipse is always swapped out if I haven't used it for a while (say, 1 day) and it is annoying it to wait for it to be resurrected from swap. My question is: would an upgrade to 8GB solve the issue of swapped out applications? With 8GB would windows 7 keep everything in memory? Or it wouldn't change anything and Eclipse would be swapped out regardless of the amount of memory, because Win 7 has a habit of kicking out every application from memory which hasn't be used for a while?

    Read the article

  • Can I use dis-similar HW for Win2008r2 DFS-R

    - by cwheeler33
    The setup: Windows 2008R2 Ent on two machines. The roles on each server will include File Servers and DC's. The machines come from two different vendors (Dell/HP) The Dell is an Athlon and the HP is an Intel. Both have roughly the same speed CPU and 8GB of RAM. They have different Raid controllers, and more or less the same amount of disk space (roughly 6TB.) Can the servers use different types of hardware? Is there any documentation about this? The last question I have is about the network. Can DFS-R be forced to use a differen subnet from the regular network?

    Read the article

  • Can't burn 8.1G iso onto 8.4GB DVD - "Media does not have enough free space"

    - by Max Williams
    I'm trying to burn a dvd on a mac with an external (firewire-connected) dvd drive. I'm checking the size of the iso thus: DVD-4:dvd_files macbook$ ls -l /tmp/hybrid.iso -rw-r--r-- 1 macbook wheel 8700884992 Aug 22 10:57 /tmp/hybrid.iso DVD-4:dvd_files macbook$ ls -lh /tmp/hybrid.iso -rw-r--r-- 1 macbook wheel 8.1G Aug 22 10:57 /tmp/hybrid.iso The "human-readable" size is 8.1 Gig but when i try to burn, onto an 8.4G dual-layer dvd, it says "Media does not have enough free space" The definition of a "Gigabyte" according to Wikipedia is 1 billion bytes, so the iso size should actually be 8.7 Gig according to this definition, in which case the disk definitely isn't big enough, and it's just that the -h option to ls is misleading. Is the discrepancy just due to the ls command using a different definition of "G" (eg 1024 Meg aka 1.07 Gig? This comes out as 8.103 which fits what ls is displaying)

    Read the article

  • Linux cannot see Windows 7 partitions on install

    - by Nash0
    I've been trying to install Linux as a dual boot with Windows 7 on my Dell latitude e6510. It is currently running Windows 7 and I have used the MS disk tools to shrink the Win 7 NTFS partition to make room for Linux. The issue I'm having is that when I run Linux installers by boot from CD they see the entire hard drive as unallocated space. I have tried Ubuntu 10.10, Kbuntu 10.10 and Fedora 14 and they all have the same problem. I have also tried the Ubuntu "install in Windows" option and could not get it to work. EDIT: Booting Gparted 0.8.0 from a usb drive did not work. It reported the entire drive as unpartitioned.

    Read the article

  • Linux mdadm software RAID 6 - does it support bit corruption recovery?

    - by user101203
    Wikipedia says "RAID 2 is the only standard RAID level, other than some implementations of RAID 6, which can automatically recover accurate data from single-bit corruption in data." Does anyone know if the RAID 6 mdadm implementation in Linux is one such implementation that can automatically detect and recover from single-bit data corruption. This pertains to CentOS / Red Hat 6 if those are different from other versions. I tried searching online but didn't have much luck. With SATA error rates being 1 in 1E14 bits, and a 2TB SATA disk containing 1.6E13 bits, this is especially relevant to preventing data corruption. Thanks!

    Read the article

  • suggest AMD machine for software developement [closed]

    - by superb1
    I need a AMD machine (Processor, Motherboard, Ram, HDD ) for software developement .. I will mainly use microsoft visual studio 2010. Operating system would be windows 7 . Suggest decent machine but i dont want to overspend....i guess the rating system of ecommerce websites may help you to suggest !? I want you to suggest from this list : http://www.flipkart.com/computers/components/processor-20246?ref=bae2643a-2cc8-4710-942f-79e7fbc1b416 plz note basic criteria for vs 2010 is 1.6GHz or faster processor 1 GB (32 Bit) Ram 5400 RPM hard disk DirectX 9 capable video card running at 1024 x 768 or higher-resolution

    Read the article

  • Replicate a big, dense Windows volume over a WAN -- too big for DFS-R

    - by Jesse
    I've got a server with a LOT of small files -- many millions files, and over 1.5 TB of data. I need a decent backup strategy. Any filesystem-based backup takes too long -- just enumerating which files need to be copied takes a day. Acronis can do a disk image in 24 hours, but fails when it tries to do a differential backup the next day. DFS-R won't replicate a volume with this many files. I'm starting to look at Double Take, which seems to be able to do continuous replication. Are there other solutions that can do continuous replication at a block or sector level -- not file-by-file over a WAN?

    Read the article

  • "Device Not Ready" when attempting to connecting an old (DOS) hard drive to windows 7

    - by Christopher Vigliotti
    I have an old Western Digital Caviar 2540 hard drive that I am attempting to connect to a Windows 7 machine by using a Bytecc BT-300 USB 2.0 to IDE/SATA Adapter. I'm connecting things in the right order (power, then sata, then usb) and the jumper settings on the drive are at present set to slave, but am still having an issue. I am recieving a message that the drive is not initialized. When I go into Computer Management I can see the drive (as "Disk 5, Unknown, Not Initialized"), but when I attempt to initialize it or right-click and select "Offline" It tells me that "the device is not ready". Is there something that I can to to get around this, connect the drive and copy the data that I need off of this drive? Is there a third party tool available that I can use?

    Read the article

  • Windows doesn't recognise my USB key anymore (it used to work)

    - by dominicbri7
    I use my friend's USB flash drive (Corsair flash voyager 16gb) to transfer files from my laptop to my desktop computer. However, since a couple of days my laptop stopped recognizing the USB key.. while there is still no problems with all other computers. I use Windows 7 64 bits if that can help. I tried uninstalling the driver, rebooting and all those kind of tricks, but it won't work. When I connect it and open the "My computer" window, I see "Removable Disk (G:)" for a moment, then it disappears... then it reappears again and it keeps doing that periodically. I can't even right click then hit "Properties" because it disappears. As I recall, it DOES work on every other computers, I think it has to do with the driver but what can I do?

    Read the article

  • Encrypt backups with GPG to multiple tapes

    - by Dan
    Currently, I use tar to write my backups (ntbackup files) to a tape drive fed by an autoloader. Ex: tar -F /root/advancetape -cvf /dev/st0 *.bkf (/root/advancetape just has the logic to advance to the next tape if there is one available or notify to swap the tapes out) I was recently handed the requirement to encrypt our tape backups. I can easily encrypt the data with no problems using GPG. The problem I'm having is how do I write this to multiple tapes with the same logic that tar uses to advance the tapes once the current one is filled? I cannot write the encrypted file to disk first (2+TB). As far as I can tell, tar will not accept binary input from stdin (it's looking for file names). Any ideas? :(

    Read the article

  • List of recent motherboards with BIOS / without UEFI [on hold]

    - by jmn
    I am building a new desktop PC and I want to have full disk encryption on it. TrueCrypt doesn't support UEFI as of now. Are there still recent motherboards out there without UEFI ? I didn't find any list and I am afraid that I will have to study each potential candidate's technical sheet before purchase. I want to buy 2 or 3 of the same model to be future proof. Newegg links will not help, I don't live in the USA ... this means that this post is a legitimate target for PRISM ;-) Thanks for your help.

    Read the article

  • people_dl_import shows millions of records

    - by amit lohogaonkar
    We have a situation now on prod in sharepoint 2007 based intranet platform and it shows thousands of records under people_dl_import category with format spsimport://?$$dl$$/domain1/domain2/domain3/ Also import was not stopping and added millions of records in database and was on verge of disk full. On other servers like dev we have very less data in this category and format is also like spsimport://doaminname?$$dl$$?... which is good and has only 6000 rows and in prod its 2 millions Crawled under people_dl_import category. I need to know the cause of this garbage data and how to fix it. I tried resetting content source and I will do full import in this weekend to see if this garbage data gets cleared. Any idea on cause for thiss issue?

    Read the article

  • More than one XP Mode virtual machines okay under Win 7?

    - by kousen
    I really like the Windows XP Mode virtual machine that comes with Windows 7 (once you download the integration components). I teach technical training classes, and having a machine configured for the classroom is very helpful. My question is, can I configure more than one? The documentation suggests you can make multiple virtual machines, but I can't tell if I can have multiple "Windows XP Mode" VMs. It can't be as simple as configuring one, then just making a copy of the primary disk file, could it? I only plan to run one at a time, but I'd like to configure multiple different VMs for different classes, all running XP if possible. Thanks for any help!

    Read the article

  • Why Windows Media Center (Vista 32-bit) unable to record TV Series?

    - by Alex R
    I can record individual shows just fine, but trying to setup the recording of a Series throws an error (something like, "unable to schedule, try restarting Media Center", pops up in the bottom right of the screen). Trying to access the "Other Showings" menu causes Media Center to crash ("check for a solution online" dialog comes up). The machine is a fresh reinstall of 32-bit Windows Vista from the HP recovery CDs, plus all Microsoft updates and service packs. Hardware: HP s3220n, AMD X2 5000+, 2GB RAM, 300GB free space on disk. Media Center came bundled. The "Media Center update for Vista" failed to install (it says, not appropriate for my computer). Any ideas? Thanks

    Read the article

  • WindowsXP+7 on C:/D: :moving the System Partition

    - by user938921
    I had installed Windows 7 on a separate 20gig partition, and I'm absolutely loving it! Plus I can now dual-boot, with my original WinXP residing on the C: drive. But I'm running out of disk space on D:, and I was able to shrink C: and expand D:. But now I would like to make D: not just a Boot Partition, but an Active System Partition, without losing my ability to boot into Windows 7 (since it was created on a separate D: partition, not the current Active System C: Partition). Any advice?

    Read the article

  • Best practice for administering a (hadoop) cluster

    - by Alex
    Dear all, I've recently been playing with Hadoop. I have a six node cluster up and running - with HDFS, and having run a number of MapRed jobs. So far, so good. However I'm now looking to do this more systematically and with a larger number of nodes. Our base system is Ubuntu and the current setup has been administered using apt (to install the correct java runtime) and ssh/scp (to propagate out the various conf files). This is clearly not scalable over time. Does anyone have any experience of good systems for administering (possibly slightly heterogenous: different disk sizes, different numbers of cpus on each node) hadoop clusters automagically? I would consider diskless boot - but imagine that with a large cluster, getting the cluster up and running might be bottle-necked on the machine serving the OS. Or some form of distributed debian apt to keep the machines native environment synchronised? And how do people successfully manage the conf files over a number of (potentially heterogenous) machines? Thanks very much in advance, Alex

    Read the article

  • Can't boot Windows after installing Linux

    - by user4035
    I have a partition /dev/sdb1, where my old Windows XP resides. All the files are there intact and I can see them, mounting the disk from Linux. Linux is on /dev/sdb2. But when I choose Windows in LILO prompt, it doesn't load. I have the following lilo.conf: boot = /dev/sdb # Linux bootable partition config begins image = /boot/vmlinuz root = /dev/sdb2 label = Linux read-only # Partitions should be mounted read-only for checking # Linux bootable partition config ends # Windows bootable partition config begins other = /dev/sdb1 label = Windows table = /dev/sdb # Windows bootable partition config ends What can be wrong?

    Read the article

  • Booting from a different partition.

    - by Legolas
    I have a question. I run Windows7 on C:\ drive and it is the only operating system. I want to install Windows8 developer installation (iso image), and I have it saved in a new partition I created, which has only this.G:\ I want to boot from this partition G:\ so that I can install Windows 8, but in my boot options I get only Internal Hard Disk, CD ROM, and USB Drive. I don't have a USB drive. How can I boot from G:\ with that iso image ?

    Read the article

  • Partition and mount my secondary hard drive on CentOS 5.5 64bit?

    - by Andrew Fashion
    I am trying to prepare my second hard drive for user image uploads. Here is the current layout: # sudo parted /dev/sda print Model: ATA WDC WD2500KS-00M (scsi) Disk /dev/sda: 250GB Sector size (logical/physical): 512B/512B Partition Table: msdos Number Start End Size Type File system Flags 1 32.3kB 107MB 107MB primary ext3 boot 2 107MB 8595MB 8488MB primary linux-swap 3 8595MB 10.7GB 2147MB primary ext3 4 10.7GB 250GB 239GB extended 5 10.7GB 250GB 239GB logical ext3 Information: Don't forget to update /etc/fstab, if necessary. I am assuming #4 is my secondary drive? How do I partition and mount it so I can begin using it? And how do I add to fstab? I understand if it's to many questions in one, just help me with whatever you can I guess :) Thank you for any help!

    Read the article

  • Recover partition table after DD command

    - by Shreedhar
    I executed the following command from a Ubuntu live cd terminal (dont ask why). dd if=/dev/zero of=/dev/sdb2 bs=512 count=1 Where sdb2 is a NTFS partition (third partition) on a disk. Suffice to say it is now messed up. When I boot into windows 7, it does show me E drive but when I click on it it asks me to format it. I am not ever sure what I did, did I mess up partition table or only the MFT? Is there any way to get the data back PLEASE HELP! this is very important :(

    Read the article

  • Poor TCP loopback throughput on Windows

    - by Yodan Tauber
    I measured the throughput of a locally bound TCP socket connection on my computer (Intel Q9550, 64 GB RAM, Windows XP 64 bit) using iperf. I got dissatisfying results (around 1.6 Gbit/s) each time, no matter how I tweaked the TCP settings (buffer length, window size, max segment size, no delay). I got similar results when I tried netperf. Now, I understand (from sources like these) that the average throughput of a loopback connection should be around 5 Gbit/s. What could be the reasons for such poor performance?

    Read the article

  • Visual Studio Development on Virtual Box, Boot Camp, or VMWare Fusion

    - by Eli
    I currently have a Mac, 2ghz and 2 gigs of ram, running OS X Leopard and Virtual Box with a Windows 7 Pro 32bit virtual machine. Performance on the virtual machine is fine for minor tasks but is very clunky while trying to multi-task or develop in Visual Studio 2008. What would be my best option for being able to use Visual Studio, keeping cost and time in mind? 1) Upgrade ram to 4 gigs ($100). Will this really improve my performance enough to use Visual Studio in a Windows 7 vm? Or am I just wasting time/money? 2) Reinstall/restore Windows 7 disk image as a Boot Camp partition. I assume this should improve my performance, yes? 3) Purchase VMWare fusion instead of VirtualBox. Does Fusion require less resources to run? I am open to any suggestions. Thanks in advance

    Read the article

< Previous Page | 500 501 502 503 504 505 506 507 508 509 510 511  | Next Page >