Search Results

Search found 8937 results on 358 pages for 'disk defragmenting'.

Page 258/358 | < Previous Page | 254 255 256 257 258 259 260 261 262 263 264 265  | Next Page >

  • Replicate a big, dense Windows volume over a WAN -- too big for DFS-R

    - by Jesse
    I've got a server with a LOT of small files -- many millions files, and over 1.5 TB of data. I need a decent backup strategy. Any filesystem-based backup takes too long -- just enumerating which files need to be copied takes a day. Acronis can do a disk image in 24 hours, but fails when it tries to do a differential backup the next day. DFS-R won't replicate a volume with this many files. I'm starting to look at Double Take, which seems to be able to do continuous replication. Are there other solutions that can do continuous replication at a block or sector level -- not file-by-file over a WAN?

    Read the article

  • Disable writing RAID degraded mode

    - by jolivier
    I have a RAID5 with 5 disks on my machine and suspect the motherboard chipset to fail at some points and make my raid going in degraded mode. Last time it happened I noticed it on the failure of the 2nd drive connected to the same chipset and lost a lot of data. So I would like to prevent this, and especially I would like to have mdadm disable writes on the raid if one of the disk fails. So that in between I get notified, I recover and can use my system again. Sadly I could not find it in man mdadm so I was wondering if this is possible via a tool or hidden option since for me it looks like a standard feature of a RAID system. If this is not possible I would also be happy with a solution to stop the raid if degraded.

    Read the article

  • Cannot delete folder - Content seems to be nested recursively

    - by RikuXan
    I cannot delete a folder located on my hard disk by any means. I don't quite know how it was created, all I know is, that it is a pretty deep structure of folders (too deep to delete it at once, since Windows restriction path name too long), but the problem in the end is, that I can't "pull out" the inner folders, because they don't seem to be folders anymore (Context menu lacks things like "Properties", "Cut", "Copy", "Delete" etc.) Here a picture of how a right click looks like on one of these "folders": As you can see, the current folder is in very deep, but that is not the problem, rather the one I left-clicked on. Has anyone any advice on how to get rid of these? I tried a chkdsk, said no errors. I also tried deleting those folder via a VMWare Ubuntu, to no success. I also tried a batch file from a volunteer at MS boards, that should automatically de-nest such folders, but I guess mine is a special case, since the tool only created more such folders.

    Read the article

  • Windows doesn't recognise my USB key anymore (it used to work)

    - by dominicbri7
    I use my friend's USB flash drive (Corsair flash voyager 16gb) to transfer files from my laptop to my desktop computer. However, since a couple of days my laptop stopped recognizing the USB key.. while there is still no problems with all other computers. I use Windows 7 64 bits if that can help. I tried uninstalling the driver, rebooting and all those kind of tricks, but it won't work. When I connect it and open the "My computer" window, I see "Removable Disk (G:)" for a moment, then it disappears... then it reappears again and it keeps doing that periodically. I can't even right click then hit "Properties" because it disappears. As I recall, it DOES work on every other computers, I think it has to do with the driver but what can I do?

    Read the article

  • Encrypt backups with GPG to multiple tapes

    - by Dan
    Currently, I use tar to write my backups (ntbackup files) to a tape drive fed by an autoloader. Ex: tar -F /root/advancetape -cvf /dev/st0 *.bkf (/root/advancetape just has the logic to advance to the next tape if there is one available or notify to swap the tapes out) I was recently handed the requirement to encrypt our tape backups. I can easily encrypt the data with no problems using GPG. The problem I'm having is how do I write this to multiple tapes with the same logic that tar uses to advance the tapes once the current one is filled? I cannot write the encrypted file to disk first (2+TB). As far as I can tell, tar will not accept binary input from stdin (it's looking for file names). Any ideas? :(

    Read the article

  • "Device Not Ready" when attempting to connecting an old (DOS) hard drive to windows 7

    - by Christopher Vigliotti
    I have an old Western Digital Caviar 2540 hard drive that I am attempting to connect to a Windows 7 machine by using a Bytecc BT-300 USB 2.0 to IDE/SATA Adapter. I'm connecting things in the right order (power, then sata, then usb) and the jumper settings on the drive are at present set to slave, but am still having an issue. I am recieving a message that the drive is not initialized. When I go into Computer Management I can see the drive (as "Disk 5, Unknown, Not Initialized"), but when I attempt to initialize it or right-click and select "Offline" It tells me that "the device is not ready". Is there something that I can to to get around this, connect the drive and copy the data that I need off of this drive? Is there a third party tool available that I can use?

    Read the article

  • Linux live cd with Broadcom Wi-fi support

    - by paul simmons
    I am looking for a live distro that has out of the box Broadcom wireless support. I am pretty happy with my Ubuntu installation and as long as I have an ethernet connection first time installed, I can install Broadcom drivers over internet. But being a little paranoid, I make my secure operations (banking etc.) with a live cd and zero hard disk access, so nothing is recorded. So far I plug ethernet to do such things with the live cd, but it would be nice if I can do same thing with wireless.

    Read the article

  • Encrypt two drives with Truecrypt with password before boot

    - by Deshroom
    i'm using laptop and PC with single HDD with full disk encryption. I know how works truecrypt on single drive because i use it everyday. My second laptop has 2 HHDs. My question is how to encrypt first 128GB SSD and second 1TB HDD in same way. I have multiple applications installed on second drive so i want to have it accessible during boot in ex. Steam in installed on second drive and it starts with windows. How to do it? can i encrypt two drives in truecrypt and unlock it via password before boot? My main reason is i want to RMA laptop without removing disks or data - my data need to be encrypted. Thank you.

    Read the article

  • XP Laptop Restarting

    - by Liath
    I have a Dell 1525 which I reinstalled xp on recently due to a hard disk problem. Apparently I've missed something, it's taken to restarting for no apparent reason (I believe it's too frequent for windows updates which was my first thought). Last night it was particularly bad, it stopped 5 times in 2 hours while playing an online game. I can't actually think of a time it's happened when it's not running something in full screen mode. I can't see anything in the event log apart from usual start up details. How can I find and fix this issue?

    Read the article

  • Linux cannot see Windows 7 partitions on install

    - by Nash0
    I've been trying to install Linux as a dual boot with Windows 7 on my Dell latitude e6510. It is currently running Windows 7 and I have used the MS disk tools to shrink the Win 7 NTFS partition to make room for Linux. The issue I'm having is that when I run Linux installers by boot from CD they see the entire hard drive as unallocated space. I have tried Ubuntu 10.10, Kbuntu 10.10 and Fedora 14 and they all have the same problem. I have also tried the Ubuntu "install in Windows" option and could not get it to work. EDIT: Booting Gparted 0.8.0 from a usb drive did not work. It reported the entire drive as unpartitioned.

    Read the article

  • php-cgi got SIGKILL on burst.net VPS?

    - by Shawn
    I got an VPS at burst.net, very cheap one, but this doesn't matter. The strange behavior of it is, the php-cgi process, which started using lighthttpd's spawn-cgi, dies every a few minutes. however, other processes are fine and great, even include a java process, and I'm sure there is no "out of memory" issue, so it's not killed by OOM killer. I used strace to trace the process, and found out it was killed by SIGKILL, hence no single log was left on the disk, just dies suddenly. Is there anyway I can find out what process/thing sent the SIGKILL to the poor php process? Filed a ticket with the vendor, but they said they won't care. strace -p 7176 Process 7176 attached - interrupt to quit wait4(-1, <unfinished ...> +++ killed by SIGKILL +++

    Read the article

  • Hardware recommendations for building an Ubuntu encrypted file server

    - by Robert Mashlan
    I would like to build a file server for my home network using Ubuntu. It will serve files from RAID1 configured disks, either in the OS or in hardware. It will be connected to a Gigabit ethernet LAN. The disks will use an encrypted file system. It will serve samba shares. I would like a recommendation on what kind of processing power/memory I would need to build a box that would be able to sustain the full capacity of the Gigabit ethernet connection in a file transfer for a single connection with the overhead of serving from an encrypted disk. I'm not looking to build a dream server, I just want enough processing capacity for high performance (and reliable) file sharing and spend as little as possible for it. This may be tangential, but what kind of hardware would I need to have a server be able to reliably go into a low power mode when no requests are being made of it?

    Read the article

  • Strange Phantom Local Disks appearing in my drive list...

    - by Paul
    Win7 Home Prem 32 bit I seem to have several phantom Local Disks mapped to different letters, they are of 0 bytes in size? Strangely they do not show up when i view my drives through windows explorer but if i open an application such as ACDSee Pro or MS Word and then go to open a file i can see all these Local Disks mapped to different letters. This means when i plug in my external hard disk it ends up mapped to letter R instead of its usual G which messes up any programs i have pointing to it by default. How did they get there and more importantly how do i get rid of them please??

    Read the article

  • Can I use dis-similar HW for Win2008r2 DFS-R

    - by cwheeler33
    The setup: Windows 2008R2 Ent on two machines. The roles on each server will include File Servers and DC's. The machines come from two different vendors (Dell/HP) The Dell is an Athlon and the HP is an Intel. Both have roughly the same speed CPU and 8GB of RAM. They have different Raid controllers, and more or less the same amount of disk space (roughly 6TB.) Can the servers use different types of hardware? Is there any documentation about this? The last question I have is about the network. Can DFS-R be forced to use a differen subnet from the regular network?

    Read the article

  • More than one XP Mode virtual machines okay under Win 7?

    - by kousen
    I really like the Windows XP Mode virtual machine that comes with Windows 7 (once you download the integration components). I teach technical training classes, and having a machine configured for the classroom is very helpful. My question is, can I configure more than one? The documentation suggests you can make multiple virtual machines, but I can't tell if I can have multiple "Windows XP Mode" VMs. It can't be as simple as configuring one, then just making a copy of the primary disk file, could it? I only plan to run one at a time, but I'd like to configure multiple different VMs for different classes, all running XP if possible. Thanks for any help!

    Read the article

  • limit the speed of writing files to NFS

    - by xgwang
    CentOS 5.6 NFS is mounted on the server for backup disk space. When the backup job started, it could reach 80MB/s and we really do not expect it took so much bandwidth. So i need to find a way to limit the speed of writing to NFS. I tried rsync with --bwlimit=5000. However, it did limit the reading speed, but the accumulated data still was written at 80MB/s, and no writing activities for seconds. Is there any way to limit the writing speed of NFS?

    Read the article

  • Why Windows Media Center (Vista 32-bit) unable to record TV Series?

    - by Alex R
    I can record individual shows just fine, but trying to setup the recording of a Series throws an error (something like, "unable to schedule, try restarting Media Center", pops up in the bottom right of the screen). Trying to access the "Other Showings" menu causes Media Center to crash ("check for a solution online" dialog comes up). The machine is a fresh reinstall of 32-bit Windows Vista from the HP recovery CDs, plus all Microsoft updates and service packs. Hardware: HP s3220n, AMD X2 5000+, 2GB RAM, 300GB free space on disk. Media Center came bundled. The "Media Center update for Vista" failed to install (it says, not appropriate for my computer). Any ideas? Thanks

    Read the article

  • WindowsXP+7 on C:/D: :moving the System Partition

    - by user938921
    I had installed Windows 7 on a separate 20gig partition, and I'm absolutely loving it! Plus I can now dual-boot, with my original WinXP residing on the C: drive. But I'm running out of disk space on D:, and I was able to shrink C: and expand D:. But now I would like to make D: not just a Boot Partition, but an Active System Partition, without losing my ability to boot into Windows 7 (since it was created on a separate D: partition, not the current Active System C: Partition). Any advice?

    Read the article

  • After moving our Servers to a virtual environment using VMware - SQL timeouts came in, why?

    - by RayofCommand
    We moved our servers to a virtual cloud (VMware) where only our servers are in. But as soon as we finished migrating everything we are fighting against SQL Timeouts and machine slowdowns we can't explain. Even though we ~ doubled the servers capacity while switching from physical to virtual. Now I googled and found that we are not alone. People are complaining about poor performance after moving to a cloud managed by VMware. Are there any known issues? Sometimes our services can't access a disk or SQL receives a timeout and we have no idea why.

    Read the article

  • Best practice for administering a (hadoop) cluster

    - by Alex
    Dear all, I've recently been playing with Hadoop. I have a six node cluster up and running - with HDFS, and having run a number of MapRed jobs. So far, so good. However I'm now looking to do this more systematically and with a larger number of nodes. Our base system is Ubuntu and the current setup has been administered using apt (to install the correct java runtime) and ssh/scp (to propagate out the various conf files). This is clearly not scalable over time. Does anyone have any experience of good systems for administering (possibly slightly heterogenous: different disk sizes, different numbers of cpus on each node) hadoop clusters automagically? I would consider diskless boot - but imagine that with a large cluster, getting the cluster up and running might be bottle-necked on the machine serving the OS. Or some form of distributed debian apt to keep the machines native environment synchronised? And how do people successfully manage the conf files over a number of (potentially heterogenous) machines? Thanks very much in advance, Alex

    Read the article

  • Moving software RAID to Linux

    - by terman
    I'm using a RAID 1 (mirrored pair) configuration in my Media Center/ NAS system. Currently it's running Windows 8 (yeah, big mistake I know) and I'm regretting it (did it for the games, not worth it). Currently I'm having two software RAID 1s (3TB + 2TB) configured with Storage Spaces and unfortunately formatted with NTFS. Now I would like to switch to Fedora (or maybe Ubuntu if there are advantages) for good. Is there a way that I could continue using the disks as they are without the need to format them with ext or something? I'm glad for every tip. Oh, the system disk is of cause not in a raid configuration.

    Read the article

  • people_dl_import shows millions of records

    - by amit lohogaonkar
    We have a situation now on prod in sharepoint 2007 based intranet platform and it shows thousands of records under people_dl_import category with format spsimport://?$$dl$$/domain1/domain2/domain3/ Also import was not stopping and added millions of records in database and was on verge of disk full. On other servers like dev we have very less data in this category and format is also like spsimport://doaminname?$$dl$$?... which is good and has only 6000 rows and in prod its 2 millions Crawled under people_dl_import category. I need to know the cause of this garbage data and how to fix it. I tried resetting content source and I will do full import in this weekend to see if this garbage data gets cleared. Any idea on cause for thiss issue?

    Read the article

  • Defrag/TRIM settings for hybrid hard drive?

    - by Joel Coehoorn
    I recently acquired a momentus XT hybrid hard drive. This is a tradition spinning HDD with a small SSD portion for frequently used files. Normally, you are not supposed to defrag SSDs, as it does not help performance and can significantly reduce the life of the drive. But you do need to defrag an HDD to keep good performance. So where does that leave us for hybrid drives? Do I need to turn off defragging in Windows to preserve the SSD portion? Is the on-disk controller smart enough to handle defragging correctly? Is there some utility I need to set it up that I missed? Additionally, for SSDs you normally want to check for and enable TRIM support... but this makes no sense for a HDDs. Where does that leave us for hybrid drives? Should I try to enable TRIM or not?

    Read the article

  • Visual Studio Development on Virtual Box, Boot Camp, or VMWare Fusion

    - by Eli
    I currently have a Mac, 2ghz and 2 gigs of ram, running OS X Leopard and Virtual Box with a Windows 7 Pro 32bit virtual machine. Performance on the virtual machine is fine for minor tasks but is very clunky while trying to multi-task or develop in Visual Studio 2008. What would be my best option for being able to use Visual Studio, keeping cost and time in mind? 1) Upgrade ram to 4 gigs ($100). Will this really improve my performance enough to use Visual Studio in a Windows 7 vm? Or am I just wasting time/money? 2) Reinstall/restore Windows 7 disk image as a Boot Camp partition. I assume this should improve my performance, yes? 3) Purchase VMWare fusion instead of VirtualBox. Does Fusion require less resources to run? I am open to any suggestions. Thanks in advance

    Read the article

  • Virtualbox: Raw linux partition not booting

    - by abalter
    I have a dual-boot laptop with Windows 7 and Ubuntu 12.04. I am trying to boot the ubuntu partition from windows using Virtualbox. I have successfully created the .vmdk, and created the virtual machine. However, I can't get it to boot (in Virtualbox). All I get is a black screen with the cursor in the top left. I wonder if I'm specifying the partitions correctly. My Ubuntu install has 3 partitions: \, \boot, \home. No swap partition. These are all in Disk 0, partitions 3,4,5 respectively. The command I used to create the .vmdk is: VBoxManage internalcommands createrawvmdk -filename C:\Users\abalter\.virtualbox\ubuntu.vmdk -rawdisk \\.\PhysicalDrive0 -partitions 3,4,5 Then I create a virtual machine based on that .vmdk. Why won't it boot?

    Read the article

< Previous Page | 254 255 256 257 258 259 260 261 262 263 264 265  | Next Page >