Search Results

Search found 31410 results on 1257 pages for 'disk based'.

Page 611/1257 | < Previous Page | 607 608 609 610 611 612 613 614 615 616 617 618  | Next Page >

  • Partition and mount my secondary hard drive on CentOS 5.5 64bit?

    - by Andrew Fashion
    I am trying to prepare my second hard drive for user image uploads. Here is the current layout: # sudo parted /dev/sda print Model: ATA WDC WD2500KS-00M (scsi) Disk /dev/sda: 250GB Sector size (logical/physical): 512B/512B Partition Table: msdos Number Start End Size Type File system Flags 1 32.3kB 107MB 107MB primary ext3 boot 2 107MB 8595MB 8488MB primary linux-swap 3 8595MB 10.7GB 2147MB primary ext3 4 10.7GB 250GB 239GB extended 5 10.7GB 250GB 239GB logical ext3 Information: Don't forget to update /etc/fstab, if necessary. I am assuming #4 is my secondary drive? How do I partition and mount it so I can begin using it? And how do I add to fstab? I understand if it's to many questions in one, just help me with whatever you can I guess :) Thank you for any help!

    Read the article

  • Would upgrading memory from 4GB to 8GB on my laptop solve swapping issues?

    - by Tom
    I have a laptop with 4GB of memory with Windows 7 on it and I often experience with Eclipse that it is swapped out to disk. On the net they usually write 4GB of RAM is more than enough for average use and aside from Eclipse+Android Emulator I don't really use other extra apps, yet Eclipse is always swapped out if I haven't used it for a while (say, 1 day) and it is annoying it to wait for it to be resurrected from swap. My question is: would an upgrade to 8GB solve the issue of swapped out applications? With 8GB would windows 7 keep everything in memory? Or it wouldn't change anything and Eclipse would be swapped out regardless of the amount of memory, because Win 7 has a habit of kicking out every application from memory which hasn't be used for a while?

    Read the article

  • raid 5 creation (using mdadm) lots of read/writes on creation: is this normal?

    - by Gbrits
    I created a software raid 5 disk using: mdadm -C /dev/md2 -l5 -n4 /dev/sd[i-l] at the same time I'm using dstat to see io-activity: dstat -c -d -D total,sda1,md2,sdi,sdj,sdk,sdl -l -m -n and notice that disks sd[i-k] are all read from and sdl is written to. Now, I do understand that raid5 has to be configured, but it takes a really long time and all disks are clean & formatted (using xfs) so I figure there might be some kind of shortcut to skip (unnecessary? ) checking.. Is it? The creation is part of a time-critical nightly batch-process (run on amazon ec2) so it's not a one-time thing. Thanks, Geert-Jan

    Read the article

  • Replicate a big, dense Windows volume over a WAN -- too big for DFS-R

    - by Jesse
    I've got a server with a LOT of small files -- many millions files, and over 1.5 TB of data. I need a decent backup strategy. Any filesystem-based backup takes too long -- just enumerating which files need to be copied takes a day. Acronis can do a disk image in 24 hours, but fails when it tries to do a differential backup the next day. DFS-R won't replicate a volume with this many files. I'm starting to look at Double Take, which seems to be able to do continuous replication. Are there other solutions that can do continuous replication at a block or sector level -- not file-by-file over a WAN?

    Read the article

  • Is it possible to run a "Windows Server 2008 R2 SP1 Hyper-V Virtual Machine" on Windows 7x64 Pro machine?

    - by tbone
    Specifically, I want to run this: http://www.microsoft.com/en-us/download/details.aspx?displaylang=en&id=27417 This download contains a three Windows Server 2008 R2 SP1 Hyper-V Virtual Machine set for evaluating and demonstrating Office 2010, SharePoint 2010 and Project Server 2010. System requirements Supported operating systems: Windows Server 2008 R2 Additionally you will need: •Windows Server 2008 R2 (SP1 recommended) with the Hyper-V role enabled. •Drive Formatting: NTFS •Processor: Intel VT or AMD-V capable •RAM: 8 GB or more recommended •Hard disk space required for install: 100 GB ------------------------- So, the above text seems to indicate that you need a physical Windows 2008 Server (r2) server running, but from the googling I've done on the subject I've yet to come across a discussion that definitively answers the question. Many posts I've read seem to indicate it might be possible to run it on Windows 7 using one of the following: VirtualBox, Windows Virtual PC, VMWare but I'm not entirely sure.

    Read the article

  • WindowsXP+7 on C:/D: :moving the System Partition

    - by user938921
    I had installed Windows 7 on a separate 20gig partition, and I'm absolutely loving it! Plus I can now dual-boot, with my original WinXP residing on the C: drive. But I'm running out of disk space on D:, and I was able to shrink C: and expand D:. But now I would like to make D: not just a Boot Partition, but an Active System Partition, without losing my ability to boot into Windows 7 (since it was created on a separate D: partition, not the current Active System C: Partition). Any advice?

    Read the article

  • List of recent motherboards with BIOS / without UEFI [on hold]

    - by jmn
    I am building a new desktop PC and I want to have full disk encryption on it. TrueCrypt doesn't support UEFI as of now. Are there still recent motherboards out there without UEFI ? I didn't find any list and I am afraid that I will have to study each potential candidate's technical sheet before purchase. I want to buy 2 or 3 of the same model to be future proof. Newegg links will not help, I don't live in the USA ... this means that this post is a legitimate target for PRISM ;-) Thanks for your help.

    Read the article

  • Create and copy a Windows Mobile 5.0 operating system image

    - by user20119
    We have several dozen Windows Mobile 5.0 devices (Symbol MC7095 handhelds equipped with embedded Verizon WLAN, if that matters) that all need the same software and configuration. We connect all of these devices via a USB cradle to add software to them via Microsoft ActiveSync, and then do several configuration changes directly on the handhelds themselves, in the OS. That process takes 30 minutes or more, per device. Is there any way to set up one device and take a 'disk image' of the entire OS/software, such that things could then be copied (quickly/easily) to the other devices? Is such a thing possible, with Windows Mobile devices?

    Read the article

  • XP Laptop Restarting

    - by Liath
    I have a Dell 1525 which I reinstalled xp on recently due to a hard disk problem. Apparently I've missed something, it's taken to restarting for no apparent reason (I believe it's too frequent for windows updates which was my first thought). Last night it was particularly bad, it stopped 5 times in 2 hours while playing an online game. I can't actually think of a time it's happened when it's not running something in full screen mode. I can't see anything in the event log apart from usual start up details. How can I find and fix this issue?

    Read the article

  • Best practice for administering a (hadoop) cluster

    - by Alex
    Dear all, I've recently been playing with Hadoop. I have a six node cluster up and running - with HDFS, and having run a number of MapRed jobs. So far, so good. However I'm now looking to do this more systematically and with a larger number of nodes. Our base system is Ubuntu and the current setup has been administered using apt (to install the correct java runtime) and ssh/scp (to propagate out the various conf files). This is clearly not scalable over time. Does anyone have any experience of good systems for administering (possibly slightly heterogenous: different disk sizes, different numbers of cpus on each node) hadoop clusters automagically? I would consider diskless boot - but imagine that with a large cluster, getting the cluster up and running might be bottle-necked on the machine serving the OS. Or some form of distributed debian apt to keep the machines native environment synchronised? And how do people successfully manage the conf files over a number of (potentially heterogenous) machines? Thanks very much in advance, Alex

    Read the article

  • linux or vmware

    - by bahar
    Hi I don't know if here is the right place to ask this question or not but its vital for me to solve it. I run open suse 10.2 on a vmware and my operating system is vista. yesterday I run my vmware and another program but both couldn't run and I end both tasks to solve it but now I can open vmware without any problem and run another os but not my open suse. the error is: Cannot open the disk 'C:\Users\hamide\Documents\Virtual Machines\SUSE Linux 2\SUSE Linux 2-000003.vmdk' or one of the snapshot disks it depends on. Reason: Failed to lock the file. please help me I don't have any recent back up and It takes too long to do all my project again.:( Bests

    Read the article

  • How to automate slipstream?

    - by Gregory MOUSSAT
    Since years I use slipstreamed Windows installations. This works very well, but preparing them is tedious : 1 - install a Windows with the last slipstreamed version we have (automated install) 2 - check Windowsupdate to see what's new, and take note 3 - download each new update available 4 - go to step 2 until no new update is available 5 - slipstream them into the last version we have (I already automated this step) I'd like a way to automate parts or all of this. Maybe a program able to know which updates are installed (already saw one, I don't remember which, and I know PowerShell can do this)... and able to download them ? Or to get them from local disk ? So the steps become : 1 - install a Windows with the last slipstreamed version we have (automated install) 2 - use Windowsupdate until no new update is available (any way to automate ?) 3 - use the magic program 4 - slipstream

    Read the article

  • Hardware recommendations for building an Ubuntu encrypted file server

    - by Robert Mashlan
    I would like to build a file server for my home network using Ubuntu. It will serve files from RAID1 configured disks, either in the OS or in hardware. It will be connected to a Gigabit ethernet LAN. The disks will use an encrypted file system. It will serve samba shares. I would like a recommendation on what kind of processing power/memory I would need to build a box that would be able to sustain the full capacity of the Gigabit ethernet connection in a file transfer for a single connection with the overhead of serving from an encrypted disk. I'm not looking to build a dream server, I just want enough processing capacity for high performance (and reliable) file sharing and spend as little as possible for it. This may be tangential, but what kind of hardware would I need to have a server be able to reliably go into a low power mode when no requests are being made of it?

    Read the article

  • php-cgi got SIGKILL on burst.net VPS?

    - by Shawn
    I got an VPS at burst.net, very cheap one, but this doesn't matter. The strange behavior of it is, the php-cgi process, which started using lighthttpd's spawn-cgi, dies every a few minutes. however, other processes are fine and great, even include a java process, and I'm sure there is no "out of memory" issue, so it's not killed by OOM killer. I used strace to trace the process, and found out it was killed by SIGKILL, hence no single log was left on the disk, just dies suddenly. Is there anyway I can find out what process/thing sent the SIGKILL to the poor php process? Filed a ticket with the vendor, but they said they won't care. strace -p 7176 Process 7176 attached - interrupt to quit wait4(-1, <unfinished ...> +++ killed by SIGKILL +++

    Read the article

  • Cannot delete folder - Content seems to be nested recursively

    - by RikuXan
    I cannot delete a folder located on my hard disk by any means. I don't quite know how it was created, all I know is, that it is a pretty deep structure of folders (too deep to delete it at once, since Windows restriction path name too long), but the problem in the end is, that I can't "pull out" the inner folders, because they don't seem to be folders anymore (Context menu lacks things like "Properties", "Cut", "Copy", "Delete" etc.) Here a picture of how a right click looks like on one of these "folders": As you can see, the current folder is in very deep, but that is not the problem, rather the one I left-clicked on. Has anyone any advice on how to get rid of these? I tried a chkdsk, said no errors. I also tried deleting those folder via a VMWare Ubuntu, to no success. I also tried a batch file from a volunteer at MS boards, that should automatically de-nest such folders, but I guess mine is a special case, since the tool only created more such folders.

    Read the article

  • Recover partition table after DD command

    - by Shreedhar
    I executed the following command from a Ubuntu live cd terminal (dont ask why). dd if=/dev/zero of=/dev/sdb2 bs=512 count=1 Where sdb2 is a NTFS partition (third partition) on a disk. Suffice to say it is now messed up. When I boot into windows 7, it does show me E drive but when I click on it it asks me to format it. I am not ever sure what I did, did I mess up partition table or only the MFT? Is there any way to get the data back PLEASE HELP! this is very important :(

    Read the article

  • Problems with the backup

    - by marcodv
    I did a script which run around 4 o'clock in the morning, for backup all the mysql databases and the config file for 250 linux vm. The problem is that it tooks ages for complete and more than 50% of these vm, need more than 8 hours for complete. More or less all the vm had the same configuration,I mean Same amount of ram same amount of disk space same number of cpu Debian 6.0.5 I am saving these backup on amazon s3, because is the cheapest solutions that I've found. Now my questions is: Has anyone some solutions or suggestions about that? On one blog I've read that probably the ionice and nice combination could be good work around about that. any thought?

    Read the article

  • limit the speed of writing files to NFS

    - by xgwang
    CentOS 5.6 NFS is mounted on the server for backup disk space. When the backup job started, it could reach 80MB/s and we really do not expect it took so much bandwidth. So i need to find a way to limit the speed of writing to NFS. I tried rsync with --bwlimit=5000. However, it did limit the reading speed, but the accumulated data still was written at 80MB/s, and no writing activities for seconds. Is there any way to limit the writing speed of NFS?

    Read the article

  • Moving software RAID to Linux

    - by terman
    I'm using a RAID 1 (mirrored pair) configuration in my Media Center/ NAS system. Currently it's running Windows 8 (yeah, big mistake I know) and I'm regretting it (did it for the games, not worth it). Currently I'm having two software RAID 1s (3TB + 2TB) configured with Storage Spaces and unfortunately formatted with NTFS. Now I would like to switch to Fedora (or maybe Ubuntu if there are advantages) for good. Is there a way that I could continue using the disks as they are without the need to format them with ext or something? I'm glad for every tip. Oh, the system disk is of cause not in a raid configuration.

    Read the article

  • Windows 7 formatting HD

    - by Rick
    I have an Acer Aspire 5610 that I am trying to put an Operating System back on. I am trying to install windows 7 32-bit and it will not go through the install. Not sure why. I am now trying to reformat the hard drive and start with a fresh formatted HD. No other partitions. The problem is that the install disk is not doing the format it is giving me an error. The error is 0x8007045d. I am at a loss. This is the first time running into this issue. Please Help.

    Read the article

  • Is there a software in windows that enables you to boot from a specific partition?

    - by Tono Nam
    I use acronis true image to mount images to my primary partition and it works great. lets say I have 3 partitions on my hard disk and all of them each is 600 GB. In the 3rd partition I keep files (documents, pictures etc), on the first partition is my primary partition where the operating system runs (windows 7). And in the 2nd partition is empty. I have an image of my primary partition and I save that image in my 3rd partition (50 GB is the size of the image so it fits in the partition number 3) and in an external hard drive. I know it is possible to install a new operating system in partition 2 such as windows xp but the only problem is that once I install that how could I tell the computer to boot from partition 1? is there a way to switch back and forth just like it's possible in the mac?

    Read the article

  • Linux live cd with Broadcom Wi-fi support

    - by paul simmons
    I am looking for a live distro that has out of the box Broadcom wireless support. I am pretty happy with my Ubuntu installation and as long as I have an ethernet connection first time installed, I can install Broadcom drivers over internet. But being a little paranoid, I make my secure operations (banking etc.) with a live cd and zero hard disk access, so nothing is recorded. So far I plug ethernet to do such things with the live cd, but it would be nice if I can do same thing with wireless.

    Read the article

  • Mounting Solaris UFS partition on Debian(with FreeBSD kernel)

    - by hayalci
    I have some disks that were being used on a Solaris system. The disks are formatted as UFS. I attached them to a Debian system (with FreeBSD kernel. Debian/kFreeBSD), but I cannot mount them. $ mount -t ufs /dev/da2s1 /mnt/diska mount: /dev/da2s1 : Invalid argument Also the tunefs.ufs does not work; $ tunefs.ufs -p /dev/da2s1 tunefs.ufs: /dev/da2s1: could not read superblock to fill out disk Is there an incompatibility between FreeBSD UFS and Solaris UFS? Is it possible to mount one, under the other OS ? Note: tunefs.ufs works on the root partition $ tunefs.ufs -p /dev/da7s2 tunefs.ufs: ACLs: (-a) disabled tunefs.ufs: MAC multilabel: (-l) disabled tunefs.ufs: soft updates: (-n) disabled tunefs.ufs: gjournal: (-J) disabled tunefs.ufs: maximum blocks per file in a cylinder group: (-e) 2048 tunefs.ufs: average file size: (-f) 16384 tunefs.ufs: average number of files in a directory: (-s) 64 tunefs.ufs: minimum percentage of free space: (-m) 8% tunefs.ufs: optimization preference: (-o) time tunefs.ufs: volume label: (-L)

    Read the article

  • New power supply, now computer doesn't recognise hard drive?

    - by Mike
    Ok, I bought a new power supply, because my old one was too damn loud. I hooked it up to my PC, turned it on, everything is looking fine, start up detects my DVD drive, 2 hard disks.. then I get the message "BOOT FAILURE INSERT SYSTEM DISK". Now I've seen some other people talk about going into BIOS and changing the start up to the HDD and not the CD.. well I've done that and it doesn't help. If I let windows load up and it asks me to which partition I wish to install windows, no partition is present. It's as if after the initial start up the drives arn't being found. I plugged my old (but loud) PSU back in, connected up all the cables, and it works perfectly. Why does the new PSU not detect my HDD's after the first BIOS screen start up? Any ideas? :)

    Read the article

  • Disable writing RAID degraded mode

    - by jolivier
    I have a RAID5 with 5 disks on my machine and suspect the motherboard chipset to fail at some points and make my raid going in degraded mode. Last time it happened I noticed it on the failure of the 2nd drive connected to the same chipset and lost a lot of data. So I would like to prevent this, and especially I would like to have mdadm disable writes on the raid if one of the disk fails. So that in between I get notified, I recover and can use my system again. Sadly I could not find it in man mdadm so I was wondering if this is possible via a tool or hidden option since for me it looks like a standard feature of a RAID system. If this is not possible I would also be happy with a solution to stop the raid if degraded.

    Read the article

< Previous Page | 607 608 609 610 611 612 613 614 615 616 617 618  | Next Page >