Search Results

Search found 16174 results on 647 pages for 'disk space'.

Page 52/647 | < Previous Page | 48 49 50 51 52 53 54 55 56 57 58 59  | Next Page >

  • Calculate disk space occupied by many .png files

    - by Alexander Farber
    I have 357 .png files located in different sub dirs of the current dir: settings# find . -name \*.png |wc -l 357 settings# find . -name \*.png | head ./assets/authenticationIcons/audio.png ./assets/authenticationIcons/bbid.png ./assets/authenticationIcons/camera.png ./bin/icons/ca_video_chat.png ./bin/icons/ca_voice_control.png ./bin/icons/ca_vpn.png ./bin/icons/ca_wifi.png Is there a oneliner to calculate the total disk space occupied by them (before I pngcrush them)? I've tried (unsuccessfully): settings# find . -name \*.png | xargs du -s 4 ./assets/support/wifi_locked_icon_white.png 1 ./assets/support/wifi_vpn_icon_connected.png 1 ./assets/support/wi_fi.png 1 ./assets/support/wi_fi_conected.png 8 ./bin/blackberry-tablet-icon.png 2 ./bin/icons/ca_about.png 2 ./bin/icons/ca_accessibility.png 2 ./bin/icons/ca_accounts.png 2 ./bin/icons/ca_airplane_mode.png 2 ./bin/icons/ca_application_permissions.png 1 ./bin/icons/ca_balance.png

    Read the article

  • Disk / system configuration for log collection / syslog server

    - by Konrads
    I am looking into building a syslog / logging infrastructure and am pondering about some architecture best practices. Essentially, I see that a syslog system needs to support two conflicting workloads: log collection. Potentially massive streams of data need to be written quickly to disks and indexed. log querying. logs will be queried by both fixed fields such as date and source as well as text search. What is the best disk/system setup assuming I'd like to keep it to a single server for now? Should I use SSDs or ramdisk to off-load some processing? some disks in stripe and some in raid5? I am particularly eyeing Graylog2 with ElasticSearch/MongoDB

    Read the article

  • Calculate minimum ext3 partition size for certain amount of data

    - by Daniel Beck
    These following ext3 partitions contain identical data. As we can see, the larger the partition size, the more space is required for the same files: Filesystem 1K-blocks Used Available Use% Mounted on /dev/loop11 3965777 561064 3199964 15% [...] /dev/loop19 573029 543843 29186 95% [...] Filesystem Size Used Avail Use% Mounted on /dev/loop11 3.8G 548M 3.1G 15% [...] /dev/loop19 560M 532M 29M 95% [...] Filesystem Inodes IUsed IFree IUse% Mounted on /dev/loop11 1024000 1656 1022344 1% [...] /dev/loop19 1024000 1656 1022344 1% [...] I start with a partition of fixed size that possibly wasted a lot of space and I want to create a partition that is able to hold that data but with (almost) minimal size. How can I reliably calculate that minimal partition size needed for storing a certain amount of data? The amount of data changes over time, and I need to automate these calculations.

    Read the article

  • Clonezilla multiple disks restore to single disk

    - by Farseeker
    I have a clonezilla image from a machine that had 3 seperate disks (one partition per disk). I want to know if I can restore that image to another computer that has a hard drive that's much larger than the original, but only has one drive. Clonezilla is stating that it can't do this automatically, and perhaps I should try cnvt-ocs-dev but I've no idea what that means (Google is less than forthcoming with information about it too). Ok so I found out what cnvt-ocs-dev is, and that allows me to move source/destination targets between physical disks, but it doesn't seem to be able to move the partition as well.

    Read the article

  • Best filesystem choices for NFS storing VMware disk images

    - by mlambie
    Currently we use an iSCSI SAN as storage for several VMware ESXi servers. I am investigating the use of an NFS target on a Linux server for additional virtual machines. I am also open to the idea of using an alternative operating system (like OpenSolaris) if it will provide significant advantages. What Linux-based filesystem favours very large contiguous files (like VMware's disk images)? Alternatively, how have people found ZFS on OpenSolaris for this kind of workload? (This question was originally asked on SuperUser; feel free to migrate answers here if you know how).

    Read the article

  • How to get partition information from non-booting server?

    - by gravyface
    Need to manually rebuild a mirrored array on a server and am in the process of reinstalling SBS 2003 on it. However, it's a Dell server, and know that there's the Dell FAT32 diagnostics partition, a system partition, and a data partition, but do not know the size of each. Planning on reinstalling SBS 2003, all applications on the server, and then doing a System State restore, but figured that not having the correct partitions will cause some grief: am I right? Almost thinking that the size of the partitions shouldn't matter, but not positive. Question: should I care about the size of the partition? If so, how can I get this partition information from a non-booting drive? We have an Acronis image of the one working disk and the partitions are mounted/viewable in Explorer on a workstation, but I'm not sure where the Logical Disk Manager/Disk Management data is stored and/or if there's a way to retrieve it without having a working Windows installation.

    Read the article

  • After repairing permissions, Mac OS X won't boot

    - by Power-coder
    This morning I ran the Repair Permissions command from inside the Disk Utility. Ever since then my MacBook wont move past the splash screen when booting. I've revolted in verbose mode and I see that it is trying to repair the disk but then terminates with 'Unable to repair the volume'. Since then I have tried running the Disk Repair from the Snow Leopard install DVD and it quits with the same error. Is there a way I can repair this thing without reformatting and installing over again? How does something so simple as a permissions repair make the system unbootable like this?

    Read the article

  • Bitdefender rescue disk does not show my C drive

    - by Nilesh
    My machine (winXP) is infected with virus and I am unable to start my machine. Therefore I have used BitDefender rescue disk to remove the viruses. But I am not able to see my C drive. All other drivers are able to see and scan. Even after I removed all viruses, my machine is not starting. Giving the same error message. I think due to viruse it is not visible to bitdefender. Please help me out. Appreciate your help.

    Read the article

  • Shrink a mounted LVM partition

    - by javanix
    I fear I already know the answer to this question, but here goes. I need to carve out a new partition on a running system. /var/ is mounted from an LVM volume (hdd1_vg-var) and has only 3% used disk space. / is mounted separately (hdd1_vg-root) and has about 80% used disk space. Filesystem Size Used Avail Use% Mounted on /dev/**/hdd1_vg-root 2.0G 1.4G 481M 75% / /dev/**/hdd1_vg-var 33G 699M 31G 3% /var Unfortunately I don't have any free extents to grow this partition organically - vgdisplay shows: Total PE 10000 Alloc PE / Size 10000 / 39.06 GB Free PE / Size 0 / 0 So seeing that I have all this free disk space on /var/, can I shrink /var/ without un-mounting it or is this just a pipe dream? I am really hoping to be able to do this work on a running system - un-mounting would of course not be difficult but it would interfere with system functionality.

    Read the article

  • CentOS - Disk Quota X% warning

    - by jfreak53
    I currently have disk quotas working perfectly for Hard Limit Quotas on a CentOS 5 box. Quotas are working fine, but I am looking for a way to alert users either in a Cron or Automatically as Quotas is already running, when they are within X% of their Hard Limit Quota? I would like this email to go out to an email address I configure somehow for each user. I've looked all over the place but can't find it. All I can find is warnquota which only works when a user goes over Quota. The problem is I use Hard Limits, so my users will never go over quota. It needs to warn them when they are within X% of their limits.

    Read the article

  • Redundant Web Space

    - by alisia123
    I have following problem. My domain is registred on service "A" My web-space (not a server) is on Godaddy. Once a week is my service unreacheble I am sure that is godaddy problem. My idea is to by some webspace of one different service to make my service redundant. Note: My service is only few html files with javascrit. I dont' need to sync But how to do it? Whre do I have to say "there are two webspaces, if one is not reacheble, so use the other" ? Thanks a lot

    Read the article

  • BSOD constantly with same ntoskrnl.exe error and disk indicator is frozen

    - by Sheep
    BSOD constantly and the disk indicator is frozen. Error do not happen immediately, usually an hour after boot up. Here is the Minidump: Bug Check Code = 0x00000124 Caused By Driver = ntoskrnl.exe Caused By Address = ntoskrnl.exe+4b094c Crash Address = ntoskrnl.exe+4b094c Seems to be hardware problem, but I checked RAM, no error. I have two HDs installed, system is on SSD, data is on HDD. Checked SSD with the properties-tools-error-checking , no error. Re-installed several times, still happens even after removed HDD. Configuration: SSD: Crucial M4-CT064M4SSD2 with Firmware 0009 Intel HM65 CPU: i7-2630QM The SSD is set correctly, SATA III 6Gb/s enabled, and everything worked perfectly for nearly a year.

    Read the article

  • Software/FakeRAID: Windows 8 Disk Mirroring vs Intel Onboard

    - by Johnny W
    So Windows 8 is out and I have a new motherboard. I wish to create a RAID 1 coupling between two HDDs -- for storage purposes only (my OS is on an SSD) -- but I don't know which is the best route to take. My motherboard (Z77 chipset) comes with the age old Intel Fake RAID, but since I only wish to use my RAID for storage, I wondered if I might be better to use Windows 8 Disk Mirroring. Can anyone advise which is better? Or perhaps the pros and cons of each, if that's too contentious? I just can't see the benefit of FakeRAID. You can see my current setup here, if that might change things(?): Thanks!

    Read the article

  • Munin "Available entropy" when using address space layout randomization

    - by clawspoon
    Having just configured munin for statistics logging on my gentoo server (hardened profile), I am noticing that my "Available entropy" is consitently in the 200-300 range. This seems way to low, so I checked it manually using the command $ cat /proc/sys/kernel/random/entropy_avail 3544 Odd. Consistently very low values in Munin and practically filled up when checking manually. After thinking about the problem for a while I came to the conclusion that the problem is probably that I'm using Adress Space Layout Randomization which is using the entropy when running commands/programs. Since Munin runs a whole slew of programs all the entropy is used up, and Munin then measures how much entropy there is, resulting in the low values. Does anyone have any experience with this? How can this be avoided?

    Read the article

  • Backup Exec backup-to-disk folder creation - Access denied

    - by ewwhite
    I'm having a difficult time creating a backup-to-disk folder in Symantec Backup Exec 12.5 and Backup Exec 2010. The backend storage is a Nexenta/ZFS-based NAS filer sharing the volume via CIFS. I've also seen the issue on other *nix-based NAS devices. I've attempted mapping the drive, providing the full paths to the folder, etc. I can browse to the share just fine from within Windows, but Backup Exec fails to create the B2D folder with different variants of a Unable to create new backup folder. Access denied error. I've attempted creating service accounts in Backup Exec to handle the authentication, but nothing seems to work. What's the key to making this work?

    Read the article

  • Ubuntu 12.04 doesn't detect hard disk partitions

    - by m0skit0
    I just installed Ubuntu 12.04 32-bit on an ASUS Laptop and wanted to install it in another one, the ASUS A55V. After booting the laptop and setting up Windows 7 (I unfortunately cannot wipe out Windows), and creating one extra partition to install Ubuntu there, when I reboot from USB and launch Ubuntu installation, it doesn't recognize any partition on /dev/sda, but shows the whole disk as unpartitioned. I guess this is due to UEFI, but I don't know how I can make Ubuntu see the partitions created by Windows. Or any workaround for this to be able to install Ubuntu on /dev/sda. Thanks in advance

    Read the article

  • Losing partitions after every reboot

    - by Winston Smith
    I have an Acer laptop with one hard disk, which up until yesterday had 4 partitions: Recovery Partition (13GB) C: (140GB) D: (130GB) OEM Partition (10GB) I read that the OEM partition has all the stuff needed to restore the laptop to the factory settings, but since I'd already created restore disks and I needed the space, I wanted to get rid of it. Yesterday, I used diskpart to do that. In diskpart, I selected the OEM partition and issued the delete partition override command which removed it. Then I extended the D: partition into the unused space using windows disk management. Everything worked fine, until I rebooted my laptop, at which point the D: drive vanished. Looking in windows disk management again, I can see that there's an OEM partition of 140GB, which is obviously my D: drive. So I used EASEUS Partition Master and assigned a drive letter to the 'OEM' partition and I was able to access my files again. However, every time I reboot, it reverts back. How do I fix this permanently?

    Read the article

  • Formatting a disk for Macintosh using Linux

    - by Ken Bloom
    I've been asked to move data from an old external hard drive to a new one, and to make the new one compatible with the Macintosh. (The old drive's USB connection has died, and I'm connecting to old the drive using a PC card that provieds an eSATA to the drive. The recipient's Macintosh doesn't have a PC card slot, so she can't access the old drive anymore. Hence, the new drive.) Naturally, I'm doing this data transfer using Linux. I've discovered that I can format the drive as HFS+ using mkfs.hfsplus from the hfsprogs package. But I need to know: do I need to do anything special with the partition table? Is there a special Macintosh partition table format that I need to format this disk to? If so, what tools can I use to get the right format for the partition table?

    Read the article

  • CentOS default installation gave 60% disk space to tmpfs partition

    - by garconcn
    I installed a CentOS server which will be used for xen hypervisor. The server has two Intel(R) Xeon(R) CPU E5620 and 148G memory. The OS was installed on a 120G SSD drive. After the installation, I found that the tmpfs partition occupied about 60% of the drive. Even though I don't need much space for the OS, will there be any problem with 71G tmp partition? Thanks for any comment. [root@cloud ~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/mapper/VolGroup00-LogVol00 55G 1.1G 51G 3% / /dev/sda1 99M 13M 82M 14% /boot tmpfs 71G 0 71G 0% /dev/shm

    Read the article

  • Disk space mismatch on OS X Server (Leopard)

    - by John Gardeniers
    My Nagios system sent me an alert to inform me that the disk space on one of the drives on our OS X server is very low. When I run df /Volumes/Apps/ I get /dev/disk0s3 117209520 114932472 2277048 99% /Volumes/Apps When I run du -c /Volumes/Apps it reports 11489944 total Why might there be such a vast difference? Even more importantly, how do I find the problem and what can I do about it? I'm essentially just a Windows admin, so am well out of my comfort zone here. I use a Mac but I'm not a Mac admin in any real sense of the word.

    Read the article

  • What are performance limits of a database?

    - by Tommy
    What are some rough performance limits (read/s, write/s) for a single database server (no master-slave architecture), assuming storage on disk? How many read/s, write/s, depending on the kind of disk? (SSD vs non-SSD) , assuming simple operations (select one row by primary key, update one row, correctly indexed). I assume this limit is dependent on disk seek/write. EDIT: My question is more about getting rough metrics of the number of operations a database supports: to be able to know for example, if a new feature triggering 300 inserts/s can be supported without scaling out with additional servers.

    Read the article

  • How to tell if linux disk IO is causing excessive (> 1 second) application stalls

    - by noahz
    I have a Java application performing a large volume (hundreds of MB) of continuous output (streaming plain text) to about a dozen files a ext3 SAN filesystem. Occasionally, this application pauses for several seconds at a time. I suspect that something related to ext3 vsfs (Veritas Filesystem) functionality (and/or how it interacts with the OS) is the culprit. What steps can I take to confirm or refute this theory? I am aware of iostat and /proc/diskstats as starting points. Revised title to de-emphasize journaling and emphasize "stalls" I have done some googling and found at least one article that seems to describe behavior like I am observing: Solving the ext3 latency problem Additional Information Red Hat Enterprise Linux Server release 5.3 (Tikanga) Kernel: 2.6.18-194.32.1.el5 Primary application disk is fiber-channel SAN: lspci | grep -i fibre 14:00.0 Fibre Channel: Emulex Corporation Saturn-X: LightPulse Fibre Channel Host Adapter (rev 03) Mount info: type vxfs (rw,tmplog,largefiles,mincache=tmpcache,ioerror=mwdisable) 0 0 cat /sys/block/VxVM123456/queue/scheduler noop anticipatory [deadline] cfq

    Read the article

  • VMWare guest OS is not using full hard drive space

    - by syuusuke
    Hey everyone, I'm wondering if you can give me some insight on how to figure this out. I've recently virtualized a w2k3 server that had 10GB C:\ drive size. Once I finished virtualizing it, I expanded the disk to 20GB for C:\ and within guest OS (W2K3) computer management settings, you can see it recognizes as 20GB partition C:. But for some reason, if I opened windows explorer and see the properties of C:\, it will only show total size of 10GB. Does anyone know how to solve this or even had this problem before?

    Read the article

  • "Disk Not Found" Error when trying to install iso image of CentOs on ESXI using VSphere

    - by kamal
    When trying to install CentOs setup as an iso image on datastore, i get an error : "Disk Not Found". I am using ESXi 4.1 as host VM. Is there a way to attach DVD/CD as iso image ? I tried "Connect at Power on" Checked and Connect CD/DVD on the VSphere Console, but i still am unable ot get the iso image recognized as a DVD mounted image: Solved: if i look at the image size, it was 11 MB, so it WAS a boot iso, but nithing else, if we compare it with other boot images with complete iso's therein lies the solution. This boot.iso was from rBUILD/Conary packaging

    Read the article

  • Alerting when a RAID Array disk fails locally on VMWare ESX or ESXi System

    - by Tim K
    With ESX and ESXi, we recently had two systems where that the boot partition became degraded due to a failed disk. The only alert we managed to capture was the visual alert on the Dell servers. We failed to received any electronic alerts regarding the failed or degraded array. Does anyone have any experience with monitoring for these types of failures? In both cases, the servers were running in a RAID 5 SCSI configuration (5 disks on one system, 3 disks on another) which if we were running a Windows Server OS, we would have had an alert created in the Eventviewer. Where would I begin to look for this solution. Can it be configured in VCenter or vFoglight?

    Read the article

< Previous Page | 48 49 50 51 52 53 54 55 56 57 58 59  | Next Page >