Search Results

Search found 30742 results on 1230 pages for 'folder size'.

Page 588/1230 | < Previous Page | 584 585 586 587 588 589 590 591 592 593 594 595  | Next Page >

  • linux disk usage report inconsistancy after removing file. cpanel inaccurate disk usage report

    - by brando
    relevant software: Red Hat Enterprise Linux Server release 6.3 (Santiago) cpanel installed 11.34.0 (build 7) background and problem: I was getting a disk usage warning (via cpanel) because /var seemed to be filling up on my server. The assumption would be that there was a log file growing too large and filling up the partition. I recently removed a large log file and changed my syslog config to rotate the log files more regularly. I removed something like /var/log/somefile and edited /etc/rsyslog.conf. This is the reason I was suspicious of the disk usage report warning issued by cpanel that I was getting because it didn't seem right. This is what df was reporting for the partitions: $ [/var]# df -h Filesystem Size Used Avail Use% Mounted on /dev/sda2 9.9G 511M 8.9G 6% / tmpfs 5.9G 0 5.9G 0% /dev/shm /dev/sda1 99M 53M 42M 56% /boot /dev/sda8 883G 384G 455G 46% /home /dev/sdb1 9.9G 151M 9.3G 2% /tmp /dev/sda3 9.9G 7.8G 1.6G 84% /usr /dev/sda5 9.9G 9.3G 108M 99% /var This is what du was reporting for /var mount point: $ [/var]# du -sh 528M . clearly something funky was going on. I had a similar kind of reporting inconsistency in the past and I restarted the server and df reporting seemed to be correct after that. I decided to reboot the server to see if the same thing would happpen. This is what df reports now: $ [~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/sda2 9.9G 511M 8.9G 6% / tmpfs 5.9G 0 5.9G 0% /dev/shm /dev/sda1 99M 53M 42M 56% /boot /dev/sda8 883G 384G 455G 46% /home /dev/sdb1 9.9G 151M 9.3G 2% /tmp /dev/sda3 9.9G 7.8G 1.6G 84% /usr /dev/sda5 9.9G 697M 8.7G 8% /var This looks more like what I'd expect to get. For consistency this is what du reports for /var: $ [/var]# du -sh 638M . question: This is a nuisance. I'm not sure where the disk usage reports issued by cpanel get their info but it clearly isn't correct. How can I avoid this inaccurate reporting in the future? It seems like df reporting wrong disk usage is a strong indicator of the source problem but I'm not sure. Is there a way to 'refresh' the filesystem somehow so that the df report is accurate without restarting the server? Any other ideas for resolving this issue?

    Read the article

  • computer plays movie fine but ps3 plays it with no sound?

    - by kacalapy
    i have a movie that plays fine on my computer but when i copy it to an SD card and play it in my PS3 it has no sound. the audio format is mpeg layer-3 bit rate is 128 frame rate is 25frames/ second data rate is 1068 kbps video sample size is 12 bit video compression is DivX i got these stats by right clicking the file and viewing properties. can i open it in some editor and re-save it? what editor? thanks all

    Read the article

  • Domain Forwarding | A Magento Store

    - by WaZ
    I have installed Magento inside a folder called magento. The URL of the site currently looks like this: http://gios.azamdevelopment.co.uk/magento/ We want our domain forwarded to the above URL and moreover, any links relative should work as well. e.g. http://gios.azamdevelopment.co.uk/magento/customer/account/login/ should ideally be www.giosconcept.com/customer/account/login and so forth. Thanks very much.

    Read the article

  • Subdomains, folders, internationalization, and hosting solutions

    - by justinbach
    I'm a web developer and I recently landed a gig to develop the US / international version of a site for a company that's big in Europe but hasn't done much expansion into the US yet. They've got an existing site at company.com, which should remain visible to European customers after the new site goes up, and an existing (not great) site at company.us, which I'm going to be redeveloping (the .us site will be taken down when my version goes up--keep reading for details). My solution needs to take into account the fact that there are going to be new, localized versions of the site in the fairly near future, so the framework I'm writing needs to be able to handle localizations fairly easily (dynamically load language packs, etc). The tricky thing is the European branch of the company manages the .com site hosting (IIS-based) and the DNS, while I'll be managing the US hosting (and future localizations), which will likely be apache-based. I've never been a big fan of the ".us" TLD--I think most US users are accustomed to visiting the .com--so the thought is that the European branch will detect the IP of inbound traffic and redirect all US-based addresses to us.example.com (or whatever the appropriate localized subdomain might be), which would point to the IP address of my host. I'd then serve the appropriate locale-specific content by pulling the subdomain from the $_SERVER superglobal (assuming PHP). I couldn't find any examples of international organizations that take a subdomain-based approach for localization, but I'm not sure I have any other options as a result of the unique hosting structure here (in that there's not a unified hosting solution for the European and US sites). In my experience, the US version of an international site would live at domain.com/us, not at us.domain.com, and I'd imagine that this has to do with SEO (subdomains are treated as separate sites, so improved rankings for the US site wouldn't help the Canadian version if subdomains are used to differentiate between them). My question is: is there a better approach to solving this problem than the one I'm taking? Ideally, I'd like to use a folder-based approach (see adidas.com as an example of what I'm talking about), but I'm not sure that's a possibility given that the US site (and other localizations) will not be hosted on the same server as the rest of the .com. Can you, in IIS, map a folder (e.g. domain.com/us) to a different IP address? What would you recommend? Thanks for your consideration.

    Read the article

  • [Mac] Recover iCal 10.5 files in 10.6?

    - by shox
    Hi, I installed Mac OSX 10.6 and have the old 10.5 on a seperate HDD. Can I now simply copy the old iCal-Data into the new iCal-installation? I tried ~/Libarary/Application Support/ but there is no Folder called iCal? Thanks

    Read the article

  • Domain Forwarding | A Magento Store

    - by WaZ
    I have installed Magento inside a folder called magento. The URL of the site currently looks like this: http://gios.azamdevelopment.co.uk/magento/ We want our domain forwarded to the above URL and moreover, any links relative should work as well. e.g. http://gios.azamdevelopment.co.uk/magento/customer/account/login/ should ideally be www.giosconcept.com/customer/account/login and so forth. Thanks very much.

    Read the article

  • GUI tool to extract iso file contents on Linux (KDE)

    - by extropy
    I have an ISO CD image file and want to extract it's contents to a folder. I know there are ways to mount the image and stuff, but it's complicated. I'm looking for a GUI tool to open up the contets and extract needed files. On windows I would use WinRar to do this. K3B only allows me to burn the stuff, Arch does not work with ISO files :( Is there a similar tool on Linux, preferably from KDE world?

    Read the article

  • Disable shift override for autostart programs per user

    - by Jens
    When starting up windows, one can normally disable all programs in the autostart folder by pressing and holding the SHIFT key. This behavior can be disabled by creating the key ignoreShiftOverride with DWORD setting of 1 in the registry key HKEY_LOCAL_MACHINE\ Software\ Microsoft\ Windows NT\ CurrentVersion\ Winlogon. I'd like to be able to configure this setting for each individual user. How can this be done?

    Read the article

  • SPF blocking some emails. How to complete disable it?

    - by Rafael Colucci
    We have configured a SMTP server on IIS. We have not configured any SPF rule. Some emails are going straight to the badmail folder. Most of them have this error message: Address does not pass the Sender Policy Framework Is there any way to disable that? I know this is insecure and spammers could use it to send spam, but I really need it to be disabled for now until I have configured and tested the SPF correctly.

    Read the article

  • Server side version of 'receipients address' rule in Outlook

    - by lethalMango
    I currently have a variety of folders on Outlook 2010 which connects to my works Exchange server. I have all the emails received from clients added to the correct folder, what I want to do is make any emails in the recipient address (i.e. I've been cc'ed in to an email sent from another user) to be also moved. Is it possible to achieve this server side opposed to using the "with @domain.com in the recipients address" rule which is client side only?

    Read the article

  • What is the recommended approach to add static subdomains to a website?

    - by shg
    I would like to create a few static subdomains like: mycategory.mydomain.com in a rather small website and would like it to point to the folder: mydomain.com/mycategory without showing such redirection in browser address bar. What is an easiest way to achieve it? I can do it in either IIS settings, asp.net, C# code, etc I guess there are better ways then creating a few separate Sites in IIS - one for each subdomain.

    Read the article

  • Is pst file enough to backup my Outlook (Gmail)?

    - by Hasan Gürsoy
    I've backup from my Gmail account at Outlook 2010 beta. I see a pst file which is equal file size with my Gmail used storage. Is this file enough to protect my mails or do I need additional files or whatever? I'm not a regular outlook user. I've read this but I don't think I'm using "Signatures, Templates, Stationery etc."...

    Read the article

  • Zend Framework: XAMPP - Redirect/Rewrite

    - by webo
    Hello, I'm using Zend Framework but I have a little problem: How could I redirect internal all requests from localhost/zendframework/ to localhost/zendframework/www/ What have to put into my .htaccess in the folder localhost/zendframework/?

    Read the article

  • Avoid Windows Explorer to load complete executable file

    - by eli.work
    On Windows Vista, when browsing to a network folder containing executables, Windows Explorer seems to load all the files completely just to be able to show the executable icon (the resource monitor indicates loads of traffic during the loading of the directory) On XP only a part of the file is loaded. Is there a way to avoid the complete loading of these files? Note that disabling my anti virus does not help.

    Read the article

  • MongoDB and datasets that don't fit in RAM no matter how hard you shove

    - by sysadmin1138
    This is very system dependent, but chances are near certain we'll scale past some arbitrary cliff and get into Real Trouble. I'm curious what kind of rules-of-thumb exist for a good RAM to Disk-space ratio. We're planning our next round of systems, and need to make some choices regarding RAM, SSDs, and how much of each the new nodes will get. But now for some performance details! During normal workflow of a single project-run, MongoDB is hit with a very high percentage of writes (70-80%). Once the second stage of the processing pipeline hits, it's extremely high read as it needs to deduplicate records identified in the first half of processing. This is the workflow for which "keep your working set in RAM" is made for, and we're designing around that assumption. The entire dataset is continually hit with random queries from end-user derived sources; though the frequency is irregular, the size is usually pretty small (groups of 10 documents). Since this is user-facing, the replies need to be under the "bored-now" threshold of 3 seconds. This access pattern is much less likely to be in cache, so will be very likely to incur disk hits. A secondary processing workflow is high read of previous processing runs that may be days, weeks, or even months old, and is run infrequently but still needs to be zippy. Up to 100% of the documents in the previous processing run will be accessed. No amount of cache-warming can help with this, I suspect. Finished document sizes vary widely, but the median size is about 8K. The high-read portion of the normal project processing strongly suggests the use of Replicas to help distribute the Read traffic. I have read elsewhere that a 1:10 RAM-GB to HD-GB is a good rule-of-thumb for slow disks, As we are seriously considering using much faster SSDs, I'd like to know if there is a similar rule of thumb for fast disks. I know we're using Mongo in a way where cache-everything really isn't going to fly, which is why I'm looking at ways to engineer a system that can survive such usage. The entire dataset will likely be most of a TB within half a year and keep growing.

    Read the article

  • How to avoid compressing compressed files

    - by Gzorg
    Most compression programs compress all files by default. But when archiving a folder containing already compressed files, there is no need to compress them a second time, such as archives, packed setup program, jpg, movies, mp3,.... Are there any compression programs that allow an arbitrary list of type of files to be stored while the others are still compressed ? It looks like Winrar can't. I expect this would be doable with tar+gz/bzip2 and some scripting in various ways. Edit : Winrar can

    Read the article

  • Root SSH/SFTP Always 777

    - by Fluidbyte
    I have an Ubuntu serve that I'm connecting to via SFTP (and also an SSHFS mount locally). When I move a file to the server via the mount I need it to have permissions set to 777. I've added umask 000 to the .bashrc file at the advice of a friend and it doesn't appear to be working. Basically I'm working completely in a restricted folder and need the root to always leave the permissions open - wether I'm SSH'ed in or moving files to the server.

    Read the article

  • MySQL Backup - incremental

    - by Tiffany Walker
    I know that you can use mysqldump. I am currently dumping the following way: ${MYSQLDUMP} --single-transaction -u ${MUSER} -h ${MHOST} -p${MPASS} $db | ${GZIP} -9 > $FILE From my understanding this locks the database and prevents any type of use of the database and can even lock up websites. Is there a better way to maybe do daily/hourly backups of the MySQL database should the database be in the 100mbs and even 1gbs in size?

    Read the article

< Previous Page | 584 585 586 587 588 589 590 591 592 593 594 595  | Next Page >