Search Results

Search found 28995 results on 1160 pages for 'sandy good'.

Page 591/1160 | < Previous Page | 587 588 589 590 591 592 593 594 595 596 597 598  | Next Page >

  • Deleting large no of files on linux eats up CPU

    - by Sanjay
    I generate more than 50GB of cache files on my RHEL server (and typical file size is 200kb so no of files is huge). When I try to delete these files it takes 8-10 hours. However, the bigger issue is that the system load goes to critical for these 8-10 hours. Is there anyway where I can keep the system load under control during the deletion. I tried using nice -n19 rm -rf * but that doesn't help in system load. P.S. I asked the same question on superuser.com but didn't get a good enough answer so trying here.

    Read the article

  • Using the right folder for the right job. Article link, please?

    - by Droogans
    There are specific folders designed for specific tasks. /var/www holds your web sites, /usr/bin contains files to run your applications...yet I still find myself putting nearly all of my work in ~. Is it possible to overuse my home directory? Will it come back to haunt me? Anyone have a good link to an article of best practices for organizing your files so that they are placed in their "correct" place? Is there even such a thing in Linux? I am referring specifically to user-generated content. I do not compile applications from source, I use apt-get for those tasks. This article has a great introduction to what I'm looking for. Table 3-2, "Subdirectories of the root directory" is the sort of thing I'm looking for, but with more details/examples.

    Read the article

  • Sql server 2000 replication error

    - by Renato
    Good morning, I have a sql server 2000 machine, with SP4. I have a transactional replication. A keep receving this error message: The process could not execute 'sp_replcmds' on 'servername' the logreader stops responding.When i click start it starts fine, and the replication starts fine.And then, works for hours, and the problem comes back. In the begining i though it could be timeouts, but i already set a couple of parameters in the logreader profile,like querytimeout/readbatchsize. Sometimes when the logreader stops,it generates a dump, but not always. In event viewer, it appears this message: 17066 : SQL Server Assertion: File: , line=1985 Failed Assertion = 'startLSN = m_curLSN'. 18052 : Error: 3624, Severity: 20, State: 1. 17066 : SQL Server Assertion: File: , line=2223 Failed Assertion = 'm_noOfScAlloc == 0'. I also executed checkdb in the databases, and they are fine. have you every experienced something similar? thanks in advance, Renato Alves.

    Read the article

  • Stability, x86 Vs Sparc

    - by Jason T
    Our project are plan to migrate from Sparc to x86, and our HA requirement is 99.99%, previous on Sparc, we assume the hardware stability would like, hardware failure every 4 month or even one year, and also we have test data for our application, then we have requirement for each unplanned recovery (fail over) to achieve 99.99% (52.6 minutes unplanned downtime per year). But since we are going to use Intel x86, it seems the hardware stability is not so good as Sparc, but we don't have the detail data. So compare with Sparc, how about the stability of the Intel x86, should we assume we have more unplanned downtime? If so, how many, double? Where I can find some more detail of this two type of hardware?

    Read the article

  • Finding text in Ubuntu (gnome) terminal output

    - by Rickson
    Imagine this scenario: You run a command at gnome terminal. This command has made a bunch of outputs to the terminal. After some time, you realize you need the value of a variable (let's say variable_needed) that was printed by the command somewhere in the terminal. How to find it? KDE terminal used to have a shortcut ctrl+shift+f which searched the terminal output. It seems that gnome-terminal doesn't have it (at least at Ubuntu 10.04.2 LTS). Is there any way of adding it? Is there any other good terminal I could use that has it? Notice that the output has already been written so I don't want (cannot) run the command again combined with grep, |, , vim, emacs, etc.

    Read the article

  • Linux stretch cluster: MD replication, DRBD or Veritas?

    - by PieterB
    For the moment there's a lot of choices for setting up a Linux cluster. For cluster manager: you can use Red Hat Cluster manager, Pacemaker or Veritas Cluster Server. The first one has the most momentum, the second one comes by default with RH subscriptions and the last one is very expensive and has a very good reputation ;-) For storage: - You can replicate LUN's using software raid / md device - You can use the network using DRBD replication, which offers a bit more flexibility - You can use Veritas Storage Foundation technology to talk to your SANs replication technology. Anyone has any recommandations or experience with these technologies?

    Read the article

  • SAN shows as unallocated in Windows Server

    - by Gareth Ferneyhough
    Hello. We have a SAN drive that shows as unallocated in Windows Server 2008. I believe it is a raid 10 with 4+ disks. The disks are in good health. I think a server that we rebuilt tried to connect to the drive and re-initialized them, or re-wrote the partition table. (excuse my poor terminology). We ran TestDisk on the drive and it shows no partitions, so now we are doing a quick search (which is not so quick). Can anyone else suggest anything? Thanks, Gareth

    Read the article

  • Speeding up remote X sessions

    - by Alexander Gladysh
    I have a remote machine running Ubuntu 11.10 Server, to which I connect via SSH from OS X 10.7.3: Host remote.example.com ForwardX11 yes ForwardX11Trusted yes Sometimes I want to launch a GUI application there (most notably, gitk). But UI is rather slow. Both remote machine and my own Mac have good internet connection. Ping between them is about 55ms, and there is no packet loss. Is there something I can do to speed things up? Setting up some other remote desktop solution is an option, but I'd like to avoid it since I don't need this application very often.

    Read the article

  • What software supports full multisession CD burning on windows?

    - by dma_k
    I know, that beating Nero in this field is very difficult. I have looked through best software listed in wiki, also checked the related posts here and here: most of the software either allows you to create ISO image and/or burn the existing image. However I am interested in the following flow: Import previous session Add/rename/remove files and directories Burn new session UltraISO is very good in (1) and (2). But it cannot append a disk with new session. InfraRecorder/CDBurnerXP can do (1) (2) (3) but: InfraRecorder crashes when trying to import some Nero sessions CDBurnerXP does not allow to move files in previous session to another directories. Is anybody here happy with any other free software?

    Read the article

  • CRL checking problem windows 2003

    - by Tim Mahy
    Hi all, we have CRL that is valid for 24 hours and has a next update in 12 hours. The CRL is valid from 12:12 AM to 12:12 AM and from 12:12 PM to 12:12 PM. In the logs of the CRL hosting webserver we see that one of our servers not always fetches the CRL at night, in most cases the server that missed the CRL IIS servers 403.16 on 12:13 PM. Is our following theory good: when a windows server misses fetching the CRL on it's nextUpdate but the current CRL is still valid, the fetching is not retried? This leads to a situation that when the CRL expires there is no overlap and gives a little time of 403.16 situations in IIS since the CRL is not thrusted and so all certificates are marked als unsafe? greetings, Tim

    Read the article

  • shared web hosting architecture in a university setting

    - by gaspol
    We're in the process of creating a shared webhosting infrastructure for our university. Departments within the university can host their sites on this infrastructure. We're thinking of setting up multiple, load balanced web servers attached to shared storage (for web content and Apache config files). There will also be database servers behind these web servers. Does anyone have any other suggestions about this? Any recommendations for an alternative setup? Would having cPanel/WHM/Plesk be a good idea to automate account creation/maintenance?

    Read the article

  • Configuring vsftpd with nginx on ubuntu

    - by arby
    I have vsftpd installed on Ubuntu 12.04LTS along with nginx, php, and sql on an Amazon ec2 instance. The web server is good to go, but I'm having trouble connecting to the FTP server. I'm not quite sure how to set the privileges or what configuration options I might be missing. By default, the location of the web root is at /usr/share/nginx/www and it is owned by root:root. The web server runs as user www-data in the group www-data. I've opened port 21 and set the passive ports in the ec2 backend and ufw firewall. In vsftpd.conf, I have: ... anonymous_enable=NO local_enable=YES local_umask=0027 chroot_local_user=YES pasv_enable=YES pas_max_port=12100 pasv_min_port=12000 port_enable=YES ... Now, I'm unsure how to create the FTP user that when I login, displays my web directory with write access. I've tried it a few different ways, but I keep running into errors (either no connection, no write access, or very slow timeouts.)

    Read the article

  • Sane(r) way to get character-encoding of CLI?

    - by danyowdee
    Hi all! I was writing a CLI-Tool for Mac OS X (10.5+) that has to deal with command-line arguments which are very likely to contain non-ASCII characters. For further processing, I convert these arguments using +[NSString stringWithCString:encoding:]. My problem is, that I couldn't find good information on how to determine the character-encoding used by the shell in which said cli-tool is running in. What I came up with as a solution is the following: NSDictionary *environment = [[NSProcessInfo processInfo] environment]; NSString *ianaName = [[environment objectForKey:@"LANG"] pathExtension]; NSStringEncoding encoding = CFStringConvertEncodingToNSStringEncoding( CFStringConvertIANACharSetNameToEncoding( (CFStringRef)ianaName ) ); NSString *someArgument = [NSString stringWithCString:argv[someIndex] encoding:encoding]; I find that a little crude, however -- which makes me think that I missed out something obvious...but what? Is there a saner/cleaner way of achieving essentially the same? Thanks in advance D

    Read the article

  • Making TT-RSS cache images

    - by Piku
    Due to Google Reader's untimely demise, I've installed tiny-tiny RSS on my Linux machine under Apache 2. It's mostly a good enough replacement and I can at least go back to reading RSS feeds in my web browser at work. Can I configure or hack TT-RSS to cache all the images it finds in its feeds? There is an option when adding a feed, but it doesn't seem to actually do anything. If I view (for example) today's XKCD comic in TT-RSS it still loads the image from the XKCD website. What I want is the image to be cached in TT-RSS and served from there instead.

    Read the article

  • Better performance with memcached cluster or local memcaches?

    - by Nicholas Tolley Cottrell
    I have a small cluster of servers balancing a Java web app. Currently I have 3 memcached servers caching data and all web apps shares all 3 memcached instances. I often get strange slowdowns and timeouts to some of the memcacheds and I wondering if there is a good way of analyzing the performance. I am wondering whether my iptables rules (or some other system limitation) are blocking/slowing connections. I am considering reconfiguring the web apps so that they only query the memcached process on their own localhost.

    Read the article

  • MySQL replicate multiple places

    - by Frederik Nielsen
    Very trick task to find a good title for this question, but here goes the q: I have a few development machines, where I develop my PHP applications on, and testing via a local webserver. This works out pretty well for each machine. However, I would like to replicate the DB from my machines to a central location. So, to sum up: DEV1 - CENTRAL DEV2 - CENTRAL DEV3 - CENTRAL CENTRAL - DEV1 CENTRAL - DEV2 CENTRAL - DEV3 I hope this makes sense, as I cannot find an easy way to tell it. Basically, it is a 2-way replication, where all 4 databases contain the same info, and each of them can be updated locally, to then be pushed out to the others. Is this actually doable? All my dev machines are running Windows 7, and my central DB server is running CentOS 6.

    Read the article

  • Tips and tricks to make NX server more stable

    - by gareth_bowles
    My shop has been using the FreeNX server on Fedora 11 for a while now and mostly getting good results, especially with performance, but we have some annoying problems with client connections. There are two main issues: Client sessions sometimes freeze after a long time (seems to be at least 2 hours of having the session active) We often have to make multiple attempts to start a new client session, especially if a previous session was suspended rather than terminated. In qwuite a few cases, we've had to restart the NX server to get around this. Our NX server configuration is the default except that we've enabled logging level 7 to /var/log/nxserver.log, and set the font server to "unix:/7100" so that it uses xfs. Does anyone have any ideas for making things more stable ?

    Read the article

  • Can somebody help me install this jBPM based workflow management suite?

    - by Eternal Saint
    1.Its a book Workflow Interface software available at sourceforge http://bookworkflowint.sourceforge.net/ Any instructions on installing and configuring it would be great especially in windows, however I can try Linux specific ones as well. I could not find any installation instructions. I posted this in stackoverflow by mistake and was directed here. 2.Can you guys suggest any good scanning/digitization workflow software (document imaging) that I can adopt to my scanner software? Infact even a simple one would do may be based on hotfolders. I just want to be able to track the uniqueid/barcode of the scanned book, its status so that its not scanned again. It books or manuscripts could be in millions page count. I thought of using some kind of generic bug tracking tool, just track a few fields, dont know if its the right choice Thank you very much

    Read the article

  • How can I share a video file during a webinar?

    - by Brien Malone
    Here is the scenario: I have a number of remote employees around the globe. I want to have a video chatting session. No problem there. Halfway through, I want to shut off all camera video feeds and simulcast (synchronous) a training video to my team. How do I do this? We have tried office communicator, but the frame rate was awful and no audio. Adobe Connect had similar trouble. In both cases we were limited by the main office's small internet pipe, but it is clear that video delivered by shared desktop is not a good solution.

    Read the article

  • NIC: Changing Link Speed & Duplex to Full Duplex Is Drastically Slower

    - by rism
    I have an old Win Xp box with a 100 PRO/100 VE NIC and connecting to a Win 7 box with gigabit NIC via an ASUS gigabit router and peer to peer. I was getting 76% network utilisation on the XP machine using AUTO as the value for Link Speed and Duplex which is fine. But I decided to see what would happen if i selected Full Duplex 100Mbps on the XP box. I thought at worst it would be as good as auto detect since the Win 7 box it's transferring to is faster. Instead network utilisation dropped to 0.12% (as in a 10th of a percent) and the transfers just timed out. No biggie to reverse back to Auto but my question is why is full duplex such a dog? Im not "stuck" but just dont understand. There are only 2 machines here connecting through a router so i cant see how that would cause alot of collisions etc (if thats the problem).

    Read the article

  • Converting a Word document to LaTeX format

    - by Mehper C. Palavuzlar
    I'm preparing a book to be published and keeping everything in .docx files. Other than text the files include graphs (jpeg) and lots of equations typed in MathType. Since MS Word is not fully appropriate to balance text and shapes according to book format, some pages are having spacings at the bottom after some text, and then comes a shape on the next page. I know that LaTeX is very good at formatting, so is it possible to convert MS Word documents (or PDF documents, since I can easily convert them to PDF) into LaTeX format so that I can handle my work in LaTeX from now on?

    Read the article

  • rsync for copying file

    - by vinayrks
    I am migrating my old server to new server . I used this server for hosting website . first I tried sftp but due to huge number of files and connection time out , it simply didn't work . then I tried rsync .rsync working good , but only problem I am facing it updating file very nicely & fastly but do not copy new files please help me . because still i need to transfer lots of file. I am using this command : rsync -anv -e ssh oldserver:/path/ /path

    Read the article

  • AWS ELB as backend for Varnish Accelerator

    - by addisonj
    I am working on a large deployment on AWS that has high uptime requirements and variable loads throughout the day. Obviously, this is the perfect use case for ELB (Elastic Load Balancer) and autoscaling. However, we also rely on varnish for caching of API calls. My initial instinct was to structure the stack so that varnish uses ELB as a backend which in turn hits an appGroup. Varnish -> ELB -> AppServers However, according to a few sources that isn't possible as ELB constantly changes the IP address of its DNS hostname, which varnish caches on start, meaning changes to the IP won't be picked up by varnish. Reading around however, it looks like people are doing this so I am wondering what workarounds exist? Perhaps a script to reload the vcl periodically? In the case of where this is really just not a good idea, any idea of other solutions?

    Read the article

  • SQL Server 2008 R2 100% availability

    - by Mark Henderson
    Is there any way to provide 100% uptime on SQL Server 2008 R2? From my experience, the downtimes for the different replication methods are: Log Shipping: Lots (for DR only) Mirroring w. NLB: ~ 45 seconds Clustering: ~ 5-15 seconds And all of these solutions involve all of the connections being dropped from the source, so if the downtime is too long or the app's gateway doesn't support reconnection in the middle of task, then you're out of luck. The only way I can think to get around this is to abstract the clustering a level (by virtualising and then enabling VMWare FT. Yuck. Good luck getting this to work on a quad-socket, 32-core system anyway.). Is there any other way of providing 100% uptime of SQL Server?

    Read the article

  • Automatizing the backup of my databases and files with cron

    - by Patrick
    hi, I want to automatize the backup of my databases and files with cron. Should I add the following lines to crontab ? mysqldump -u root -pPASSWORD database_name | gzip > /home/backup/database_`date +\%m-\%d-\%Y`.sql.gz svn commit -m "Committing the working copy containing the database dump" 1) First of all, is this a good approach ? 2) It is not clear how to specify the repository and the working copy with svn" 3) How can I run svn only when the mysqldump is done and not before ? Avoiding conflicts Any other tip ? thanks

    Read the article

< Previous Page | 587 588 589 590 591 592 593 594 595 596 597 598  | Next Page >