Search Results

Search found 20904 results on 837 pages for 'disk performance'.

Page 347/837 | < Previous Page | 343 344 345 346 347 348 349 350 351 352 353 354  | Next Page >

  • Which SSL certificate to buy [closed]

    - by Sparsh Gupta
    I am reading several notes on SSL certificates and comparison. What matters to me the most is speed. I can read that encryption is same with all different certificates available but I was wondering if there is any difference in the performance of the website with different certificates involved. I am ofcourse interested in end to end response times and I wonder if the type of encryption or number of certificates required as Chain Certificates makes a difference in speed. I dont really care for cost but looking for a good SSL certificate which ideally gives me absolutely no pain and best performance. Recommendations?

    Read the article

  • Why doesn't Ghost 2003 offer to fill the destination drive?

    - by Neil
    Because it is dangerously low on disk space, I want to upgrade an SBS 2003 server by replacing its existing 72GB drive with a 364GB drive. When I tried to use Norton Ghost 2003 to clone the disk it didn't suggest that I use the entire new drive. I'm worried that I caused the process to fail by overriding its decision - although the cloned drive boots in Safe Mode, if I try booting it normally then none of the SQL Express instances start and something causes the server to reboot before even the Ctrl+Alt+Del screen appears. Does Ghost 2003 know something that I don't? Or should I be using some other software?

    Read the article

  • How to access both partitions on a bootable USB flash drive in Windows

    - by Maccyjam
    I have a 16GB USB Flash Drive that is partitioned into two different sizes. The first partition contains a bootable version of Ubuntu, the second partition is for general saving of files. Windows will only recognise the first partition. I have tried using Bootice but this breaks the bootable partition. Disk Management recognises the second partition but does not allow me to do anything with it. Is there a way to make both partitions accessible by Windows and keep the USB disk bootable?

    Read the article

  • Tips for maximizing Nginx requests/sec?

    - by linkedlinked
    I'm building an analytics package, and project requirements state that I need to support 1 billion hits per day. Yep, "billion". In other words, no less than 12,000 hits per second sustained, and preferably some room to burst. I know I'll need multiple servers for this, but I'm trying to get maximum performance out of each node before "throwing more hardware at it". Right now, I have the hits-tracking portion completed, and well optimized. I pretty much just save the requests straight into Redis (for later processing with Hadoop). The application is Python/Django with a gunicorn for the gateway. My 2GB Ubuntu 10.04 Rackspace server (not a production machine) can serve about 1200 static files per second (benchmarked using Apache AB against a single static asset). To compare, if I swap out the static file link with my tracking link, I still get about 600 requests per second -- I think this means my tracker is well optimized, because it's only a factor of 2 slower than serving static assets. However, when I benchmark with millions of hits, I notice a few things -- No disk usage -- this is expected, because I've turned off all Nginx logs, and my custom code doesn't do anything but save the request details into Redis. Non-constant memory usage -- Presumably due to Redis' memory managing, my memory usage will gradually climb up and then drop back down, but it's never once been my bottleneck. System load hovers around 2-4, the system is still responsive during even my heaviest benchmarks, and I can still manually view http://mysite.com/tracking/pixel with little visible delay while my (other) server performs 600 requests per second. If I run a short test, say 50,000 hits (takes about 2m), I get a steady, reliable 600 requests per second. If I run a longer test (tried up to 3.5m so far), my r/s degrades to about 250. My questions -- a. Does it look like I'm maxing out this server yet? Is 1,200/s static files nginx performance comparable to what others have experienced? b. Are there common nginx tunings for such high-volume applications? I have worker threads set to 64, and gunicorn worker threads set to 8, but tweaking these values doesn't seem to help or harm me much. c. Are there any linux-level settings that could be limiting my incoming connections? d. What could cause my performance to degrade to 250r/s on long-running tests? Again, the memory is not maxing out during these tests, and HDD use is nil. Thanks in advance, all :)

    Read the article

  • Dell Inspiron 530 - SSD Worth it?

    - by DrFredEdison
    I'm going to be upgrading my Dell Inspiron 530 (2.0 Ghz Intel Dual Core CPU, 3 GB RAM) to windows 7 soon, and rather than backup and reformat my existing drive, I'm planning on getting a 2nd drive to replace my current primary, and moving it to a secondary. Thus, this seems like an excellent time to get a solid-state drive, if its going to be worth it. As far as I can tell this machine has a SATA-I controller, and I'm unsure if I'll see a noticeable performance increase with an SSD without going to SATA-II. So I have a three part question here given all that: Will spending the money on a SSD be worth it if hook it into a SATA-I controller? Is it reasonable to upgrade the controller on this machine to a SATA-II controller? Given that this PC is kind of old to begin with, am I better off performance wise to just stick with a faster HDD?

    Read the article

  • cdrom drive doesn't work on laptop

    - by bozdoz
    Here's as best as I can describe it: When starting up, the boot order doesn't recognize the cdrom drive, but I can open and close the drive during this time. In Windows 7, I can't open the cdrom drive, and it doesn't recognize it in device manager, disk management, or my computer. In Ubuntu Linux, I can open the cdrom drive, but it still doesn't recognize the cds, and it won't mount. If I reformat everything, would my cdrom drive work again? Can I reinstall Windows without a cdrom drive? I've deleted the upper and lower filters as was suggested in Google searches. Took the disk drive out and checked that it was installed correctly (no reason it shouldn't have been). Still: nothing works.

    Read the article

  • "Device not ready" on a network share in Windows 7

    - by user60689
    I have two computers C1 and C2 runing Windows 7 and both of them are members in a domain. On C1 I have an USB hard-disk which I shared for the users U1 and U2 giving them Read-Only permissions on the entire drive. However, even if I can see and browse the hard-disk localy (IOW from C1), from the other computer (C2) where I'm logged with U1, trying to access the C1's shared device, the C2's Windows 7 throws an error saying "Device Not Ready". Why? How can I fix this? PS: Tried to un-share and re-share again. No luck.

    Read the article

  • repair/recovery tool for Western Digital Elements (1TB)

    - by Pennf0lio
    My Hard Disk Drive just got corrupted due to electricity fluctuation. When I plug my Western Digital Elements, It ask me if I want to format it or not... I can't see my files or even the capacity of my disk from it's properties. Is there any solution you would suggest? looking for a software that can give me access to my files. I just need to copy it then I can retire my Drive and will buy a new one... Thanks!

    Read the article

  • Is it safe to compress my Windows 7 %USERPROFILE%\AppData folder?

    - by Kev
    Having just read Scott Hanselman's latest blog entry, Guide to Freeing up Disk Space under Windows 7, he suggests turning on NTFS compression which I already do for a number of less travelled folders that contain static files such as downloads or images. However I am wondering if it's wise to turn on NTFS compression for the whole of my %USERPROFILE%\AppData folder? My system drive is a 128 GB SSD residing in a Dell Precision T5400 3Ghz Quad Core Xeon workstation so I ought not to notice the extra cycles used to compress and decompress files on their way to and from the disk. However would there be any good reasons not to do this? In fact could I safely compress the whole of my %USERPROFILE% folder?

    Read the article

  • Should I choose KVM/XEN over OpenVZ or use them together?

    - by Krystian
    I've got a dual xeon e5504 server, with [for now] only 8GB of ram. Storage is'n impressive either: 3x 146GB sas in raid5 + 500GB sata drives. Currently it works as a development server, but it's over speced for our needs and since our development methods changed through last 2 years we decided it will work as a production system for some of our applications + we would like to have a separate system for testing/research. Our apps are mainly web apps deployed on tomcats [plural as some of the apps require older versions] and connected to Postgres. I would like to have a production system, where only httpd+tomcat+db are setup and nothing else runs there. Sterile system. Apart from that, I would like a test system, where I can play with different JVM settings, deploy my test apps, play with tomcat/httpd settings and restart them without interfering with the production system. Apart from that, I would like to be able to play with different linux flavors, with newer kernels to test how they work etc. I know, this is not possible with OpenVZ and I would have to choose KVM for that. I am thinking about merging the two, and setting up a KVM to be able to work with different systems [linux only to be frank] + use openVZ to setup separate machines for my development needs. I would simply go with that, but reading here and there about the performance impact full virtualization has over containers and looking at the specs of my server makes me think twice about it. I don't want to loose too much performance, especially because of the nature of my apps [few JVMs running at the same time]. It will be my first time with virtualization, apart from using desktop virtualbox/vmserver. Although I am a fast learner I don't want to mess with the main system so much that it will break the production apps or make them crawl. Although they are more or less internal apps and they don't produce much load, they need to be stable. I've read, that KVM host is a normal linux installation and it allows to run normal processes on it. If that is so, does it allow to run openVZ as well? I mean... can I have KVM and OpenVZ running on the same system/kernel? Or do I have to setup another system to run OpenVZ containers? How much performance impact can this have for me? Will my hardware suffice? oh and one more thing... unfortunately I'm quite limited with the funds... I'm looking for a free solution only :/

    Read the article

  • Raid 5 scsi fault

    - by HaLaBi
    I have no much knowledge about servers and I was looking all day around the internet about finding a solution to my raid 5 problem. All of a sudden two disks failed. The server won't boot (HP Proliant, windows 2003 R2, very old maybe 10 years old). I know that if one disk is faulty then I can add a new disk and rebuild it and things will be fine, the problem is two went faulty :( is this normal? two at the same time? is there any other thing I can do and I am not aware of? other than taking them out and reinserting them back? Windows won't boot. The Array menu shows that disks 0 and 4 are "Missing". Any other tricks or things to do? It is important because for some unknown reason the back up job did not work for a month and I just found out, so I need to make these raid 5 back online again.

    Read the article

  • SDcard /dev/sdb2 is apparently in use by the system; will not make a filesystem here

    - by user171223
    I divided my sdcard into 2 partitions, but It got an error and couldn't create a new partition. Error: /dev/sdb2 is apparently in use by the system; will not make a filesystem here! My /dev/sdb was not mounted, and the output of command lsblk was: cxphong@cxphong:~/Desktop$ lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sr0 11:0 1 1024M 0 rom sda 8:0 0 465.8G 0 disk +-sda1 8:1 0 118.8G 0 part +-sda2 8:2 0 147.7G 0 part /media/DATA +-sda3 8:3 0 137.1G 0 part +-sda4 8:4 0 1K 0 part +-sda5 8:5 0 1023M 0 part [SWAP] +-sda6 8:6 0 61.2G 0 part / sdb 8:16 1 3.7G 0 disk +-sdb1 8:17 1 70.6M 0 part +-sdb2 8:18 1 3.6G 0 part +-sdb1 (dm-0) 252:0 0 70.6M 0 part +-sdb2 (dm-1) 252:1 0 3.6G 0 part I couldn't delete /dev/sdb1 (dm-0) & /dev/sdb2 (dm-1). What are they?

    Read the article

  • User WinWget to keep web site alive in a Windows Server 2003

    - by Menelaos Vergis
    I have a site that must stay alive due to a service that runs and check a directory for changes. The site is running in IIS at a Windows Server 2003 and the solution I came up it that I will Schedule a task that requests the home page every 5 minutes. I am sure that this way the site will stay alive almost all the time. I have downloaded Wget from Wget from Windows and I have installed it at my windows server 2003 but I don't know how to use it in order to ping the server but not download anything. Since I want to use this forever I don't want to save anything on the disk, can you provide me with the command that pings a web page but don't save anything on the disk?

    Read the article

  • Black screen with cursor after BIOS screen

    - by Radio
    Here is a weird one, Got computer with Windows XP. It's getting stuck on a black screen with cursor blinking. What did I do: - Boot from installation CD (recovery option - command line): chkdsk C: /R copy D:\i386\ntdetect.com c:\ copy D:\i386\ntldr c:\ fixmbr fixboot Chkdsk showed 0 bad sectors and no problems during scan. dir on C:\ shows all directories and files in place (Windows, Program Files, Documents and Settings). BIOS shows correct boot drive. Still does not boot. Not sure what to think of. Please help. UPDATE: Just performed these steps: Backed up current disk C: (without MBR) using True Image to external hard drive Ran Windows XP clean installation with deleting all partitions and creating new one. Hard drive booted fine into Windows GUI installation!!! Then: I interrupted installation. Booted from True Image recovery CD and restored archive of disk C to an new partition. Same issue with black screen.

    Read the article

  • Best alternatives to recover lost directories in FAT32 external hard drive?

    - by Sergio
    I have an 320 GB ADATA CH91 external hard drive. I guess it has some problems with the connector of the USB jack. The point is that in certain occasions it fails in write operations generating data losses. Right now I lost a directory with several GB's of very useful information. Since then I have not attempted to write to the disk any more. What tool would you recommend to recover the lost data? The disk is FAT32 formatted (only one partition) and I use both Linux and Windows. What filesystem format would you recommend to avoid future data losses? I currently only use this external hard drive in Linux so there are several available choices (FAT, NTFS, ext3, ext4, reiser, etc.).

    Read the article

  • ubuntu boots into gnu grub 1.99

    - by greenish
    I've tried set root=(hd0,2) chainloader +1 boot set root=(hd0,2) linux /boot/vmlinuz... and the loopback (loop0) /ubuntu/disks/root.disk command etc. When I try the boot command it tells me there's no kernel and when I boot Win7 (it's a dual boot) the root.disk says 0kb. nothing boots from the live usb I've made and I've tried to use programs to mount the partitions to no effect - they only show me what's on my windows file drives. I've got some really important docs on the linux harddrive I need to get to. any ideas?

    Read the article

  • Server OS: put it on a separate drive? Yes, no, or depends on the situation?

    - by captainentropy
    Hi, I would like opinions, or facts, both preferably, on whether it's ok to install a server's OS on the RAID array or not. I would predict installation on separate drives is the best but I'm interested in the performance. The server in question will have 8 cores (2.4GHz ea.), 24GB RAM, and ~16TB of usable space of server-class drives in RAID10. There is also a subsytem of an ~equivalent size for backup. I will be running CPU/memory intesive applications on this server in addition to it being file storage for my work (research lab). IF I install the OS (haven't decided which one, probably Ubuntu or Fedora or some other good linux distro) on separate drives will there be any performance problems if they aren't configured in RAID10? IF it is better to have the OS on separate drives should I go for 150GB velociraptors in RAID1 or smallish SSD drives in RAID1? Money is unfortunately a factor as I think I'm close to maxing my budget as it is. Thanks!

    Read the article

  • How can I use target mode in Linux with USB?

    - by dash17291
    Kernel 3.5 introduces: This release includes a driver for using an IEEE-1394 connection as a SCSI transport. This enables to expose SCSI devices to other nodes on the Firewire bus, for example hard disk drives. It's a similar functionality to Firewire Target Disk Mode on many Apple computers. This release also adds a usb-gadget driver that does the same with USB. The driver supports two USB protocols are supported that is BBB or BOT (Bulk Only Transport) and UAS (USB Attached SCSI). BOT is advertised on alternative interface 0 (primary) and UAS is on alternative interface 1. Both protocols can work on USB 2.0 and USB 3.0. UAS utilizes the USB 3.0 feature called streams support. http://kernelnewbies.org/Linux_3.5 I have an Arch Linux with kernel 3.5.3-1 and wanna try out this feature.

    Read the article

  • Defrag/TRIM settings for hybrid hard drive?

    - by Joel Coehoorn
    I recently acquired a momentus XT hybrid hard drive. This is a tradition spinning HDD with a small SSD portion for frequently used files. Normally, you are not supposed to defrag SSDs, as it does not help performance and can significantly reduce the life of the drive. But you do need to defrag an HDD to keep good performance. So where does that leave us for hybrid drives? Do I need to turn off defragging in Windows to preserve the SSD portion? Is the on-disk controller smart enough to handle defragging correctly? Is there some utility I need to set it up that I missed? Additionally, for SSDs you normally want to check for and enable TRIM support... but this makes no sense for a HDDs. Where does that leave us for hybrid drives? Should I try to enable TRIM or not?

    Read the article

  • Cloning Fresh Windows 7 with external HDD, DVD or CDs?

    - by hhh
    I want to create Windows installation disk (not necessarily CD) from my Windows 7 Pro laptop, not sure what it actually means here. The material is about 50-60GB, requiring about 12 pcs of 5GB disk (not going to happen, too much work to use them later). I have the Windows serial on the laptop bottom and this is so-called firm -laptop (no idea what it actually mean, better warranty and some instant support thing apparently). Now how to do the clone with external material such as DVDs? how can I create mock-windows-installation medium or some real windows-installation medium? I am now not sure whether Windows offers W7 -installations medium online, well my W -peer mentioned some MSDAA -something. Perhaps related Cloning Fresh Windows 7 -fsed HDD to Linux Server because having no external HDD or disks for the Backup

    Read the article

  • Memcached clustered alternative

    - by Johan Kooijman
    I'm looking to replace memcached. We have a LOT of traffic to our central memcached node which I'd like to split. There's only so much trunking networks I can do. My general idea is to install a memcached-type daemon on every webserver and have the daemons replicate set/delete/updates over all the daemons, so that each webserver connects to a socket or on localhost. All data should be available on all nodes. The alternatives: - repcached (max 2 masters) - redis (single master) - couchdb/mongodb/handlersocket - persistent data on disk, I'd like to remove the disk part to gain more performance. Any hints?

    Read the article

  • "The requested operation could not be completed due to a file system limitation" 3202

    - by user46529
    I backup SQL Server database and it fails BACKUP DATABASE dd TO DISK = '\backupServer\backups\dd.bak' WITH COMPRESSION, CHECKSUM, NOFORMAT, INIT , BlockSize = 65536 , BufferCount = 2200 , MaxTransferSize = 4194304 The backup size is 3TB and I have 6TB free space on bacup server. I am using backup parameters per SQLCAT whitepaper. Everything works ok when I backup to local HDD and it always fails when I backup to network share. After about 6 hours. Can't find why. Thank you. Yes. The backup over the network is fastest and saves me 3Tb of local disk space :) Thanks for pointing to the memory issue. I left 4Gb to OS and it worked!

    Read the article

  • Windows 7 stopped booting

    - by jstawski
    I have a laptop which after shutdown it stopped booting. I tried repair, safe mode, and even start with Windows 7 Installation. The screen just goes blank with a mouse pointer that I can move around. I removed the harddrive from the laptop and connected it to my desktop using an External HD casing. The computer recognizes the disk, but it seems like it can't read it. If I go to My Computer it shows up, but it doesn't display usage information. When I double click on the drive it sits there as if it was loading something and eventually shows "G:\ is not accessible. The parameter is incorrect." Disk Management and Diskpart also take forever to load and when it does it shows the drive. My question, do you think this is a hardware problem or some corrupted sector? How can I try to fix the drive without formatting it? Thanks...

    Read the article

  • Is it possible to boot from internet?

    - by Calmarius
    Let's assume the following situation: you have a computer with empty hard disk, and you don't have a CD, floppy, pen drive nearby to boot the computer from. But you have connection to the Internet. Modern computers support network booting using PXE, but I haven't found anything regarding booting via the internet. So, is it possible to use PXE to load an image from the internet and boot it? By having a running system (even a minimal Linux) in RAM, it should be possible to install it on the hard disk, and build up a working system from here.

    Read the article

< Previous Page | 343 344 345 346 347 348 349 350 351 352 353 354  | Next Page >