Search Results

Search found 9715 results on 389 pages for 'bad passwords'.

Page 195/389 | < Previous Page | 191 192 193 194 195 196 197 198 199 200 201 202  | Next Page >

  • RAID10 without write-back cache = horrible write performance?

    - by Harry Mexican
    I have just provisioned a dedicated server on singlehop. I'm running it through some tests to know what to expect performance-wise. On the I/O side (with 4 1TB disks in RAID 10) I get: write-cache disabled 200 MB/s read throughput 30 MB/s write throughput I thought that was really low compared to my desktop HD which gets 150-150 or so. So I had a chat with them and they suggested enabling the write cache. New results: write-cache enabled 280 MB/s read 260 MB/s write which is great and all but means I'd have to add a BBU for an additional monthly cost. Is it normal for the write throughput to be 1/4 of a regular drive on RAID10, if you don't have write cache? It almost feels like its intentionally bad to force you to pony up for the BBU. I'd be happy with normal non-raid performance of 150/150.

    Read the article

  • Is a cluster the most cost effective redundancy method for windows server 2003?

    - by Ryan
    We had a server with bad ram which caused a long outage while they figured it out and our client facing apps had to go down for a while. We are coming up with a solution for instant fail-over but are not sure what the most cost effective method would be. Is a windows server cluster the best method for this? Also note we are using Parallels Virtuozzo if that makes any difference here. We found Parallels has a documented method for setting this up but it said it required a Domain Controller as well as a Fiber connection to shared storage, is all that really needed? Thanks.

    Read the article

  • Pros/cons to turning off cable modem

    - by Jay
    A little off the wall perhaps, but ... I have a cable modem and a router for a wireless home network. Is it a good or a bad idea to turn it off at night and during the day when we're all at work or school? Or should I leave it on 24/7. I was thinking that leaving it on constantly makes me more vulnerable to hackers, not to mention wasting electricity. (Though I'd guess the amount of electricity used by a cable modem and a router is probably pretty trivial. Still, every little bit helps.) When I have turned it off and turned it on again, it takes several minutes for it to go through its little dialog with the cable company and get me connected to the Internet again, which is annoying but not a big deal. Anyone know any good reasons one way or the other?

    Read the article

  • IIS 7 + ASP.NET 4

    - by user26712
    Hello, I have an ASP.NET application that I am trying to convert to an ASP.NET 4 application. The application is fairly simple. I have created a new web application in IIS 7.5 pointing to the directory that the ASP.NET application exists in. When I attempt to execute the application, but entering http://localhost%3A%5Bport%5D into my browser, I receive the following error: Error Summary HTTP Error 500.21 - Internal Server Error Handler "PageHandlerFactory-Integrated" has a bad module "ManagedPipelineHandler" in its module list Most likely causes: * Managed handler is used; however, ASP.NET is not installed or is not installed completely. * There is a typographical error in the configuration for the handler module list.

    Read the article

  • Does Dynamic DNS require separate subdomains?

    - by kce
    Hello. I have a functioning DHCP/DNS (ISC Bind 9.6, DHCP 3.1.1) server running on Debian that I would like to add DynamicDNS functionality to. I have a pretty simple question: Does DynamicDNS require (or recommend) separate sub-domains? I have seen a few tutorials where the the clients that are acquiring their IP addresses and other networking information via DHCP are on a different sub-domain as the servers which are statically configured (both in terms of IP, and DNS). For example: All the clients are on ws.example.org and the servers on example.org. Right now all of our servers and clients are in the same domain (example.org) but spread across different zone files (because we have multiple subnets). The clients are configured with DHCP and the servers are configured statically. If I want to setup DynamicDNS for the clients should I use a separate sub-domain? What's the best practice here (and why or why not would it be a bad idea to do otherwise)? Thanks.

    Read the article

  • Memcached failover

    - by user25164
    We have 2 memcached servers configured and use the Eniym client. When one of the server is down, it appears this server is added to the deadServers list (ServerPool.cs) and tries to resurrect the server every 10seconds (we have configured deadTimeOut to be 10seconds). Attempting to connect to the failed server causes a TCP timeout, the pages take a long time to load which results in bad user experience. 1) What is the standard way of resolving this issue? There are some posts about removing the server from the deadServers list. Is it okay to do this? 2) What is the recommended deadTimeOut setting (I understand by default it's 2 mins and we've changed it to 10seconds in our implementation) 3) Am I correct in my understanding that the cached data is not replicated across Server 1 and Server 2? If Server 1 is down, then it goes to the database to fetch these values (and it doesn't really check Server2)? Any help is really appreciated.

    Read the article

  • How to install Oracle's exp / imp binaries

    - by Sotapanna
    I was browsing the net to find a solution, but I had no luck. :( How can I install Oracle's imp / exp binaries to my CentOS box without installing an Oracle Server? I'm accessing an existing Oracle database on another machine, so I only need a client here. I already installed sqlplus and its dependencies (oracle-instantclient11.2-basic-11.2.0.3.0-1.x86_64.rpm, and oracle-instantclient11.2-odbc-11.2.0.3.0-1.x86_64.rpm), but that didn't provide those tools. Do I really need to fully install Oracle? My aim is to suck a database dump from the Oracle to my CentOS box. If there is another way to do this without using exp I wouldn't feel bad about NOT using it, but it seems to me there is no alternative.

    Read the article

  • I want to install an MSI twice

    - by don.vince
    I have a peculiar wish to install an msi twice on a machine. The purpose of the double install is to first install under the pre-production folder, run the deployment in a safe environment prior to deploying in the production folder. We typically use separate machines to represent these different environments however in this case I need to use the same box. The two scenarios I get are as follows: I've installed pre-production, I'm happy, I want to install production, I run the msi, it asks whether I want to repair or remove the installation I've production installed, I want to install the new version of the msi, it tells me I already have a version of the product installed and I must first un-install the current version The first scenario isn't too bad as we can at that point sensibly un-install and re-install under the production folder, but the second scenario is a pain as we don't want to un-install the live production deployment. Is there a setting I can give to msiexec that will allow this? Is there a more suitable different approach I could use?

    Read the article

  • Does anyone use the L-Track trackball?

    - by thethinman
    I've been using the Logitech Trackman Marble Mouse for years. Now I'm looking for a trackball with a scroll wheel, larger and heavier ball, and preferably rollers instead of pins. It must be finger (not thumb) operated. The Kensington Expert Mouse is close, but from what I've read the scroll wheel is poorly implemented. They also switched from rollers to pins. I bought a Kensington Orbit Trackball and it's not bad but the scroll wheel is rough and the ball is the same as the marble mouse. I'm still looking for something better. I found the L-Trac and it looks good but there's little info on the web. Has anyone used it and can provide their impressions? Or can you point out another option?

    Read the article

  • installing plugins in vim

    - by Ying
    I had a real difficulty installing vim plugins on mac. I know for a plugin to be installed it has to be seen by runtimepath. According to vim docs, this is Macintosh: "$VIM:vimfiles, $VIMRUNTIME, $VIM:vimfiles:after" First off, what does $VIM:vimfiles mean?? It is certainly not $VIM/vimfiles, as (~/.vim/vimfiles) does not work. So my options are 1) put my plugin in /usr/share/vim/vim73 (my $VIMRUNTIME, which i know to be BAD) 2) find a way to make vim look at my home directory when including plugins. I have putting my plugin in: ~/.vim/plugin.vim ~/.vim/vimfiles/plugin.vim ~/.vim/plugin/plugin.vim All to no avail. Can somebody help me out here? Thanks!

    Read the article

  • Horrible performing RAID

    - by Philip
    I have a small GlusterFS Cluster with two storage servers providing a replicated volume. Each server has 2 SAS disks for the OS and logs and 22 SATA disks for the actual data striped together as a RAID10 using MegaRAID SAS 9280-4i4e with this configuration: http://pastebin.com/2xj4401J Connected to this cluster are a few other servers with the native client running nginx to serve files stored on it in the order of 3-10MB. Right now a storage server has a outgoing bandwith of 300Mbit/s and the busy rate of the raid array is at 30-40%. There are also strange side-effects: Sometimes the io-latency skyrockets and there is no access possible on the raid for 10 seconds. The file system used is xfs and it has been tuned to match the raid stripe size. Does anyone have an idea what could be the reason for such a bad performing array? 22 Disks in a RAID10 should deliver way more throughput.

    Read the article

  • How to format HDD with MHDD Tool.

    - by PP
    I have copied MHDD application on a bootable CD and i wan to use it to low level format my HDD which had some bad sectors on it. so how can mark these badsecotrs using MHDD application? Pluse can i partition/format my HDD using MHDD because after low level format i want to setup window xp on that drive. can some one explain me step by step procedure to how to lovel format driven and partition/format it and then setup xp on that. Link to MHDD app. http://hddguru.com/content/en/software/2005.10.02-MHDD/ Thanks, PP.

    Read the article

  • is it worth defragging an iPod

    - by alimack
    Essentially my 5G iPod was cutting tracks off and generally misbehaving. So I did the following: 1) Use Diskwarrior - heavy directory fragmentation which it fixed; 2) Use iDefrag - some fragmentation but it kept halting as it couldn't move files; 3) Try to write out drive with Disk Utility - got a warning from DU so gave up before I started; 4) Completely restored using iTunes; 5) Reran Diskwarrior - still heavy directory fragmentation; 6) Reran iDefrag, still fragmentation although limited to two bands; The iPod is much quicker to traverse menus and no more track skipping. My question is this - is defragging worth it or does the heat generated by the process kill the drive and make it a self-defeating process. Anyone have any metrics/ figures? Clearly it's a bad idea for solid state drives like the nano & touch.

    Read the article

  • How do you test if a PCI slot is working - motherboard diagnostic software?

    - by Adam
    I suspect that my motherboard is bad, and that my PCI-E slot is not working. I have a Gigabyte motherboard that had a Hauppage 2250 working fine in the PCI-E slot for many months. However - recently - would not get TV channels that I used to get --- until finally Media center cant detect any channels at all. Got a new tuner card (ATI 350 card) - cant detect any channels. The cable signal is good - I have it connected directly to a TV - all channels come in clear. SO that is why I suspect a motherboard problem. Are there any motherboard diagnostic scanner utilities that can verify I need a new motherboard

    Read the article

  • Partition error being reported by only some programs

    - by Mahmoud20070
    I am getting an error about the partitions on one of my drives. I checked my hard disk with Acronis Disk Director 11 Home, HDTunePro, MiniTool Partition Wizard, Hddscan, HDD Regenerator 2011, WD Data Lifeguard Diagnostics, GParted, and Parted Magic, but none of them found any problems with the drive. However, when I perform a check with PartitionGuru, DiskGenius, and an old version of Partition Magic (8.0), I get the following error (seen below) about the partition being bad and needing to reformat to fix it: Error - The number of sectors record in DBR overflow - Reformat Should I reforamt or not? I checked two other hard drives, one of them like my 500 WD Blue, and neither has any errors with any progras. Are there any other programs that can check partitions for errors? Finally, how should a large SATA hard disk be formatted? I currently use programs like GParted, Acronis Disk Director, or MiniTool Partition Wizard. Are these programs good for formatting a new hard disk? Could the choice of format tool be the cause of the error?

    Read the article

  • Database modularity with EBS volumes

    - by Eclyps19
    I would like to add modularity to my websites on EC2 instances by encapsulating the site files and the mysql files in their own EBS volumes. The end result that I'm going for is the ability to quickly mount a volume or two to different servers running the same AMI (for testing/development/emergency maintenance, etc), as well as maintain separate snapshots of each. I'm able to do this fairly easily with a single database by symlinking my mounted database EBS to the appropriate places (/var/lib/mysql, /etc/my.cnf, /var/log/mysqld.log), but I'm not sure if it would even be possible be possible to have multiple databases on different EBS volumes running concurrently. Example: /website1/www.website.com /database1/ /website2/www.otherwebsite.com /database2/ Could anybody shed some light on this for me? Is it possible? Is it a bad idea? Thanks.

    Read the article

  • How to move Mailboxes over from old Exchange 2007 to new EBS 2008 network?

    - by Qwerty
    This q is similar to: http://serverfault.com/questions/39070/how-to-move-exchange-2003-mailbox-or-store-from-2003-to-2007-on-separate-networks Basically I am trying to move our exchange mailboxes over to a test domain that is hosting EBS2008 with Exchange 2007. We plan to move as soon as we can when we have our exchange data over. I have tried moving a db with mailboxes over but cannot get it to mount in the new Exchange in any way possible, including mounting it onto a recovery store. From my understanding the ONLY prerequisite for moving Exchange DBs across is that it must have the same Organizational name (unlike previous versions of Exchange). If anyone has any insight as to why I cannot mount and simply reattach the mailboxes, please give me an idea as to what could be wrong. It should be as simple as this. Note that the DBs I have are in a clean state. I cannot use ExMerge because I am not running any mailboxes on 2003. I have also tried using a 32bit Vista machine with the Export-Mailbox cmdlet to extract mailboxes but anything I do to it results in Permission errors. I have tried to troubleshoot these with no success. I am running in full admin with proper exchange roles and yet it still gives me access denied errors: Export-Mailbox : MapiExceptionNetworkError: Unable to make admin interface conn ection to server. (hr=0x80040115, ec=-2147221227) Also some errors show in the management console: get-MailboxDatabase Completed Warning: ERROR: Could not connect to the Microsoft Exchange Information Store service on server TATOOINE.baytech.local. One of the following problems may be occurring: 1- The Microsoft Exchange Information Store service is not running. 2- There is no network connectivity to server TATOOINE.baytech.local. 3- You do not have sufficient permissions to perform this command. The following permissions are required to perform this command: Exchange View-Only Administrator and local administrators group for the target server. 4- Credentials have been cached for an unpriviledged user. Try removing the entry for this server from Stored User Names and Passwords. Why I have to use a 32bit machine to export a simple .pst file is beyond me... So yeah I am now out of ideas and any help would be great! Thanks in advance.

    Read the article

  • Writing my first bash script (line 2: testscript: No such file or directory)

    - by skyeagle
    I am attempting to write my first shell script (I'm running Ubuntu 10.x) This is what my 'script' looks like cd /path/to/directory/cronscripts . testscript I put the following debug statements (after the cd), in the script above to make sure everything was ok: pwd echo `ls -lhrt` and it displayed all the files in the directory. So I dont understand the error message I am getting. I have the following two questions Why am I getting the above error message and how do I fix it? My script seems to run fine without a shebang in the first line - infact when I try either of the following: #!/usr/bin/bash #!/usr/bin/bash env I get the following error message: /usr/bin/bash: bad interpreter: No such file or directory What is the purpose of the shebang statement and why do I get the above error when I try to include it in my script/ how do I fix it, so I can add the shebang (is it necessary? - it seems most script examples have the shebang - but I'm not sure if it is a must, as the script runs without it).

    Read the article

  • Does sending e-mail in the name of customers increase the risk of being marked as spammer?

    - by Adrian Grigore
    Hi, We are developing a SaaS website application that lets users send invoices to their clients. Ideally, these e-mails should appear to be originating from our customers, so the sender e-mail address domain will not match the reverse IP entry for our server. In effect we would be forging their e-mail address, but of course with their consent. Will that result in a higher probability of being marked as a spammer / their e-mails being marked as spam? If yes, how bad is the penalty? And what about people who have an e-mail address originating form an SPF-enabled domain? I guess it should be the majority of the big e-mail providers.

    Read the article

  • Blacklisting: IP's or domains?

    - by johnnietheblack
    I am implementing a blacklisting system on my website that monitors contact forms for suspicious usage (both spam content and excessive frequency). When I find somebody / robot that meets my criteria for blacklisting, I want to send them to my DB as a blacklisted entity. My question is, should I blacklist them as an IP or as a domain? As far as I can see, blacklisting an IP is going to be far more effective, because I allow people to enter their email address in the form, and they can easily just change their domain on a regular basis. However, the downside is that if I blacklist an IP, I could potentially be blacklisting a large group of people who share an IP, when only one person is bad (ie - college campuses, coffee shops, etc). Is there a solution I'm missing?

    Read the article

  • PS/2 vs USB keyboards: performance and energy consumption

    - by Mister Smith
    As far as I know, PS/2 keyboards are interrupt driven, while USB are polled. Typically a PS/2 keyboard was assigned IRQ_1 on Windows. I'm no hardware expert, but at a first glance seems like the PS/2 keyboards are more efficient. So here are my questions: On modern day computers, are PS/2 keyboard better (or faster), and if so, would it be noticeable at all? (e.g.: in gaming) Since they don't need polling, do PS/2 keyboards save energy compared to USB? (notice I'm not talking only about the peripheral here, but about the overall computer energy consumption). In case PS/2 had any advantage over USB, would adding a PS/2 adapter to my USB keyboard make the device as good as an actual PS/2 keyboard? Conversely, would adding a USB adapter to a PS/2 make it as bad as a USB KB? Thanks in advance.

    Read the article

  • How can I set the fan speed to 100% on a laptop?

    - by Codemonkey
    I recently purchased a HP Pavilion dv7-4031. When it's cool, it works smoothly and efficiently. However when CPU and GUP temperatures reach 60c and above, the PC starts freezing up and stuttering. I can hear that the fan speeds steadily increase all the way up to 70-80c. This is what pisses me off: I want the fan speeds to run 100% all the time, perhaps preventing the high temperatures in the first place. The way it is now, fan speeds only increase to keep internal temperatures at above 60c. I've searched all over for any sort of speed control, finding nothing. Any help appreciated. I have tried Speedfan. In "Fans" there is nothing listed. I took that as a bad sign. The BIOS is pathetic, and only has 4 or 5 changeable settings, including "Quickstart" and "Boot order"

    Read the article

  • Windows - Delayed Write Failed error on USB hard drive

    - by ndngrd
    I've got a new Verbatim 1.5TB USB hard drive (Samsung HD154UI) and I'm finding myself completely unable to fill it. I'm using Windows XP. Whenever I try to copy a load of files over, it works for some time (will copy over between 20 and 90GB) but eventually stops with an error saying "The specified path is too deep" - the specified is not too deep, there's nothing more than 2 dirs deep that I'm copying. A balloon pops up at the bottom saying "Windows - Delayed Write Failed" telling me the data could not be copied. This wouldn't be too bad if I could just restart the transfer, but after this error has happened I can't write anything else to the disk - including if I eject it and then connect it to another machine. It just seems completely locked. The only way I can unlock it is to delete everything that I was copying to it. I've tried various USB cables and copying from different machines, and the same thing keeps happening.

    Read the article

  • What could cause Vista machines to degrade in performance?

    - by gencha
    We have installed several machines running Windows Vista at a client site. These machines run the same set of applications each day and are not operated with by any user. They mainly run a 3D application for digital signage purposes. After a few months I noticed that several machines suffer from really bad stuttering. The performance is unacceptable. At first I figured it was faulty hardware. But after putting a fresh image on them, they run perfectly fine again. These are Core i7 machines with GTX 280 graphics adapters. They should be able to handle the workload without any trouble (and they do, at least for a while). I really have no idea how to track down the source of this issue. Especially since not all of the machines display the same issue.

    Read the article

  • What ports does Advantage Database Server need?

    - by asherber
    I have an application which uses ADS and I am attempting to deploy it in a Windows network environment with a rather restrictive firewall. I am having a problem configuring firewall ports appropriately. ADS lives on \\server, and it's listening on port 1234. When \\client tries to connect to \\server\tables, I get Error 6420 (Discovery process failed). When \client tries to connect to \\server:1234\tables, I get error 6097, bad IP address specified in the connection path. \\server is pingable from \\client, and I can telnet to \server:1234. If I try to connect from a client machine inside the firewall, either connection path works fine. It seems there must be something else I need to open in the firewall. Any ideas? Thanks, Aaron. Edit: I should have specified that the firewall is open to \\server:1234 specifically for TCP traffic. Is UDP involved here in some way?

    Read the article

< Previous Page | 191 192 193 194 195 196 197 198 199 200 201 202  | Next Page >