Search Results

Search found 11553 results on 463 pages for 'bad programmer'.

Page 264/463 | < Previous Page | 260 261 262 263 264 265 266 267 268 269 270 271  | Next Page >

  • What can I do to configure my SSH such that it pauses for a flaky connection?

    - by kfmfe04
    I run ssh from OSX to a Ubuntu box - under home WIFI, everything works perfectly. However, when I ssh from certain external networks, ssh may be working fine for a minute or two and then drop due to a flaky network (diagnosed by pinging 8.8.8.8). How can I set up ssh so that it simply waits/pauses rather than drop (when dropping, I have to setup my development environment from scratch - a big hassle). EDIT When I say a bad connection, I mean, for example, ping 8.8.8.8 will work fine for a while, and then fail for 40-50 pings, and then come back on again.

    Read the article

  • Is there a way to import a scheduled task from windows 2003 (.job) to windows 2008 (.xml)?

    - by Rodrigo
    I had some jobs to be moved from the old production server (windows 2003 server standard) to the new machine (windows 2008 server standard), but the new server is unable to read the old .job format, also the import wizard only imports from .xml job files (same version). Obviously I don't want to rebuild all the jobs by hand, but can't find a tool that makes the process a very little easier. I don't trust in Microsoft for this kind of tools, my previously experiences had been to bad (DTS - SSIS). Any ideas? Thanks in advance.

    Read the article

  • Does NTFS performance degrade significantly in volumes larger than five or six TB?

    - by Josh Yeager
    One of my customers is planning to set up a new document store, which will probably grow by 1-2TB per year. One of my co-workers says that Windows performance is extremely bad if it has a single NTFS volume that is bigger than five or six TB. He thinks that we need to set up their system with multiple volumes so that no single volume will exceed that limit. Is this a real problem? Does Windows or NTFS slow down when the volume size reaches several terabytes? Or is it possible to create a single volume of 10 or more TB?

    Read the article

  • How do I find broken symlinks automatically on Windows?

    - by HughG
    Not sure if this is bad style, but I'm asking this question here because I couldn't find the answer elsewhere, and then I worked out one solution on my own. I'd be interested to see other people's solutions, but after a few days I'll post my own. In my specific case, I'm running on Windows 7, but I'd be interested in answers for other/older versions of Windows. I realise that one answer is "install a version of Unix find, and then solve as for Unix", but I wanted a more "native" solution. EDIT 2012-07-17: Clarification: by "automatically" I ideally mean something I can run as part of a script, rather than a GUI tool which does all the work at the press of a button, because I'll want to do this unattended.

    Read the article

  • Client PC not booting when certain TFT plugged in - TFT or graphics card failure?

    - by Chake
    here comes something quite strange: On a client machine (DELL Vostro 420) we experienced problems when booting: when turning on the machine beeps normal but doesn't display anything and doesn't boot. After some testing I found out, that this only happens if one (of the two) monitors (Iiyama ProLite E2472HDD) is plugged in while booting. If the other monitor (TFT 2) is plugged in everything is fine. Here a small illustration, TFT 1 is the bad guy: TFT 1 | TFT 2 | failure x | x | x x | | x | x | After BIOS-Phase I can safely plug in TFT 1 and everything works just fine. The question is, what can be done to avoide this behavior: Change monitor? (Iiyama ProLite E2472HDD) Change graphics card? (GeForce 9800 GT) Other suggestions?

    Read the article

  • Exchange Server 2003 Drive Error

    - by Saif Khan
    My exchange box went offline this AM. The logs show The device, \Device\Harddisk1, has a bad block. I tried running chkdsk D: /x /r but that didn't help either. I went to the Exchange Manager and tried to move the .stm file from D: to C: but got Exchange System Manager The database files in this store are corrupted. ID no: c104173b Exchange System Manager What else is there to try just to get this online for a few hours until I can get onsite to change the drive. This is a Windows Server 2003 server.

    Read the article

  • How to accelerate and notice failure of potentially faulty disks

    - by rainier
    Hey, I got a bunch of 'used' servers, whose disks should have been checked, but they have been shipped around the county in crate which can't help. I just had one disk go bad (despite being mirrored, currently trying to get more details). The server was fine for about a week before everything ground to a halt this afternoon. Is there any way 'accelerate' the failure of faulty disks, with the goal of bringing the disk to failure before we launch production services? Would doing lots of I/O with 'dd' or 'iozone' be a good way to test these potentially faulty disks? Any other tests/tools that would help recognized failures before they happen?

    Read the article

  • BES 5.0 and MAPI calls to exchange system

    - by nysingh
    We have been using BES 4.1(5) for a while now and it has been a resource hog on exchange due to high number of MAPI calls. I have heard that BES 5.0 is even worse. the comparison i heard is that BES 4.1 is makes MAPI calls equal to 5 outlook clients per BB user and BES 5.0 makes MAPI calls equal to 10 outlook clients per BB user. can someone confirm if it is true? is BES 5.0 is really that bad in MAPI calls and for exchange performance. ? thanks

    Read the article

  • How can I make my browser(s) finish AJAX requests instead of stopping them when I switch to another page?

    - by Tom Wijsman
    I usually need to deal with things on a page right before switching to yet another page, this ranges from "liking / upvoting a comment or post" up to "an important action" and doesn't always come with feedback on whether the action actually proceeded. This is a huge problem! I assume the action to proceed once I start the particular AJAX request, but because I switch to another page it didn't actually happen because the AJAX request got aborted. This has left me several times with coming back to the page and seeing my action didn't take place at all; to give you an idea how bad this is, this even happened once when commenting on Super User! Is there a way to tell my browser to not drop these AJAX connections but simply let them finish?

    Read the article

  • Virtualbox crashes quite often in Windows 8

    - by user1776158
    I just installed Windows 8 on my computer. I got the ISO and the product key from my university so the software itself is sound. I use alot of virtual box. And ever since I moved to Windows 8, I have noticed that virtual box crashes more often. In particular, it is very very bad at opening multiple guests. My CPU usage will be at like 20% and I only have 3 guests open and my entire computer just freezes. Cursor and all. In Windows 7, I was able to open like 6 (not that I ever needed to) and really push my CPU. I havent experience any other issues with Windows 8 yet. Has anyone encountered this? Thanks!

    Read the article

  • How to I test if mod_rewrite is enabled?

    - by user124130
    I'm setting up an environment for wordpress on apache2, on a fresh install of ubuntu 12.04. In order to get friendly URLS working, I'm trying to set up mod_rewrite. I followed some instructions I found on the net, and used a2enmod. Now. after restarting apache, I'd like to check if the module is actually loaded. The command that I've found for getting a list of loaded modules is this: apache2 -t -D DUMP_MODULES However, this returns an error: apache2: bad user name ${APACHE_RUN_USER} So, how do I actually list all loaded modules, or otherwise check to see if mod_rewrite has been enabled?

    Read the article

  • Anyone have any experience with bargain laptop batteries?

    - by chris
    I've got an oldish D820 that's got a 100% dead battery. I know that I could, in theory, take it apart and replace bad cells in the battery. I'm not really comfortable with doing that. I also know that there are various places that sell replacement batteries for 20% to 80% of the cost that Dell would charge. Does anyone have any experiences with buying more than a couple of these off-brand batteries? If a battery goes boom, it could be really ugly, so I'd rather not risk it, but at the same time, the dell batteries are really expensive... Any opinions on these ebay / off-brand battery vendors? Thanks!

    Read the article

  • Big Excel File Freezing/Running Slowly

    - by ktm5124
    Hi, My co-worker has a very large Excel file (over 7 MB) that suffers from the problems of (A) running slowly (B) taking forever to open/save/close and (C) freezing the computer, requiring a restart. I set the calculations to Manual, and I repaired the file, but the file didn't change in file size and it is still having these problems. My questions are: (1) Is there any way around this problem or is Excel just bad at handling ~7MB files? (2) Would upgrading RAM make a big difference? (3) It's possible that we can't afford to spend the money on a RAM upgrade. Are there are any other ways of mitigating the problem? Thanks.

    Read the article

  • Why would Linux VM in vSphere ESXi 5.5 show dramatically increased disk i/o latency?

    - by mhucka
    I'm stumped and I hope someone else will recognize the symptoms of this problem. Hardware: new Dell T110 II, dual-core Pentium G860 2.9 GHz, onboard SATA controller, one new 500 GB 7200 RPM cabled hard drive inside the box, other drives inside but not mounted yet. No RAID. Software: fresh CentOS 6.5 virtual machine under VMware ESXi 5.5.0 (build 174 + vSphere Client). 2.5 GB RAM allocated. The disk is how CentOS offered to set it up, namely as a volume inside an LVM Volume Group, except that I skipped having a separate /home and simply have / and /boot. CentOS is patched up, ESXi patched up, latest VMware tools installed in the VM. No users on the system, no services running, no files on the disk but the OS installation. I'm interacting with the VM via the VM virtual console in vSphere Client. Before going further, I wanted to check that I configured things more or less reasonably. I ran the following command as root in a shell on the VM: for i in 1 2 3 4 5 6 7 8 9 10; do dd if=/dev/zero of=/test.img bs=8k count=256k conv=fdatasync done I.e., just repeat the dd command 10 times, which results in printing the transfer rate each time. The results are disturbing. It starts off well: 262144+0 records in 262144+0 records out 2147483648 bytes (2.1 GB) copied, 20.451 s, 105 MB/s 262144+0 records in 262144+0 records out 2147483648 bytes (2.1 GB) copied, 20.4202 s, 105 MB/s ... but after 7-8 of these, it then prints 262144+0 records in 262144+0 records out 2147483648 bytes (2.1 GG) copied, 82.9779 s, 25.9 MB/s 262144+0 records in 262144+0 records out 2147483648 bytes (2.1 GB) copied, 84.0396 s, 25.6 MB/s 262144+0 records in 262144+0 records out 2147483648 bytes (2.1 GB) copied, 103.42 s, 20.8 MB/s If I wait a significant amount of time, say 30-45 minutes, and run it again, it again goes back to 105 MB/s, and after several rounds (sometimes a few, sometimes 10+), it drops to ~20-25 MB/s again. Plotting the disk latency in vSphere's interface, it shows periods of high disk latency hitting 1.2-1.5 seconds during the times that dd reports the low throughput. (And yes, things get pretty unresponsive while that's happening.) What could be causing this? I'm comfortable that it is not due to the disk failing, because I also had configured two other disks as an additional volume in the same system. At first I thought I did something wrong with that volume, but after commenting the volume out from /etc/fstab and rebooting, and trying the tests on / as shown above, it became clear that the problem is elsewhere. It is probably an ESXi configuration problem, but I'm not very experienced with ESXi. It's probably something stupid, but after trying to figure this out for many hours over multiple days, I can't find the problem, so I hope someone can point me in the right direction. (P.S.: yes, I know this hardware combo won't win any speed awards as a server, and I have reasons for using this low-end hardware and running a single VM, but I think that's besides the point for this question [unless it's actually a hardware problem].) ADDENDUM #1: Reading other answers such as this one made me try adding oflag=direct to dd. However, it makes no difference in the pattern of results: initially the numbers are higher for many rounds, then they drop to 20-25 MB/s. (The initial absolute numbers are in the 50 MB/s range.) ADDENDUM #2: Adding sync ; echo 3 > /proc/sys/vm/drop_caches into the loop does not make a difference at all. ADDENDUM #3: To take out further variables, I now run dd such that the file it creates is larger than the amount of RAM on the system. The new command is dd if=/dev/zero of=/test.img bs=16k count=256k conv=fdatasync oflag=direct. Initial throughput numbers with this version of the command are ~50 MB/s. They drop to 20-25 MB/s when things go south. ADDENDUM #4: Here is the output of iostat -d -m -x 1 running in another terminal window while performance is "good" and then again when it's "bad". (While this is going on, I'm running dd if=/dev/zero of=/test.img bs=16k count=256k conv=fdatasync oflag=direct.) First, when things are "good", it shows this: When things go "bad", iostat -d -m -x 1 shows this:

    Read the article

  • "shell not found" on start of backtrack 5

    - by Jor-el
    I am using backtrack5 in VMware on my system. I was trying to change my default shell to ksh. After executing the 'chsh' command. On restarting the OS, it shows "Cannot execute ksh: no such file or directory", which apparently because of giving a bad path while changing shell. Because of this I am not able to start the OS, as I have no terminal to interact with (cannot start the GUI). Now I need to restore the previous shell, but I don't have any access to the terminal. Thanks for the help.

    Read the article

  • WebsitePanel 2 totally NOT working on Windows Server 2012 on Azure

    - by Carmine Giangregorio
    I’m having many troubles installing WebSitePanel on an Azure Virtual Machine, with Windows Server 2012. I followed the steps in http://www.websitepanel.net/documentation/deployment-guide/server-configuration/preparing-windows-server-2008-r2-for-websitepanel-installation/ and installed everything I needed. Then, I installed the WebSitePanel Standalone Server package with the installer. I opened the endpoint for the port 9002 on Windows Azure; so I pointed my browser to myhostname.cloudapp.net (note: in Azure you don’t have a static IP address, instead you have an hostname like [hostname].cloudapp.net). So, loading myhostname.cloudapp.net:9002 fails, and any browser shows something like “Unable to load page”. Notice: if I try to load the WebSitePanel Portal directly on the server, I get an error HTTP 400 Bad Request. How come? IIS works perfectly on the server, in fact the default website runs without problems on port 80.

    Read the article

  • Open Directory authenticated bind succeeds, but creates incomplete record

    - by Jay Thompson
    I have about a dozen Macs running 10.6.7 or 10.6.8, which are all failing to bind properly to my new 10.7.4 Server OD. I can bind them just fine via Directory Utility or dsconfigldap, and it reports success. However, when I look at the record, it is failing to write the MAC address. Even if I manually update the record with the MAC address, MCX doesn't do anything and clients can't log in to OD accounts. All of the affected clients have hundreds of lines in the /Library/Logs/DirectoryService.error.log like so: 2012-09-15 22:23:18 EDT - T[0x00007FFF70292CC0] - GetMACAddress returned 0x *** bad control string *** 8x I do know that all of these clients were previously managed with the Guest computer account, and I also know that they were all imaged with a DeployStudio image when they were purchased. I've tried dscacheutil -flushcache, but after that I'm drawing a blank. Google has a few hits, but nothing very helpful. Re-imaging would be ideal but probably isn't going to happen. Anyone come across this before?

    Read the article

  • Log with iptalbes which user is delivering email to port 25

    - by Maus
    Because we got blacklisted on CBL I set up the following firewall rules with iptables: #!/bin/bash iptables -A OUTPUT -d 127.0.0.1 -p tcp -m tcp --dport 25 -j ACCEPT iptables -A OUTPUT -p tcp -m tcp --dport 25 -m owner --gid-owner mail -j ACCEPT iptables -A OUTPUT -p tcp -m tcp --dport 25 -m owner --uid-owner root -j ACCEPT iptables -A OUTPUT -p tcp -m tcp --dport 25 -m owner --uid-owner Debian-exim -j ACCEPT iptables -A OUTPUT -p tcp -m limit --limit 15/minute -m tcp --dport 25 -j LOG --log-prefix "LOCAL_DROPPED_SPAM" iptables -A OUTPUT -p tcp -m tcp --dport 25 -j REJECT --reject-with icmp-port-unreachable I'm not able to connect to port 25 from localhost with another user than root or a mail group member - So it seems to work. Still some questions remain: How effective do you rate this rule-set to prevent spam coming from bad PHP-Scripts hosted on the server? Is there a way to block port 25 and 587 within the same statement? Is the usage of /usr/sbin/sendmail also limited or blocked by this rule-set? Is there a way to log the username of all other attempts which try to deliver stuff to port 25?

    Read the article

  • Office 2010 & Windows 7 - 'File' is currently in use. Try again later

    - by thing2k
    The issue: when saving a document from either Word, Excel or PowerPoint 2010, every so often it will show the message 'file' is currently in use. Try again later. We started our rollout of Windows 7 beginning of this year, and this issue was infrequent, but is now affecting enough people to be a problem. Usually, if you clear the alert, then save again, it works fine, though only in Word or Excel. Annoyingly, PowerPoint has a bad habit of changing the file to read-only after the error. So the only choice is to save to a new file name. The issue seems to only happen to files in the user's My Documents, which is a redirection folder from their HomeShare. The HomeShares are spread across 3 different file servers, 2 Windows 2003 and 1 Windows 2008 R2. Has anyone seen this issue and know how to fix it?

    Read the article

  • Windows 7: from Geforce 8800 to three monitors?

    - by lance
    I've got a GeForce 8800 that I've quite happy with. It drives my two 23" widescreen displays well. Now I've got a 19" standard display that I want to stick between the two widescreens. My second PCIe 16x slot is unused (as is the PCI slot below that), and I want to add a card to my Win7 x64 system. This 19" display won't be used for gaming, so I don't need anything fancy. Here are two cards I was considering, but I'm wondering if they're bad choices for some reason? If they're both fine choices, which is better and why? Again, I'm needing to power only the 19" standard display with this card, and it won't play games. I just need 1280x1024 in Win7 x64. NVidia: Galaxy 95TFE8HUFEXX GeForce 9500 GT Video Card - 512MB DDR2, PCI Express 2.0 ATI: ASUS EAH4350 SILENT/DI/51 Radeon HD 4350 Video Card - 512MB DDR2, PCI Express 2.0

    Read the article

  • Zfs Drive config on FreeNas

    - by Martyn
    Couple of related questions. Background: I have a stock of drives lying about that I want to use in a HP Proliant Microserver which has a 4 slot SAS cage with FreeNas. I don't want to spend any more on parts. 1) With Zfs and RAIDZ, how bad would it be to mix 2 x 1TB Seagate Barracudas with 2 x 1TB WD Green drives. The latter Green drives are I believe only 5,400 RPM, but I can't find that info for sure. 2) Which would be better, the above mis-match in a RAIDZ or 2 x 3TB Seagate Barracuda in a mirror. Considering both performance and data security 3) I have a spare 250GB drive and a spare SATA slot (for optical drive, not sure on speed) as well, would that be a performance gain to use that for the ZIL and L2ARC (over just using the main drives). Thanks in advance.

    Read the article

  • Can a processor upgrade cause BSOD?

    - by Daniel
    I Had the following hardware setup: Phenom II X4 945 Asus M4A97 4GB DDR2 OCZ Radeon HD5850 OCZ Agility 2 120 GB Windows 7 x64 Pro Fully updated (latest drivers and windows update patches) Then I bought an used Phenom II X6 1090T and installed it without formatting. Since that my computer started BSODing almost every time I'm playing any game and with different error messages, like: page fault in non-paged area bad pool header the video memory manager found a problem error in dxgmm1.sys(or something like that) And when it doesn't BSOD the game simply crashes. I have tried: Updating BIOS Reseting BIOS to defaults Reinstalling Video Drivers Installing the latest DirectX All that's left is to do a full format and I don't want to do that since it's going to be a lot of work to fine tune windows to my preferences again. So is my "new" processor defective or do I really need to format the computer? Update: I use a (properly installed) custom cooler from Coolermaster and both the BIOS and the Open Hardware Monitor(Application) attest the CPU is not overheating, so I guess the CPU its defective and since I bought it from a guy over the Internet I'm probably screwed

    Read the article

  • Domain name in server other than the website

    - by med
    Hello! I'm not very popular with hosting, but I have a special situation: I live in Tunisia, I can buy a domain name with .tn extention, but the problem is that: 1- The domain could not be pointed to a server outside Tunisia 2- All servers in Tunisia are bad, no one provides really reliable hosting so, I want to use the .tn domain with a basic hosting in Tunisia, and Make other db queries and rich media on another remote server outside Tunisia. How to do it? Is there better alternatives? All suggestions are welcome :) Thank you.

    Read the article

  • Macbook Pro - Disk Locked HD needs repair, can't reinstall OSX

    - by Rob
    I basically have the same problem. A friends HD was acting badly. Ran Disk Repair Util many times. Says disk needs repair but it won't repair it. I re formatted and installed 10.8 but that hasn't fixed it. Ive tried partitioning it and that won't work. Now I'm trying to get system 10.8 to reinstall and it says the disk is locked. Your fix requires Terminal, is that right? How do I open terminal to run your fix? Please help if you can. Disk Warrior couldn't fix it. I think the disk is bad. What do you think?

    Read the article

  • Is allowing remote Sql Server Management Studio safe?

    - by dave thieben
    I administer a website that runs on IIS on one box, and SQL Server 2008 Workgroup on another box. typically I remote into the DB box and run SSMS to work on the db, but I would like to be able to access the db directly with SSMS on my local box. I've seen the other questions about allowing remote access to the database, but my question is, is this safe? I'm concerned that I'm opening a hole in the firewall and potential for hack attempts. Is this just a bad idea in general?

    Read the article

< Previous Page | 260 261 262 263 264 265 266 267 268 269 270 271  | Next Page >