Search Results

Search found 908 results on 37 pages for 'the worst shady'.

Page 16/37 | < Previous Page | 12 13 14 15 16 17 18 19 20 21 22 23  | Next Page >

  • Motherboard max memory supported

    - by Ashfame
    I have an Intel DG965RY motherboard and its specification says it supports 8GB with 533 or 667Mhz RAM sticks and only 4GB with 800Mhz RAM sticks. I am running a 64bit OS. I earlier had 2 X 1GB sticks (800Mhz), so I bought 2 X 2GB sticks (800Mhz) and I underclocked them in the settings to run at 667Mhz. Shouldn't it support all 6GB RAM now? It would be a bummer if I will specifically need 667Mhz sticks thinking that at the worst they will underclock and then run at 667Mhz. I tried this because I saw someone posted at some forum that he put in 4GB+ of RAM in the same board @ 800Mhz and the system uses it all. In my case (On Ubuntu), it only shows 3.2GB as of now (link to Question) so needed to confirm if this is a hardware limitation.

    Read the article

  • What configuration management solutions exist in a non-networked environment?

    - by Rob Spieldenner
    My servers exist in an environment without outside network connectivity (this is a requirement), so when I deploy updates all packages, binaries, config files, etc. must be included on the delivered media. And of course I want some sort of configuration management so I can tell what has and hasn't been installed. So I was wondering if people had experience with chef, puppet, or another configuration management type tool for dealing with this type of environment. Worst case I deploy my updates as an RPM. EDIT: My setup has both Linux servers and Windows servers.

    Read the article

  • Can a 32-bit RHEL4 userland work with a 64-bit kernel?

    - by James
    Is there a way to change an i386 RHEL4 machine to run an amd64 kernel, but ensure that it still builds software into same i386 binaries? On Debian this seems quite straightforward: just install an amd64 kernel (worst case, build one like this guy: http://www.debian-administration.org/users/jonesy/weblog/1) and prefix everything with "linux32". Then everything that considers uname -m will be unchanged, I just need to handle the few cases that consider uname -r. What is the Red Hat equivalent? Is the only way a full 64-bit installation on another disk and then chrooting back to the 32-bit system before anyone builds anything? (Even the best examples of that seem to be Debian-based.) Background: We make a large system that runs on (a variant of) i386 RHEL4. However, some of the larger RHEL build machines now have enough RAM that they might benefit from going 64-bit (for the kernel and maybe some of the bigger build steps). Our build system doesn't support cross-compilation.

    Read the article

  • NIC: Changing Link Speed & Duplex to Full Duplex Is Drastically Slower

    - by rism
    I have an old Win Xp box with a 100 PRO/100 VE NIC and connecting to a Win 7 box with gigabit NIC via an ASUS gigabit router and peer to peer. I was getting 76% network utilisation on the XP machine using AUTO as the value for Link Speed and Duplex which is fine. But I decided to see what would happen if i selected Full Duplex 100Mbps on the XP box. I thought at worst it would be as good as auto detect since the Win 7 box it's transferring to is faster. Instead network utilisation dropped to 0.12% (as in a 10th of a percent) and the transfers just timed out. No biggie to reverse back to Auto but my question is why is full duplex such a dog? Im not "stuck" but just dont understand. There are only 2 machines here connecting through a router so i cant see how that would cause alot of collisions etc (if thats the problem).

    Read the article

  • About the External Graphics Card and CPU usage

    - by Balaji
    Hi, We are Rendering 16 live Streams at our client machine through one of our applications and the resolution of the video streams are as 4CIF/MPEG4/25FPS/4000Kbits. The configuration fo the client machine is below. HP Desktop Machine: Microsoft Windows XP Intel (R) Core2 Duo CPU E8400 @ 3.00 GHz 2.99 GHz, 1.94 GB of RAM Intel (R) Q45/Q43 Series Express Chipset (Inbuild) The CPU usage of the machine peaks 99% for 16 straems. After some discussion, we had decided to install external graphics card to reduce the CPU usage. So that, we have tried following graphics cards. NVIDIA Quadro NVS 440 - 128 MB Radeon HD 4350 - 512 MB GDDR2 Redeon HD 4350 - 1GB DDR2 ASUS EAH 4350 Silent 1GB DDR2 But the performance wise no difference, even worst. So, what is the pupuse of these external graphics cards? Really it will reduce the CPU usage? What parameters have to check, if we want to reduce the CPU usage? Please do the needful as soon as possible. Regards Balaji

    Read the article

  • Integrated Graphics and Audio as Media Center?

    - by Will
    I'm considering setting up a PC as a Media Center. Mainly to watch movies (ideally HD quality) and listening to music, but also to perform tasks like e-mail, web browsing, ... I quite like the looks and the price of this barebone: http://www.asus.de/Barebone_PC/S_Series_7L/S2P8H61E However it comes with integrated graphics and audio and only has one free PCI-Express slot. Which would mean in the worst case, where both integrated graphics and audio turn out to be insufficient, I could only upgrade one. So is integrated graphics and audio sufficient for a media center solution? Cheers, Will

    Read the article

  • How to delete just one LINE of text (NOT a table-row!) with a single KEYBOARD shortcut in Microsoft Office Word 2010?

    - by Sk8erPeter
    Are there any shortcuts to delete just one row (which is NOT a table row, just a single row in a text) in Microsoft Office Word 2010? If not, how can I assign one to do it? In worst case, can I make a macro (in VB) which could do the same with a custom shortcut? To clarify my problem: I would like to avoid multiple clicks and/or pushing multiple buttons, even if I click in the middle of the line of text. :) For example, in Notepad++ I can delete the entire current line with Ctrl+L, in NetBeans, I can delete an entire line with Ctrl+E, in Eclipse, I can delete current line with Ctrl+D, etc., where it doesn't really matter where my mouse cursor is actually... so there are these simple solutions, which I look for in Word too. It really would simplify my work in huge documents.

    Read the article

  • MacPro will freeze after first startup and first time screen saver activates

    - by Rob Segal
    I've run into a bizarre situation every time my MacPro first starts up in the morning when I get to work and the screen saver activates for the first time. When I try to deactivate the screen saver hitting the keyboard or mouse nothing will happen. At best I will get one of the following... Able to move the mouse but clicking does nothing Login prompt will show up but typing on the keyboard does nothing At worst the screen will be fully locked. In all cases I will have to reboot my machine to get it to work again. After this initial restart the machine is fine for the rest of the day. Anyone encountered something like this before?

    Read the article

  • Performance issues with new dedicated server [closed]

    - by Pierre Espenan
    I have just subscribed to a new dedicated server and am getting worst than expected PHP execution performance. Execution times are twice as high as on my old mutualized server! I'm definitely not an expert at server management, so I'm wondering what I missed. Here are some stuff that can help you understand what's wrong here : My server (in french but easy to understand) : http://www.online.net/fr/serveur-dedie/dedibox-sc phpinfo(); output : http://jsfiddle.net/E8b7W/embedded/result/ PHP bench script (dedicated server) : http://jsfiddle.net/EhXzK/embedded/result/ PHP bench script (old mutualized) : http://jsfiddle.net/ANbWt/embedded/result/ Is it normal to get such poor performances after a kernel update and basics "apt-get install" for apache2 and php ? Thanks !

    Read the article

  • Funnelling http traffic

    - by spencer p
    I have a situation where a large batch of servers (X), on demand, need to request data from a smaller set of web servers (Y). The worst case scenario is if all servers in X decide to fetch different requests to one server in Y. That would be X amount of connections, which could be a very large burst of traffic. The best case scenario is if 1 server in X hit 1 server in Y in tandem. Life does not work like this. One idea to entertain is placing a proxy, similar to squid between X and Y. All of X servers can connect to this proxy, but would result in a few persistent (http keepalive) connections to Y. If The few were say, 3 or 4, then it would funnel. If we could then rate limit those connections and traffic decides to spike unusually high, we wouldn't hurt anyone but ourselves. Thoughts?

    Read the article

  • ID Card printers and software

    - by Chris
    My company of about 60 people currently uses a Magicard Pronto ID Card printer and the Magicard ID card software. The printer is would be OK (not great) if they released 64 bit drivers for it, but Magicard ID is without a doubt one of the worst pieces of software I have ever used, ever. It would be OK for occasional hiring, but my company is looking to add a lot of people rather quickly. Does anybody know of decent ID-card printing software, and possibly a replacement printer? I might just "accidentally" run over this one with my truck a few times.

    Read the article

  • Can I change the system's "Browse for Folder" dialog globally?

    - by Chris Phillips
    As far as I know, everyone hates the "Browse for Folder" dialog: This dialog is always too small, rarely remembers locations well, and worst of all: forces you to navigate your entire computer using a tedious tree structure. Now, to be fair, some of the problems are likely to do with how apps are invoking the control -- not setting a size or a default directory, etc. But the problem about the tedious tree control remains. Is there any way to customize your Windows installation to use a different control? Preferably an app/installer that does it for you safely, but dropping in a compatible DLL or similar technique would be okay too. Or are we stuck with this terrible control forever?

    Read the article

  • Upgrading from php 5.3 to php 5.4 with Macport

    - by dr.stonyhills
    PHP5.4 has been available for sometime now and Macport recently caught up with the release of port php54 but the process of upgrading is not as clear as possible. Even worst for those who are new to maintaining multiple versions of PHP on the same machine. I am keen on trying out some of the new features in PHP5.4 like traits, new array form etc but falling back on to php5.3 for other compatibility stuff. So i sudo port install php5+ (all the variants, apache2 etc) Then i tell it what PHP port to use as default sudo port select --set php php54 Check what version of PHP is active in the terminal using php -v outputs php 5.4.3. But i seem to be having issues with choosing the right non cli version as in the version of the module run by apache etc is still php5.3.12. Do i have to change the reference to the libphp5 in apache httpd.conf? Any advice on the right workflow for switching between php version on macport greatly appreciated!

    Read the article

  • Accessing a shared folder using other credentials with Windows 7

    - by Nicolas Buduroi
    I was at a client office trying to connect to a shared drive on their network but was greeted by a "you do not have permission to access" error! Couldn't find any way to enter the required credentials as this message didn't had any other options. Tried to map the drive, selected the option to enter the correct credentials (with \\HOST\user) but it wouldn't work at all. The worst thing in all of this is that my coworker who is using OS X has been able to connect to that drive without any problem, he clicked on it, entered the credentials and could open the folder! The folder is shared by a Windows Small Business Server 2008 machine.

    Read the article

  • Can't use command line – "command not found" after editing PATH

    - by MEM
    I'm running OS X Mavericks and was trying to install MAMP PRO 2.2. I was trying to configure the PATH variable to have the PHP binaries of MAMP PRO. I added the following line on my ~/.bash_profile file: export PATH=/Applications/MAMP PRO/bin/php/php5.5.3/bin:$PATH As you may notice, since I have MAMP PRO and not just MAMP, I've added a space. As a consequence, I know have the following error each time I run the terminal: -bash: export: `PRO/bin/php/php5.5.3/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/bin': not a valid identifier Worst: I can't get any command to run, like: ls, clear etc. I always get: "command not found" I don't even know the absolute path for ls. How can I make the commands work again, so that I can properly fix the path I was trying to setup on the .bash_profile file?

    Read the article

  • Best method(s) to backup VMs running on HyperV?

    - by Kara Marfia
    We're in the middle of P2V'ing most of the network, so the current backup method is likely the worst - the backup agent is still installed on the guest OSs, and the backup device is dutifully pulling them onto tape, one file at a time. I suspect there's a clever way to script (PowerShell?) a suspend on the VMs, then backup of the .vhd files, and unsuspend the VMs. This seems like it would provide big speed benefits, while losing file-level restore (might be best for things like DCs and app servers). What methods/policies have you hammered out?

    Read the article

  • Can I run Linux on the HP Mini from Verizon Wireless, with a built-in 3G modem, and connect using the built in modem?

    - by Parker
    I apologize if my question made little sense, I am very tired at the moment. Anyway, Recently I purchases the HP Mini from Verizon Wireless. It runs on the Intel Atom processor and came with Windows 7 Starter Edition, which is quite possibly the worst and most restrictive OS I have ever used. I plan on installing CRUX, after doing a little research on minimal, free, Linux operating systems, as I am working on web programming using Python, among other things. My only worry is that, because my HP Mini uses a built in wireless card (and I have no wifi connection, only a 3G connection) that has a "mobile access" connection to Verizon wireless, I am not sure if after installing CRUX that I would still be able to connect using this option. I have never used Linux before, and am hoping this will work. So if anybody has experience with this...please answer, can I connect to vzw 3G with linux, using an internal mobile access card...

    Read the article

  • Verifying Time Machine backups

    - by jtimberman
    I'm preparing my system for a Snow Leopard upgrade, and I prepare for the worst case scenario: full reinstall and restore. I would like to verify that my Time Machine backups are valid, and will restore correctly. My Time Machine backups go to a Linux server running Netatalk, and the backups complete successfully. How do I do a test restore to an alternative location, or otherwise verify my data without overwriting any existing files? Do I need to save anything in particular externally to make sure I can access the backups if I have to reinstall from scratch?

    Read the article

  • Use ImageMagick to convert TIFF to PNGs, how to improve the speed?

    - by Woo
    I am using "convert" from IM to get PNGs from multi-page TIFF files, everything is good except the speed. From "convert" documentation, I found: For the MNG and PNG image formats, the quality value sets the zlib compression level (quality / 10) and filter-type (quality % 10). For compression level 0, the Huffman-only strategy is used, which is fastest but not necessarily the worst compression. The default PNG compression is 75. So I tried "-quality 0", but almost no changes with the spreed. Anyone can share the ideas of how to improve the spreed? Here are my command: convert 100Pages.tif[0,1,2,3,4,5] -quality 0 100Pages.png Thanks!

    Read the article

  • Time difference between servers after disaster recovery

    - by Sandokan
    We are running an old training system based on Windows Server 2003 and XP-clients. The solution is rather simple with four servers, two of them beeing DC:s. Everything is preconfigured and that goes for backup scheme as well. The backup software is Symantec BackupExec 2010. The scheme is a standard GF-F-S routine with full backups running once a week on Sundays. The other six days run differential backups. Now let's say in a worst case scenario, a server crashes on Saturday and we have to restore it from backup. The last backup will then be six days old and thus it will come online with six days old configurations. Will this pose a problem for the other servers or will the recovered server "get in line" eventually?

    Read the article

  • What are possible causes of keyboard lag on my desktop machine?

    - by Jer
    I am running Windows 7 and began experiencing keyboard lag in most applications, and it seems to be getting worse. Certain websites are the worst - on some, I can type a sentence, take my hands off the keyboard, and watch the characters continue to appear on the screen for several seconds. Others are not as bad, but still noticeable and annoying. I just started noticing it in non-browser applications (e.g. Outlook) as well. I've disabled all extensions in Firefox, rebooted my machine, and that did nothing. There is nothing using much memory or cpu cycles, even when the lag is occurring. This is a machine at work with very strict controls over what can be installed, so the chances of any kind of malware are very slim. I don't believe anything as been installed since before the problem started. What could be causing this, and/or what can I do to debug?

    Read the article

  • Restoring file properties but not the complete files, from backup

    - by Jon
    While copying data from my old storage on a Linux computer to the new (linux-based) NAS, I accidentially failed with getting the properties (most important: the modify dates) along to the new location. I also continued to use/modify the files at the new location and hence, cannot just copy it all over again. What I would like to do is a diff between files in the old vs. the new storage, and for those being identical, restore the properties from Linux storage to the NAS storage files. Is there a clever way such as a script or a tool to do this? I could either run it on the Linux box or in worst case from a remote Windows computer. Grateful for any suggestions. /Jon

    Read the article

  • Why does my CPU Usage reach 100% too often?

    - by deathlock
    I'm using a dual-core processor and often see my CPU usage reaches 100%. I realize this may happen if I'm running too much applications, so when I know the computer starts to run slowly, I start to close my applications. I usually run 4-5 applications simultaneously. Usually those are: web browser (Google Chrome), Adobe Photoshop, Notepad++, XAMPP, and Windows Task Manager. Usually I close tabs in my Chrome first, because I often browse the net with about 20 tabs/4 windows open, so I presume that would take much memory (bad habit, I know). But even after closing Chrome's tabs or closing other applications, my CPU Usage often stays at high percentage - 72% at best, 100% at worst. I check the Processes tab on Windows Task Manager and usually found the System, System Idle Process, or services.exe taking the highest CPU process (could reach 60). Why is this happening? And is there any solution? EDIT I have T2250 @ 1,73 Ghz and 2.5 GB RAM

    Read the article

  • How to setup email server in ubuntu 12.04LTS(debian 7 wheezy/sid) running on linode vps

    - by shihon
    I am working on email server, since i tried several times to create email server on ubuntu12.04LTS with postfix + dovecote + postfixadmin + courier + clamav + spamassassin. But everytime i install these packages i face new problems, like mails send to localhost users and found in users maildir. But I can't determine how to configure/setup for send an email to external smtp like gmail, yahoo. The most worst thing i can't determine how to use sasl, because i am not using SSL so it is not worthy for my domain. This is so complicated, i search everywhere on google: links are https://help.ubuntu.com/community/PostfixCompleteVirtualMailSystemHowto http://www.starbridge.org/spip/spip.php?article1&lang=fr http://knopix.wordpress.com/2008/01/16/postfixadmin-postgresql-courier-squirrelmail-on-debian-etch-howtotutorial/ http://flurdy.com/docs/postfix/ Is there any article for install email server on ubuntu 12.04LTS. Please help me to understand these things.

    Read the article

  • Amazon and bandwidth limits

    - by Dave
    This question may sound weird to some of you but I have never really used cloud and above all I'm still beginner in the web development and would be really thankful if someone could answer few of my question though they may sound weird So I would like to deploy simple website to Amazon, however, I'm concerned about bandwidth as they charge 0.12GB and I'm not able to set budget limit. My problem is that I wouldn't like to pay for 1000GB of bandwidth if someone for some reason decides to download one file constantly So could some of you, who have experience with amazon, tell what happens if my app is able to handle (say 50 req/sec 30kb/page) does that mean that in the worst case I would have to pay req * sec * min * hours * days * page size 50 * 60 * 60 * 24 * 30 * 30kb = 3888GB

    Read the article

< Previous Page | 12 13 14 15 16 17 18 19 20 21 22 23  | Next Page >