Search Results

Search found 13249 results on 530 pages for 'performance tuning'.

Page 365/530 | < Previous Page | 361 362 363 364 365 366 367 368 369 370 371 372  | Next Page >

  • Memcached clustered alternative

    - by Johan Kooijman
    I'm looking to replace memcached. We have a LOT of traffic to our central memcached node which I'd like to split. There's only so much trunking networks I can do. My general idea is to install a memcached-type daemon on every webserver and have the daemons replicate set/delete/updates over all the daemons, so that each webserver connects to a socket or on localhost. All data should be available on all nodes. The alternatives: - repcached (max 2 masters) - redis (single master) - couchdb/mongodb/handlersocket - persistent data on disk, I'd like to remove the disk part to gain more performance. Any hints?

    Read the article

  • If I double my ram on a x86 processor, does that double the ram I can use for each individual process?

    - by Derek Reitz
    I don't understand how 32-Bit OS's use RAM on a per process basis. I've read the max RAM my x86 processor running a 32-bit OS can use is 2^32 = 4gb; but that's just for one process, right? 3DS Max keeps crashing, but it typically can never use more than 2GB of RAM before it crashes, if I increase my RAM from 4-8GB, would that double how much RAM I can use for each individual process or actually cause no change in my performance? Also would increasing my VRAM and getting a better graphics card increase the extent to which individual programs can preform? Lastly, is there any way to upgrade a 86-bit processor to be able to run a 64-bit OS? I feel like it would be ridiculous to sell modern processors that are capped at 4GB of RAM? Thanks. Quad-Core Intel i7 Q 720 @ 1.6GHz

    Read the article

  • VirtualBox: Ubuntu installed in Windows 7 OR Windows 7 installed in Ubuntu

    - by user584968
    The title of the post sums up my questions: Is performance better, on average, assuming Ubuntu and Windows are used equally and for the same purposes, when Ubuntu is installed inside of Windows 7 or when Windows 7 is installed inside of Ubuntu using VirtualBox? My gut instinct tells me that because Ubuntu drivers are mostly open source, optimizations could be made when translating the virtual system calls of Ubuntu through Windows 7 that could not be made if it were flipped the other way around, because of the closed nature of Windows drivers. However, this is just a guess, and I have no evidence to back this up. I'm not sure, so I thought it would be best if I opened up this question to the experts of the community. I look forward to reading your responses!

    Read the article

  • After moving our Servers to a virtual environment using VMware - SQL timeouts came in, why?

    - by RayofCommand
    We moved our servers to a virtual cloud (VMware) where only our servers are in. But as soon as we finished migrating everything we are fighting against SQL Timeouts and machine slowdowns we can't explain. Even though we ~ doubled the servers capacity while switching from physical to virtual. Now I googled and found that we are not alone. People are complaining about poor performance after moving to a cloud managed by VMware. Are there any known issues? Sometimes our services can't access a disk or SQL receives a timeout and we have no idea why.

    Read the article

  • FPS lags with new acer aspire 5755G

    - by Calvin
    The title is kind of self explanatory, as my new laptop lags and has FPS drops. For example my FPS in Starcraft 2 hovers around 20 and constantly drops to 1 with low settings when I know it should run smoothly in high settings. I've updated my Nvidia driver, and set the preferred global settings to the 'High-performance Nvidia processor'. Here are some screen shots. Screen Shot One - Screen Shot Two - Screen Shot Three I'm not sure how to fix this problem, any feed back would be nice!

    Read the article

  • Is it possible to cause artificial network packet loss or latency?

    - by nbolton
    I'm trying to reproduce some issues on a deployed application where the MSSQL server and client are running in two separate machines. I think there may be network issues between the two machines, so I'd like to try and reproduce these conditions on two Hyper-V virtual machines (on the same virtual server). Of course, the network for these virtual machines is "local" so it's actually far from the conditions in a live environment. Is there a program I can run on either virtual machine which will degrade the network performance? Or maybe any other work arounds? For example, one way to reproduce the conditions may be to run the VMs on separate Hyper-V servers in geographically dispersed locations (so the SQL traffic goes over VPN or something) -- but this is a little long winded I think. There must be a simpler way.

    Read the article

  • Openvpn client-to-client connection reencrypted at Server?

    - by user1684411
    currently i'm using a site-2-site openvpn setup. The routers en/decrypt all traffic that goes from one net to another. One of them is the Openvpn server. This works but performance is not as good as possible. I think the limiting factor is the cpu power of the router. Would it be better if i use client-to-client connections and access the fileserver in the one net from a pc in the other, because the openvpn-server does not have to decrypt the (whole) packets?

    Read the article

  • Prevent Windows 7 from unpluging External HDD [migrated]

    - by marverix
    I have installed Windows 7 as my media server. I pluged in 500GB external HDD via USB. I have changed power plan to Best Performance and changed advenced power settings to never turn off HDD etc. I even yesterday wrote powershell script (create and delete folder on this disk) and I have added it to harmonogram to run every 5 minutes starting from system boot. And nothing! Disk after some time (I realy can't say when) is turning off and Windows show "Unknown Device" in Device Manager. Then only system reboot or disk reboot helps. Any ideas how to prevent Windows 7 from stopping my external HDD? Cheers!

    Read the article

  • How can I use rsync with a FAT file system?

    - by Kim
    I would like to write a simple backup script that saves some data to a FAT drive. Should I reformat the drive and use a better file system or is it possible to use rsync with FAT? If so, what problems might I run into? Would performance be a lot worse? EDIT: This is on linux, didn't even know there was a rsync for windows. The sources are various file systems (it's a mess), and the destination is currently formatted with FAT32. Thank you for your answers, I'll probably go for a reformat, since I'm not completely sure about the file sizes we'll have.

    Read the article

  • Drobo Pro or SAN for backups

    - by PHLiGHT
    I have multiple offices and was thinking of using two Drobo Pros as a storage location for my backups, with one at each site being synced. It seems like on the surface its a cheap alternative to a SAN. Because it will just be used for storing backups I don't think that performance is as big of an issue. Related question. How do you sync your backups from your main location to a remote site. I am using Backup Exec and wondered if I should use a feature in the program or use something like Delta Copy.

    Read the article

  • large RAID 10 vs small RAID1

    - by user116399
    The machine will store and serve millions of small files (<15Kb each), and all those files require a total storage space of 400G Considering the exact same SATA hard drives maker and models, on the exact same environment (OS, cpu, ram, raid controller, etc...) which one of the setups bellow would be faster? A) RAID 1 with 2 drives of 2T each, making up total storage of 2T B) RAID 10 with 4 drives of 2T each, making up total storage of 4T [EDIT]: I'm aware RAID10 is faster than RAID1. The larger the disk, at least in theory, the longer will take to do seeks/writes. So, will the performance gain of RAID10 will be outweighed by the "drag" caused the larger disk area when seek/write operations happened?

    Read the article

  • What's better for deploying a website + DB on EC2: 2 small VM or a large one?

    - by devguy
    I'm planning the deployment of a mid-sized website with a SQL Server Standard DB. I've chosen Amazon EC2 to deploy it. I now have to choose between these 2 options: 1) get 2 small instances (1 core each, 1.7 GB of ram each): one for the IIS front-end, one for running the DB. Note: these "small instances" can only run the 32-bit version of Win2008 Server 2) a single large instance (4 cores, 7.5 gb of ram) where I'd install both IIS and the SQL Server. Note: this large instance can only run the 64-bit version of Win2008 Server What's better in terms on performance, scalability, ease of management (launch up a new instance while I backup the principal instance) etc. All suggestions and points of view are welcome!

    Read the article

  • KVM guest storage difference with NBD and NFS

    - by WojonsTech
    I am setting up my own little private cloud for my own use maybe for a project or to. I am using linux kvm on debian 6. I have 3 servers 2 of them for compute nodes and 1 storage node. I would I have already installed kvm made a few test machines got my networking setup. I have 2 nics on each server 1 nic is for web traffic other nic is for network traffic. My first Idea was to use NFS for storing the guest machines which can range in size, maybe 8gb maybe 100gb, it just depends. I was doing have heard of nbd before seems like it could work but I dont know what the performance differences are and if it will effect my enviroment, nfs looks like it will be easier to use.

    Read the article

  • What maintenance is required for a Postfix setup?

    - by JonLim
    I've taken a look at the setup and configuration process for a Postfix server, planning to use it for just sending emails out from my server. So far, I have these steps: Setup Postfix Configure Postfix Install DKIM Set SPF records Tune for performance Debug Seems rather straightforward. However, I was just wondering: are there any actions I should be taking for periodic maintenance of my Postfix setup? Thanks! EDIT: Also, just curious, how long would this entire setup ideally take? 30 - 60 minutes? More?

    Read the article

  • Sharing RAM resources between 2 or more computers

    - by davee44
    I know there was a somewhat similar question before: How to share CPU or RAM? But, let me just specify it a little more... When Microsoft Windows requires more RAM capacity than available it uses a swap-file to temporarily store the data there, this is actually something like a hard-drive-based RAM. This technology is used for many years. Theoretically, it shouldn't be too hard to implement a similar technology that uses the RAM of different computer(s) in the network for temporary data storage. This just requires a software that runs on computers in the network that accepts and returns data from/to the main computer and keep that data in the RAM; plus the operation system of the main computer must have the ability to use computers in the network instead of (or in addition to) the swap-file. I wonder, are there any implementations of this idea? This would allow users to build RAM clusters using all of their home or office computers, that will boost the performance of a single computer for some development/gaming/video tasks, etc.

    Read the article

  • Which modules can be disabled in apache2.4 on windows

    - by j0h
    I have an Apache 2.4 webserver running on Windows. I am looking into system hardening and the config file httpd.conf. There are numerous load modules and I am wondering which modules I can safely disable for performance and / or security improvements. Some examples of things I would think I can disable are: LoadModule cgi_module others like LoadModule rewrite_module LoadModule version_module LoadModule proxy_module LoadModule setenvif_module I am not so sure they can be disabled. I am running php5 as a scripting engine, with no databases, and that is it. My loaded modules are: core mod_win32 mpm_winnt http_core mod_so mod_access_compat mod_actions mod_alias mod_allowmethods mod_asis mod_auth_basic mod_authn_core mod_authn_file mod_authz_core mod_authz_groupfile mod_authz_host mod_authz_user mod_autoindex mod_dav_lock mod_dir mod_env mod_headers mod_include mod_info mod_isapi mod_log_config mod_cache_disk mod_mime mod_negotiation mod_proxy mod_proxy_ajp mod_rewrite mod_setenvif mod_socache_shmcb mod_ssl mod_status mod_version mod_php5

    Read the article

  • How (much) is virtualization used today?

    - by BLAKE
    I know that where I have worked, I have pushed alot for virtualizing our servers. I think that it is much easier to implement and maintain than physical servers. I have been using Microsoft's Virtual Server 2005 R2 since it was released. Right now at my workplace we have 12 VMHosts that hold about 55 VMs. We have 6 other servers that we have been unable to convert to VMs. I want to know how other people in our field view virtualization. I know that I have had developers dislike the notion of VMs claiming major performance hits. What do other Sys Admins think about virtualized servers?

    Read the article

  • XDMCP is slow any ideas? (looking for alternative remote desktops)

    - by peteri
    I've been used to using RDP on Windows to remote to machines, and I've got an asus eee 701 which I want to use to do some *nix programming on. While the eee is a lovely little machine the screen and keyboard are a little small to use for lots of programming. I've tried using Xming (the free version) to remote login into the eee from my desktop using XDMCP (or even using a ssh session as a straight X11 server and no desktop on the eee) the whole thing seems seriously slow over wifi the initial desktop takes at least 5 seconds to paint (might even be 10 seconds I haven't actually timed it). So my real question is what do other folks use for remote control with a GUI for their *nix boxes? I am finding it hard to believe the performance is so bad over a wifi network (It makes the Mac IIs I used to use a college in 1988 seem fast) or is this just a problem with Xming and using say the Cygwin X11 server would be better.

    Read the article

  • what are the rules for SLI ( GTX 550 Ti )

    - by equivalent8
    I got ASUS GTX 550 Ti and I want to SLI it with another graphic card. I heard that not all graphic cards are good idea to SLI, (or not all combinations) because sometimes the final performance could be even worse that with one graphic card. Is that true? What are the rules ? ( maybe chip-set needs to be same or something ? ) I was wondering if you can recommend me what Graphic card should I use as with mine. Should I use same one (GTX 550 Ti) ?

    Read the article

  • Fastest security check of file tree on NFS

    - by fungs
    I am currently experiencing very bad performance using the following on an NFS network folder: time find . | while read f; do test -L "$f" && f=$(readlink -m $f); grp="$(stat -c %G $f)"; perm="$(stat -c %A $f)"; done Question 1) Within the loop permissions are checked using the variables grp and perm. Is there a way to lower the amount of disc I/O for these kind of checks over the network (e.g. read all meta data at once using find)? Question 2) It seems like the NFS isn't tuned very well, the same operation on a similar network link via SSHFS take only one third of the time. All parameters are auto-negotiated. Any suggestions?

    Read the article

  • Solaris 10: Identify a PID and the CPU it's running on

    - by Marcus
    I have multiple instances of a database running on a Solaris system. I'd like to prove that each database process is being handled by a different CPU. Essentially, I want to be able to do something like a ps -ef | grep <process_name> to get the PIDs and then run another command (if required) to identify the CPU... Is prstat able to do this? I'm making an assumption that as each database instance is started each one uses a different CPU. I'm not sure if I'm understanding this correctly... The reason I want to do this is because Sun hardware has slow CPU's, but lots of them. Therefore, to get the best performance out of it, I need to try and spread the load among CPU's... Thanks

    Read the article

  • Solaris 10: Identify a PID and the CPU it's running on

    - by Marcus
    I have multiple instances of a database running on a Solaris system. I'd like to prove that each database process is being handled by a different CPU. Essentially, I want to be able to do something like a ps -ef | grep <process_name> to get the PIDs and then run another command (if required) to identify the CPU... Is prstat able to do this? I'm making an assumption that as each database instance is started each one uses a different CPU. I'm not sure if I'm understanding this correctly... The reason I want to do this is because Sun hardware has slow CPU's, but lots of them. Therefore, to get the best performance out of it, I need to try and spread the load among CPU's... Thanks

    Read the article

  • How to measure that a host is good for users in Egypt ?

    - by Sherif Buzz
    Hi all, I currently have a site that's hosted in Texas. The majority of my users are from Egypt and I'm a bit concerned that the current hosting is not the optimal in terms of performance. The site is not slow but for how can I know if, for example, hosting it in Europe or Asia is better ? To clarify I need to know there is a way that I can test different hosting options - for example how can I test the average response time between Egypt and a host in Texas, the average response time between Egypt and a host in the UK ?

    Read the article

  • Hardware changes to require XP Activation ? (for a Virtual Machine)

    - by NVRAM
    I have WXP-64 running on a VM and, for testing and performance reasons, I would like to occasionally change the allocations for it. Changes might include: Number of CPU cores, Amount of RAM Add/remove network adapters. But I'm concerned that XP will demand re-activation and that I might eventually have licensing issues if I do this. So, can anyone tell me: What kind of changes trigger re-activation in XP? Is there limits or caveat with regard to re-activation? I've perused this question and the article it references, but wanted more recent and verified info. (FWIW, I'm not trying to cheat: the OS copy was purchased explicitly for the VM.)

    Read the article

  • Web Server Setup

    - by gustyaquino
    Hello, In my workplace, we want to implement your own web server for at leat 100 Apache/PHP/MySQL web pages. My boss is opposed to hiring skilled personnel, he think we can do ourselves. Currently, we are working with hostgator reseller account. I chose CentOS as the operating system, but I don't know the best hardware solution. HP, Dell ? What about the setup on these platforms? Thanks. PS: sorry for my bad english Edit: The purpose of this migration isn't related to performance issues. But independence.

    Read the article

< Previous Page | 361 362 363 364 365 366 367 368 369 370 371 372  | Next Page >