Search Results

Search found 13249 results on 530 pages for 'virtualized performance'.

Page 360/530 | < Previous Page | 356 357 358 359 360 361 362 363 364 365 366 367  | Next Page >

  • Ways to setup a ZFS pool on a device without possibility to create/manage partitions?

    - by Karl Richter
    I have a NAS where I don't have a possibility to create and manage partitions (maybe I could with some hacks that I don't want to make). What ways to setup multiple ZFS pools with one partition each (for starters - just want to use deduplication) exist? The setup should work with the NAS, i.e. over network (I'd mount the images via NFS or cifs). My ideas and associated issues so far: sparse files mounted over loop device (specifying sparse file directly as ZFS vdev doesn't work, see Can I choose a sparse file as vdev for a zfs pool?): problem that the name/number of the assigned loop device is anything but constant, not sure how increasing the number loop device with kernel parameter affects performance (there has to be a reason to limit it to 8 in the default value, right?)

    Read the article

  • When adding second processor to SQL Server, will it automatically balance the load?

    - by ddavis
    We have a SQL Server 2008 R2 (10.5) on a dedicated box with a single 2.4Ghz processor, which regularly runs at 70-80% CPU. We are going to be adding a significant number of users to the application and therefore want to add a second processor to the box (scale up). Will SQL Server automatically use the second processor to balance threads, or is there additional configuration that will need to be done? In other words, will adding the second processor drop my CPU usage to 35-40% per CPU, automatically balancing the load? Based on what I read here, it seems that it will: http://msdn.microsoft.com/en-us/library/ms181007.aspx However, I've read elsewhere that CPU performance gains can be made by assigning database tables to different filegroups, but I'm not sure we want to get that complicated at this point.

    Read the article

  • Redirect traffic to local address so iOS speedtest app measures LAN speed

    - by ivan_sig
    I have mounted a Speedtest Mini server on a local LAMP, so I can test my LAN speeds effortlessly just by opening the URL with a Flash enabled web browser, the thing is, I want my iOS and Android devices to test with the LAN server too, not with the WAN, as I'm trying to measure LAN-Only performance. Is there a way so I can redirect the traffic intended to an specific external IP (The one of the real server) to my local server?. I know the servers IP as a short Wireshark analysis gave me the data, but still searching for a way to make that redirect. I have Jailbreak and root on my devices, so playing with system files is not a problem. I've tried mounting a proxy and making redirects by the hosts file and domain names, but it looks like Ookla's app relies on IP address only.

    Read the article

  • Dual Monitors don't turn off/sleep when they should

    - by Mario
    Details: Dual 25" monitors connected to HP Envy via HDMI and display port via DVI adapter. Power scheme is set up in high performance (Dim Display: Never - Turn Off Display: 15 mins - Computer Sleep: Never) Screensaver is set up to kick in after 10 mins of idle (which happens) 5 minutes later, the screensaver stops. The "Monitor Going to Sleep" notice appears on screen and monitors go to sleep briefly. All is well thus far. Suddenly, the Windows 7 alert sound when a device is unplugged is heard. Monitors then turn back on. Screens are black. Only the mouse cursor is displayed. Backlighting is back on. This only started happening after I obtained and connected the second 25" monitor a few days ago. However, I had a 24" in its place before, and this wasn't happening. Why is this happening and how do I correct this behavior? Thanks in advance.

    Read the article

  • How to Increase Memory Allocated to IIS .NET Application?

    - by Mark Hansen
    We are using Windows 2008 R2 and IIS 7 running on Amazon EC2. IIS is running a single .NET application written in C#. We are having performance issues and I want to give the application more memory, but I cannot figure out how to do it. How do I control the amount of memory that the CLR gets? I'm a total newbie with IIS, .NET and the CLR. If I were working with Java, I would just use the -Xmx flag to increase the memory available to the JVM (e.g., -Xmx3000m for 3GB). But, I cannot seem to figure out how to do this in the Windows world.

    Read the article

  • Can NFS be forced to refresh stale files/directories when not using noac on the mount?

    - by johnnycrash
    We mount without using noac. I have a file that I append to once every 20 minutes. Then it will be read with mmap about 5,000 times a minute. We only mmap a couple blocks for each read. Needless to say, noac just kills the access performance, so we don't use it. I add data to the end of the file using a mount with noac and read from a mount without noac. The mounts that are reading are not seeing the new data. I want to know if there is a function I can call from c to refresh the attributes of a path and all its files. EDIT: I should add we cannot mount and unmount since there are 16 servers running on each system and they are constantly accessing the files. Well...maybe we could mount and unmount if each server used their own mount. I'd like to avoid that if possible. thanks!

    Read the article

  • BES 5.0 and MAPI calls to exchange system

    - by nysingh
    We have been using BES 4.1(5) for a while now and it has been a resource hog on exchange due to high number of MAPI calls. I have heard that BES 5.0 is even worse. the comparison i heard is that BES 4.1 is makes MAPI calls equal to 5 outlook clients per BB user and BES 5.0 makes MAPI calls equal to 10 outlook clients per BB user. can someone confirm if it is true? is BES 5.0 is really that bad in MAPI calls and for exchange performance. ? thanks

    Read the article

  • Windows 8 / Server 2012 RDP connection is slow

    - by Chris
    I recently installed Windows Server 2012 for development purposes at our office and noticed immediately that connecting via RDP is slow. It can take 5-10 seconds to connect at times, where as connecting to any of our Win7 or Win2008R2 boxes takes at most 1-3 seconds. At first, I chalked this up to the box itself needing a driver update or something, but just yesterday, I installed Win8 on my desk PC and connecting from home to that machine produces the same result. There is a 3-4 second pause at "securing remote connection" and then again at "configuring remote session". I don't see any warnings in the event log, and once connected, there do not appear to be any performance issues. Is there a known problem with RDP connections on Windows 8 systems? Anything I should look for?

    Read the article

  • Odd Language In a BIOS Message

    - by Josh
    So I started up my laptop today and was greeted with the following message (not a direct quote): The type of the AC adapter cannot be determined. This may interfere with your computer's performance. Try unplugging the AC adapter and then plugging it back in, thanks. The problem was that I hadn't fully secured the plug into the back of the computer. However, I was a little taken aback when a message from BIOS said, "thanks." Is this normal? Any chance the message was illegitimate (virus)?

    Read the article

  • Determine compression ratio for Windows compressed drive

    - by munrobasher
    Is there a Windows 7 native way to display the overall compression ratio on a Windows compressed drive? As part of our disaster recovery process, we're copying some key system folders onto 2TB external hard drive, encrypted using TrueCrypt and copied using robocopy. The drive is compressed and I'd like to see what kind of compression ratio we're getting and whether it's actually worth the performance overhead. I know that TreeSize can possibly do this (as mentioned in another post) but want a OS native way if possible. Thanks, Rob.

    Read the article

  • Which PSU should one chose? The biggest is the best?

    - by Shiki
    I'm fully aware of PSU's "Active PFC" and that they won't consume the written W all the time. (Makes sense). But now I'm before a PSU replacement (Guys: NEVER buy a Chieftec. Seriously.) The question is: If one can get a bigger one (in my case 750W and 650W) ... should that person go for the bigger one ? (The difference in price is not much). No, I don't think I'll soon use all that much. (Please help (if you want of course) to make the question more generic if the question is really not OK in this form. I've been wondering about this for a time already. In my case it would be XFX Black Edition Silver 750W and 650W) (Basically about "which one" I would go with XFX/Antec/something which comes with industry qualified parts. Like Duracell but in a PSU. :) But the performance is a different thing.)

    Read the article

  • Running OpenVZ virtual servers within a Xen XCP vritual server? Bad practice?

    - by Damainman
    I have a 1 server with 8GB RAM and 2xQuadcore Processors. It currently has the Xen XCP installed on it, and centos6.2 x64 running on a virtual machine. I have a server control panel software that I want to use and it allows the administration via a web interface for Openvz machines. My questions are: Would this be considered bad practice? Would there be a big performance hit? Should I avoid this all together or am I going about it all wrong? Thank you in advance.

    Read the article

  • An international mobile app - Should I set up EC2 instances in multiple regions?

    - by ashiina
    I am currently trying to launch an mobile app for users around the world. It is not a spectacular launch which will get millions of users in weeks - just another individual developer releasing an app. I know enough about the techniques of managing timezones, internationalizing string, and what not ( the application layer ). But I cannot find any information on how I should manage my EC2 instances... Should I be setting up EC2 instances in different regions around the world? Is that a must-do, or is it an overkill? I'm aware that it's the ideal solution in terms of performance, but it becomes very tough managing servers in multiple regions. DB issues, AMI management, etc... I'd much rather NOT do so. So I would like to know the general best practice when launching an international app/website. Note: For static contents, I know it's better to use a CDN, so I'm planning on doing so.

    Read the article

  • Do I need more RAM for programming if I switch to a 64 Bit OS?

    - by Buttercup
    Hi, Given that today (for performance reasons?) variables are usually aligned to the "bit-width" (I wanted to use "word-width" here, but on x86 a "word" is still 16 bits right?) of the processor, would switching from a 32 bit OS to its 64 bit version double the RAM usage? Would this then in turn mean that to do the same work a 32 bit OS can do with 4 GB RAM (well, the 3.x GB actually...) with a 64 bit OS I would need 8 GB of RAM for programming? Please note that I'm only talking about everyday x86 computers here.

    Read the article

  • SSD runs faster on Windows as compared to Linux [closed]

    - by wushugene
    Windows 7 seems to install, boot and run much smoother & faster than each the three linux distros I have recently tried (Ubuntu 12.04 unity, Linux mint 13 MATE, and Fedora 17 on gnome 3.4). Why am I facing bad performance in Linux? I have tweaked my Linux installs for the SSD (enabling trim, disabling swap, etc.) I'm using an Acer TravelMate with i5-2410m processor, intel hd 3000 graphics, 8 gigs of ram, and a 256 gb samsung 830 ssd. Edit: Boot times are 10-15 seconds slower, there is noticeable delay from login to fully loaded desktop, and in general does not appear to be as responsive as my old windows 7 install or the Linux guests I had running on it.

    Read the article

  • Missing 16:10 resolutions with Nvidia drivers (Can't add resolutions)

    - by Wuinny
    Hello, I have a laptop with a Nvidia 9650M GT and used the drivers that Seven brought me. It works fine but Metro 2033 tells me that i have to upgrade my drivers to play the game. So i did it. But since i did a clean install of the new Nvidia drivers, i just have 1440*900 or 4:3 resolutions. I usually played with 1280*800 or 1184*740 (for performance issue) With the "old" drivers i was able to create custom resolution (1184*740) in Nvidia control panel but now when i try it tells me that "my monitor cannot support this resolution". When i insist, it works but soon as i shut down my computer i have to recreate it.. Do anyone have a fix ? Thank you

    Read the article

  • Hardware recommendations for building an Ubuntu encrypted file server

    - by Robert Mashlan
    I would like to build a file server for my home network using Ubuntu. It will serve files from RAID1 configured disks, either in the OS or in hardware. It will be connected to a Gigabit ethernet LAN. The disks will use an encrypted file system. It will serve samba shares. I would like a recommendation on what kind of processing power/memory I would need to build a box that would be able to sustain the full capacity of the Gigabit ethernet connection in a file transfer for a single connection with the overhead of serving from an encrypted disk. I'm not looking to build a dream server, I just want enough processing capacity for high performance (and reliable) file sharing and spend as little as possible for it. This may be tangential, but what kind of hardware would I need to have a server be able to reliably go into a low power mode when no requests are being made of it?

    Read the article

  • Windows file Sync

    - by Deane Venske
    So I have a big problem at the moment. Trying to find a reliable solution for syncing 2 windows IIS servers. I need to keep the web content imaged on both. Now I have been trying to use Rsync to this point, but unfortunately file permission errors are a nightmare to manage this way. I'm testing out dropbox, but the performance sucks. I'm more familiar with Linux stuff and I've used Rsync in the past but isn't there a native windows solution that will work?

    Read the article

  • what to use instead of laptop-mode?

    - by playcat
    hello, i have ubuntu 10.10 64bit on hp 6735s (turion processor). it overheats, and i'm forced to use turion power control in order to keep core temperature to a reasonable level. one more measure that i use is putting my processors to conservative mode. that way, i'm perfectly happy with its performance, and heat is where it should be. however, after my latest upgrade, something happened - cores are back to ondemand by default, and i'm not sure if turionpowercontrol is working any more (ps axu | grep urion shows no process). in addition, i read somewhere that laptop-mode uses hdd spindown for preserving data/energy, and that hdds have only a limited amount of those spindowns, so laptop-mode usage can actually shorten the life of my hdd. i'm wondering if there is a good way to set my cores to automatically go to conservative mode? also, what's the good way to see what is the voltage my cores use? on windows i use cpuz tools. thx and sorry for the long explanation.

    Read the article

  • Copy an Amazon EC2 Instance to use locally

    - by Excolo
    Ok, so we have a spare server I have installed Debian Wheezy on, and setup Xen on for virtual machines. It has better performance than all our ec2 instances combined, and will cost less to run (for a few various reasons) I would like to get the EC2 instances downloaded to my server, and converted to run for Xen, but im having difficulty finding anything specific. I did not setup the EC2 instances myself, and am not very familiar with them. Everything I have found (which isnt much) just says "Do XYZ" and I have no idea how to do those. So being as specific as possible would be helpful. Also, confusingly I see people writing in forums saying you can only export linux images (which mine are, Ubuntu images) but then I see on amazons export tool saying you can only export Windows server? Am I missing something here? Is that not the right place to be looking? Thanks

    Read the article

  • useful JMX metrics for monitoring WebSphere Application Server (and apps inside it)?

    - by Justin Grant
    When managing custom Java applications hosted inside WebSphere Application Server, what JMX metrics do you find most useful for monitoring performance, monitoring availability, and troubleshooting problems? And how do you prefer to slice and visualize those metrics (e.g. chart by top 10 hosts, graph by app, etc.). The more details I can get, the better, as I need to specify a standard set of reports which IT can offer to owners of applications hosted by IT, which those owners can customize but many won't bother. So I'll need to come up with a bunch of generally-applicable reports which most groups can use out-of-the-box. Obviously there's no one perfect answer to this question, so I'll accept the answer with the most comprehensive details and I'll be generous about upvoting any other useful answer. My question is WebSphere-specific, but I realize that most JMX metrics are equally applicable across any container, so feel free to give an answer for JBoss, Tomcat, WebLogic, etc.

    Read the article

  • SSL on Apache seems to significantly affect WebDAV performace

    - by takesides
    I'm using Apache 2.2 running on Windows Server 2008 R2 as a WebDAV server for clients to upload large media files (roughly 100-2000MB). I am finding that when I have SSL enabled (openSSL 0.9.8o) and use HTTPS for the uploads the throughput is around 13Mbps but when I disable it and just use HTTP I get around 80Mbps. I can't understand why this is happening as my understanding was that the heavy SSL work was done at the beginning of the connection. Does anyone have any idea why the performance is so drastically affected by enabling SSL? Cheers.

    Read the article

  • In BASH, are wildcard expansions guaranteed to be in order?

    - by ArtB
    Is the expansion of a wildcard in BASH guaranteed to be in alphabetical order? I forced to split a large file into [10Mb pieces][1] so that they can be be accepted by my Mercurial repository. So I was thinking I could use: split -b 10485760 Big.file BigFilePiece. and then in place of: cat BigFile | bigFileProcessor I could do: cat BigFilePiece.* | bigFileProcessor In its place. However, I could not find anywhere that guaranteed that the expansion of the asterisk (aka wildcard, aka '*' ) would always be in alphabetical order so that .aa came before .ab ( as opposed to be timestamp ordering or something like that ). Also, are there any flaws in my plan? How great is the performance cost of cating the file together?

    Read the article

  • VirtualBox: Ubuntu installed in Windows 7 OR Windows 7 installed in Ubuntu

    - by user584968
    The title of the post sums up my questions: Is performance better, on average, assuming Ubuntu and Windows are used equally and for the same purposes, when Ubuntu is installed inside of Windows 7 or when Windows 7 is installed inside of Ubuntu using VirtualBox? My gut instinct tells me that because Ubuntu drivers are mostly open source, optimizations could be made when translating the virtual system calls of Ubuntu through Windows 7 that could not be made if it were flipped the other way around, because of the closed nature of Windows drivers. However, this is just a guess, and I have no evidence to back this up. I'm not sure, so I thought it would be best if I opened up this question to the experts of the community. I look forward to reading your responses!

    Read the article

  • Multiple iSCSI Targets or 1 that's shared?

    - by Joost Verdaasdonk
    On my network I have several types of files I want to save on a SAN like: SQL db's and logs Exchange data Random files Now I'm wondering if I should create one iSCSI Target with a large volume and initiate that from one of the servers. (and share it so other servers can use it too) Or I should create separate Targets to have each server use its own storage. For the record the storage could be separated because the servers aren't using the shared data. For one reason I was thinking of one storage is ease of backup. (but perhaps performance could be a problem?) What would be an advisable configuration for these type of data?

    Read the article

< Previous Page | 356 357 358 359 360 361 362 363 364 365 366 367  | Next Page >