Search Results

Search found 22065 results on 883 pages for 'performance testing'.

Page 539/883 | < Previous Page | 535 536 537 538 539 540 541 542 543 544 545 546  | Next Page >

  • 10,000 RPM HDD (WD VelociRaptor) vs SSD for OS?

    - by GiH
    I currently have a 10,00RPM 150GB Raptor that I use for Vista. I'm about to upgrade to Windows 7, and while doing that I thought I'd buy another drive and install Ubuntu 9.10 on it. I don't want to partition the current drive I have, but I don't need 150GB for another OS. So, I'm having trouble deciding whether its worth it to buy a 64 GB SSD at the same price point as the 150GB WD VelociRaptor? Or should I just get a 7,200 RPM drive for really cheap (around $50)? Would it be better to use an SSD for the OS than a mechanical drive? I could always get a 32GB SSD too... Oh, and I don't want to virtualize Ubuntu because I'm going to be testing to see the differences in networking and overall performance.

    Read the article

  • Unable to Get IIS Response Times with Hyperic Monitoring Tool

    - by jwmajors81
    We have a very large .NET application (vendor) that I am trying to gather performance metrics for using Hyperic. In general I wanted to be able to report response times for each of the components within the application, which include: Web Services ASP.NET Pages MSMQ I am currently unable to successfully monitor the response times for IIS on my windows machines. I have successfully auto-discovered them and I am getting diagnostic information, but am not getting back the response times. After looking online at what other people are seeing, I found that I am missing a tab for the response times which is usually next to the metrics tab. Also, when I look at the configuration screen for IIS I do not see a field which enables me to specify where the log files are located at. Please note that my logs are located at e:\LogFiles\ instead of the usual location because our IT staff doesn't allocate much space for the C: drive. Please note that I have the MSMQ monitors up and running great. Any assistance would be greatly appreciated. Thanks, Jeremy

    Read the article

  • Why does a group policy not applied to the domain administrator account?

    - by Saariko
    I have a working policy on my entire domain. I just found out, when logging with the domain administrator, that this policy is not applied (EDIT: Running : gpresult shows that the GPO's are applied - but, this GPO is for Drive Mappings, and the actual drive mappings are NOT shown) The administrator account - does not have any login script on his profile tab. My GPO's are mainly small/atomic settings: single GPO to handle each settings: UAC, Firewall, printers. GPO status for the object is enabled That's an overview of the Drive Maps: Reading on MS support site, I checked the delegation tab, and it is marked as applied to domain and enterprise admins. Every user gets these policies correctly. The OU that is set is the root of the domain. (for testing purpose - I did that to eliminate hierarchy issues - did not help) Block Inheritance is disabled. (never used it anyway) GPO link GPO Security Filterings

    Read the article

  • NTFS disk mounted as fuseblk in ubuntu 12.10 is very slow and a lot of errors when rsync. Is that not a rare thing?

    - by Pablo Marin-Garcia
    I am having problems with a NTFS disk mounted as a fuseblk in my ubuntu 12.10 through external usb3. When I did a 1.1TB backup with rsync the speed was 1-2MB/s (wiht a ext4 disk speed was 70 MB/s before and after trying the NTFS disk). Also after one hour errors started to appear: rsync: write failed on "xxx": No such file or directory recv_files: "yyy" is a directory #but this file is a FILE not a dir ??!! .... As this is the first time I have mounted the NTFS in linux for heavy usage (the data would be used in windows afterwards), I would like to know if this kind of thinks are common o was only that something became unstable in my system and a simply restart would probably have solved it. This leads me to the these questions: Can I trust fuse for manage NTFS disks? Or is a problem of the NTFS tools in linux not yet totally stables for writing? Do people is still suffering from low performance with fuse-NTFS vs ext4 (in the past I have read about people complaining about this)?

    Read the article

  • dell u2410 3dMark Benchmark distortion problems

    - by Scanningcrew
    Ive been doing burn in testing for a new system I have put together and I am running into some video distortion problems with running the 3DMark benchmark tools (Both 06 and Vantage). The graphics will be fine, then sometimes during a test switch the screen will light up with thin horizontal ranibow lines (Something that looks very "glitchy") If i turn the monitor off and back on it clears up. All the tests "pass" and my system gets good marks but it concerns me if I might have problems with games (The screen returns to normal if I dont resest monitor and just let tests pass). I want to return a problem component now before its too late if it is something with the new hardware. Also, I am monitoring temp with thermal laser gun and the card itself is not going above 65c. Any ideas? System: Asrock x58 Xtreme - Last BIOS (1.80) EVGA Geforce GTX 285 w/ latest nvidia drivers (Connect via DVI1) Dell U2410(Set to 59hz refresh 1900x1202 -although I believe benchmarks run 1200x1024) Windows 7 Ultimate 64 12Gb DDR3 1600 RAM

    Read the article

  • External hard drive encryption

    - by Kragen
    I've got a complete backup of my main PC on 1.5 TB external hard drive that I carry around with my laptop so I can have access to all of my files while I'm on the move, however it has just dawned on me that if someone nicks my external hard drive they now have access to everything! Hence I'm looking for a way to encrypt my external hard drive. I'm after something that is: Secure (if I need to carry around a USB dongle to keep the key on so be it) Fast (the performance of the drive should still be reasonable) Cross-platform (I regularly use other peoples computers - Sometimes they are not windows based and might not even have internet access, however I still want to be able to access my files) Cheap (preferably free / open source!)

    Read the article

  • video streaming infrastructure advice

    - by Alchemical
    We would like to set-up a live video-chat web site and are looking for basic recomendations for software and hardware set-up. Most streams will be broadcast live from a single person with a web cam, etc., and viewed by typically 1-10 people, although there could be up to 100+ viewers on the high side. Audio and video do not have to be super-high quality, but do need to be "good enough". The main point is to convey the basic info in the video (and audio). If occasionally the frame-rate drops low and then goes back to normal fairly soon, we could live with that. Budget is an issue, so we are in general looking for a lower cost solution that will give us most of what we need in temers of performance and quality. We are looking at Peer1 for co-lo. The rest of our web site will be .Net / Windows platform. We are open to looking at any platform for the best streaming solution, although our technical expertise is currently more on the Windows side.

    Read the article

  • windows server 2008 vs ubuntu 11 [closed]

    - by user472875
    I am working on implementing a custom server application that should be capable of handling a very large volume of traffic. I am aware that this type of question has been asked a lot, but I haven't been able to find a good answer. What I'm really looking for is for a server with given specs which OS will be able to handle a larger traffic faster and more reliably. I do not care about rights management or any other features. I am fairly good with both platforms, and so I would like to pick the OS with better performance on a clean install, and with nothing else running. Thanks in advance.

    Read the article

  • alternatives to altaris/symentec SVS

    - by The Journeyman geek
    I use an old copy of altaris SVS (link to lifehacker - symentec gutted the old altaris website) for testing applications - it lets me capture the changes one executable makes, or changes to a whole system into a bundle, and remove the whole bundle once i'm done. However symentec bought over altaris and apparently killed off the product, or merged it into some other bigger product. I'd like the same fine grained control SVS had, something that just does this, and preferably free (for personal use) or not too expensive. I'd like this to run on at least windows XP 32 bit, but windows 7 and 64 bit support would be nice too

    Read the article

  • What does the "Maximum Frequency" number mean in the Windows Resource Monitor?

    - by nhinkle
    In the Windows Resource Monitor's CPU tab, there is a status box and graph for the "Maximum Frequency", right next to the "CPU Usage" values. What does this mean? The value is sometimes over 100% on my system... what could that imply? By looking at CPU-z's real-time report of the processor's clock speed, it seems to be loosely related to what frequency the CPU is running at, which would imply that it means "percent of maximum possible frequency the CPU is running at"; this would be of relevance on systems with SpeedStep and/or TurboBoost technology (or similar). Furthermore, setting the system to "power saving mode" lowers the "maximum frequency" value to around 60%, while setting it to "high performance" mode sets it to around 110%. However, the percentage does not seem to exactly correlate to the CPU speed being shown. What value is this actually representing then?

    Read the article

  • XPP-32 over W7-64 on music production laptop

    - by quarlo
    I need to upgrade my laptop and need high performance for music production (recording and mixing). My audio interface manufacturer seems to be unable to successfully convert their drivers to 64-bit. I do not trust a virtual machine to handle real-time audio recording at low enough latency so ... I would like to install XP Pro 32-bit on a separate partition and dual boot since most of the machines that can handle this application now ship with Windows 7 64-bit flavors. I'd like to transit to 64-bit over time assuming M-Audio does eventually get a handle on 64-bit drivers, but really need to ensure that I can stay at 32-bit for now. Does anyone have any experience with this or something similar?

    Read the article

  • Cheap windows shared hosting service

    - by Elangovan
    Hi I have recently purchased a domain from go daddy, now I am looking for a cheap windows hosting. Following are my requirements Shared Windows hosting Should be cheap in price Should have at least one SQL server Db and one mysql db. Should support atleast asp.net 3.5 and php Will be good if it has support for asp.net mvc (no problem even if it is not available also) Should be able to install third party blog sites. Bandwidth, total space and performance are not very important. Silverlight is also an added advantage (no problem even if it is not available also). There should be no advertisement or banner added by the hosting company in the site. Should have support for subdomains

    Read the article

  • Drivers for Quick Cam pro 9000 on Windows 7

    - by runaros
    I have a Logitect Quick Cam Pro 9000 that I want to install on my Windows 7 installation. The user manual specifies that I have to install the software before I install the camera. But my experience is that software from hardware vendors tend to decrease computer performance, so I was wondering if this camera will work by only plugging it in and letting Windows 7 find and install the drivers for it. I could've just tried installing it of course, but again, hardware vendors are notorius for fucking up things, and I wouldn't want to make it impossible to install the camera by doing things in the wrong order, hence the question here on SuperUser because I assume somebody is more knowledgeable than me on this subject.

    Read the article

  • EC2 out of space on root disk, moving it to ephemeral

    - by Joseph Misiti
    I am spawning a few test servers on ec2 that happen to be m1.larges. I am using these test servers for load balancin testing. Anyways, most of the servers I have used before have been backed by EBS, but these instances (ubuntu 11.04) obviously come with a lot of ephemeral space located @ /mnt. What I noticed that is happening is I am running on space on the root disk. I am trying out this tutorial http://www.turnkeylinux.org/docs/using-instance-storage moving my /home + /usr directories to /mnt and then remounting them. This works except it does not survive a reboot. Am I missing something here or is this tutorial not completely correct. How do I make space on my / drive so I can do stuff and survive re-boots.

    Read the article

  • Unable to run VMs on hyper-v

    - by PRAWAT-DS
    Folks/Mates, I need some advise and assistance regarding the testing of Hyper-V. Here is my h/ware configuration: 1) Intel i5 processor (i5-750) 2) Intel M/B DP55WB 3) 6 GB DDR3 RAM OS = Server 2008 R2 Standart (evaluation copy). I installed 2008 r2 on my machine and added hyper-v role to it. I created 2 VMs and installed OS. But after finishing the OS installation the VMs are not booting up. After finishing the OS installation, the VM reboots automatically (normal behaviour) and shows "preparing your system for first time" after that it reboots and didn't come online. Few things to notice, when I am running "securable" on my server 2008 R2 OS it shows that processor is not supporting h/ware virtulization, but (since my desktop is dual boot) when I am running "securable" on my windows 7 OS, it shows that process "does" supports hardware virtulization. VT option is already enabled in BIOS. Any help and suggestions are highly appreciated :) Thanks in advance. Pradeep Rawat

    Read the article

  • Why does TeamViewer launch a web server?

    - by OverTheRainbow
    Hello I just read this article which I find concerning: "After 90 minutes of troubleshooting, Nash traced the problem to TeamViewer, which he used to remotely administer the client's servers. It turns out the program had opened up its own webserver on the client's machine as soon as Apache went down and in the process made it impossible for the client, a large provider of business software, to restart its proper website." http://www.theregister.co.uk/2009/07/24/teamviewer_snafu/ I started using TeamViewer recently because it's just much better than UltraVNC (runs on both XP and Vista, the server dials out instead of requiring the user to open a port on their router + better performance than UVNC)... but I thought that TeamViewer had both the server and client connect out to TV's web server for data to flow between the two hosts. Why does TV even need to open a web server on the server host? Thank you.

    Read the article

  • size of extent on LVM2

    - by piotrek
    in LVM1 there was a limit of 65k extends. So size of extent had to been chosen carefully between wasted space on partitions (to big extent) and maximal possible size of logical volume (too small extent). in lvm2 (according to http://docstore.mik.ua/manuals/hp-ux/en/5992-4589/apa.html) the limit is ~16 million extents. so the default size of 4mb gives ~60TB of LV size. so is there any point in making the extent larger than 4-16mb on a desktop? is there any performance degradation or other costs of having big number of extents?

    Read the article

  • NTFS write speed really slow (<15MB/s) on Ubuntu

    - by Zulakis
    When copying large files or testing writespeed with dd, the max writespeed I can get is about 12-15MB/s on drives using the NTFS filesystem. I tested multiple drives (all connected using SATA) which all got writespeeds of 100MB/s+ on Windows or when formatted with ext4, so it's not an alignment or drive issue. top shows high cpu usage for the mount.ntfs process. AMD dual core processor (2.2 GHz) Kernel version: 3.5.0-23-generic Ubuntu 12.04 ntfs-3g version: both 2012.1.15AR.1 (Ubuntu default version) and 2013.1.13AR.2 How can I fix the writespeed?

    Read the article

  • Server 2008 Hard Faults

    - by claw
    Hey all, plase bear with me as I haven't looked at a server in a very long time. The problem I am having is with a Windows 2008 Standard FE Service Pack 2 Intel Xeon X3430 @ 2.40 2.39 GHZ 4 GB Memory 64 Bit There seems to be no problems other than the physical memory peaking at 91%, always with over 100 Hard Faults Per Second. To my understanding hard faults should be fairly rare on a machine with. Are there any logs I can show you? Or investigate myself. The general performance of the machine is ok, i can access SBS2008 and change settings fairly smoothly without hangs etc. However, we connect to the server and do quite a bit of SQL via an application. For a record to retrieve say 20 rows, it can take 20+ seconds. Thanks in advance, Jamie EDIT: What the server is used for: IIS ASP Web Service SQL 2008 List item Exchange unable to upload screenshots due to low reputation - why doesnt my SO work here :)

    Read the article

  • Worth it to move /var to physical disk vs logical?

    - by Tammer Ibrahim
    Brief question about partition layout. I use an SSD for /, /boot, /usr, & /home partitions. I'd like to move /var to a mechanical disk to minimize writes to the SSD. I'm mainly concerned about maximizing drive life rather than maximizing performance (although I obviously wouldn't want to cripple my server). My mechanical disks consist of two drives sharing LVM, and a third used for nightly rsync backups. I also have a bunch of old 2.5in hard disks lying around. My question is, should I simply create a new LVM volume '/var' on my primary data store, or would it be worth the increased energy consumption (in terms of maximizing the lifetime of the LVMed drives) to install a low volume 2.5in disk to use just for /var? On a more general level my question is about the trade offs of placing OS mounts on the same physical volumes as my data. Thanks for any help!

    Read the article

  • Toshiba Satellite error 10053A0000 when re installing windows xp home on an existing windows 7

    - by Jayapal Chandran
    I had installed windows 7 for testing. Now i want to re-install windows xp home original. I am using the toshiba installation(recovery) disk. The installation process asked a few questions. I selected the option to retain other partitions and to delete only c drive. In the next step i got this error. http://web1.toshiba.ca/support//techsupport/tsbs/all/-tsb001404.htm So, what should i do to retain my files in d drive and only allow the installation to delete c drive?

    Read the article

  • Unmounting a zfs pool while it is shared with sharenfs

    - by Ted W.
    I have a Solaris (open indiana) system which is getting poor disk write performance. In order to enable ZIL in this version of zfs I need to add a line to /etc/system. This will not take affect until I've unmounted and remounted the zpool. The trick is that this spool is shared via nfs to about 200 other servers to host users' home directories. I can guarantee that no users will be accessing the disks during this period of maintenance but I would like to avoid having to issue an unmount for 200 systems in order to unmount the disk on the Solaris box. My question is, with sharenfs, is it necessary to have all systems disconnected before unmounting the filesystem on the host? If it's possible, how do you go about it? I've tried unmounting already, the normal way, and it reports the disk is busy. There is no lsof in Solaris and pfiles (I think that's what it was) does not show anything obviously using the mounts.

    Read the article

  • Best choice for off-site backup: dd vs tar

    - by plok
    I have two 1TB single-partition hard disks configured as RAID1, of which I would like to make an off-site backup on a third disk, which I am still to buy. The idea is to store the backup at a relative's house, considerably far away from my place, in the hope that all the information will be safe in the case of a global thermonuclear apocalypse. Of course, this backup would be well encrypted. What I still have to decide is whether I am going to simply tar the entire partition or, instead, use dd to create an image of the disks. Is there any non-trivial difference between these two approaches that I could be overlooking? This off-site backup would be updated no more than two or three times a year, in the best of the cases, so performance should not be a factor to be pondered at all. What, and why, would you use if you were me? dd, tar, or a third option?

    Read the article

  • Slow old notebook Hardy => Karmic

    - by Mailo
    Hi, i have one very slow notebook from year about 2000. On the computer is running icewm with firefox (in this times chromium for testing). My question is if it's good step to upgrade the system to Karmic Koala? I can't install another OS on that. It doesn't have CD-ROM, it can't boot from flash, or network. The new wanted state is little bit faster system for browsing web and copying photos to local NAS. I don't mention hardware configuration, becouse it's real speed is really deep below the paper parameters.

    Read the article

  • Trying to determine the correct number of XFS allocation groups for postgresql server on Linux

    - by HBlend
    I am running a postgres 8.4.5 server on the linux 2.6.33.7 kernel on an 8 disk raid array with an LSI controller. Most of the tables are around 1GB or less. I know that XFS uses allocation groups (AG) to achieve I/O parallelism. My first question is, does this mean that if two tables are in the same AG, all I/O requests are queued to both of them if either is being read from/written to? If so, I assume I would want to spread my tables across as my allocation groups as possible, correct? Wouldn't this ensure that multiple users querying different tables would get the best performance?

    Read the article

< Previous Page | 535 536 537 538 539 540 541 542 543 544 545 546  | Next Page >