Search Results

Search found 2042 results on 82 pages for 'average'.

Page 39/82 | < Previous Page | 35 36 37 38 39 40 41 42 43 44 45 46  | Next Page >

  • What is your IT-department to staff ratio?

    - by Bart Silverstrim
    This is a topic that may be rather political but it related to system administration still...how many people do you have supporting your users in your IT department? We have three people supporting something near 1500 systems and around 2400 people and a recent question was raised about why we have long times to get systems replaced, repaired, work orders fulfilled, etc. along with the system admin duties of maintaining the servers (mail, proxy, filters, VOIP, wireless, backup, etc.) Does anyone have numbers on what the "best practices" average is of support people to employees in an organization? EDIT:...from some of the answers that are appearing I'm starting to feel depressed...

    Read the article

  • Can't connect to domain computers until reboot

    - by thealliedhacker
    I have a domain with about 300 Windows 7 and XP machines, with the domain controllers running Server 2003. Sometimes, I lose the ability to communicate/authenticate with some of the machines until I reboot my computer. This also happens from other computers and regardless of user account and operating system. In other words, say I'm on ComputerA, and I can't connect to ComputerX. I can go to ComputerB and connect to ComputerX, but ComputerB may not be able to connect to ComputerY. If you reboot ComputerA, then it will be able to connect to ComputerX again. Here are some messages from various utilities: sc: [SC] OpenSCManager FAILED 1722: The RPC server is unavailable. mmc (compmgmt.msc): Computer (computer name) cannot be managed. The network path was not found. explorer (\\computer): Windows cannot access \\(computer name). ping: Packets: Sent = 4, Received = 4, Lost = 0 (0% loss) / Average = 1ms

    Read the article

  • Recommended Setup

    - by Chris Ryan
    I have been running into issue with my MSSQL Database setup with speed. Here is my scenario. About 100M Rows Average: 1k Updates Per Second Hard Drives: RAID 10 SSD MDF --Active Time: 0 Log Drives: 1 SSD LDF - Simple Recovery --Active Time 99.9 --Queue: 8 I do not need a back up of the log so it is set to simple recovery but my bottleneck is still at my log. I get high WAITLOG times and thus it can not update any faster. I can't do bulk updates/transactions and each update needs to be one at a time. Is my only option to increase write performance of the log drives, add a RAID drives? Any suggestions on increasing the performance?

    Read the article

  • Internet keeps getting Disconnected & Reconnected

    - by Paul
    I have an Internet connection thru a Phone company which uses DSL modem to connect. For past few weeks my net connection keeps getting disconnected for few seconds & it reconnects automatically. I asked the service provider & they checked & even replaced some wires but of no help. Funny thing is that sometimes there seems to be a rhythm where it disconnects for an average of 13 seconds or so plus or minus few seconds, for hours at a time. I don't know how to attach an image to this post where it shows the pattern of connect & disconnect. I have taken an image using snagit to show the timing pattern. If someone can explain it, I can attach it here. Thanks

    Read the article

  • Downloading movies with Bit torrent

    - by Quintin Par
    I come from a part of the world where the average Hollywood movie is released a year + later than in US. For example: Transformers: Revenge of the Fallen will be released only 5-6 months down the lane. The Reader, I think will not be released at all. Is this a right justification for me to download movies from the torrent networks and watch them? Even if I don’t share it? P.S I don’t think my country has all these American type cyber laws to catch me. It’s just the guilt feeling. Note: SuperUser is not a legal resource.

    Read the article

  • Monitoring over Time with Nagios: How?

    - by David
    Nagios in its standard usage monitors with point-in-time checks: either something is - or is not - true. Other tools like SGI's PCP, HP's MeasureWare, and SEC provide monitoring over time - monitoring things like average disk access time over the last five minutes, or other similar items. Is there anything like this for Nagios? I'm already running NDOUtils, which seems like a natural source for such data. I'd like to have something that would monitor and fire off alarms based on a time-based check using historical data. Is there anything like this for Nagios?

    Read the article

  • Torrent download (µTorrent and FlashGet) suddenly got limited speed!

    - by el
    Both µTorrent and FlashGet are being throttled and I don't know why or how. I use both and they usually report a download speed of 70kbps for torrent speed. FlashGet would report a speed of 70kbps yet actually download at another speed which is about the same as the actual which is a consisten 70kbps. Now, all of a sudden, my torrent downloads are only up to 20kbps. FlashGet still shows a reading of 70kbps, maybe even 80kbps. But, the average actual speed of download is really only 20kbps. I am using Windows 7 on my laptop and no matter where the internet is coming it's the same thing. I have no idea what could be limiting my download rates.

    Read the article

  • Samsung syncmaster SA300

    - by lee
    I recently bought the above moniter. i am now using the DVI port in the moniter but the picture quality stil seems average like the same when i was using the VGA port. THere hasnt been a noticeable change. I am using the DVI output from my graphics card but i thought the picture quality should be alot more High Def than what it is. im just really disappointed in the out come as was expecting nearly High Def quality from my new moniter but just seems a bit normal. Could i need to change any settings?

    Read the article

  • wifi routers and concurrent devices

    - by Joelio
    We have a Linksys WRT54G WiFi router in our office which was working great when we had 5-6 folks. Now on peak days we have 10-15 people, each with a computer, smartphone, etc, and an ooma VOIP device. On average 1-2 times a day I need to go hard reboot the router, and sometimes the border router (Cox-supplied device). I assume this is just because the router cant handle this many concurrent users. So my question is can these consumer routers handle this kind of load? If not, would adding more devices solve the problem, and how close proximity can I put 2 routers without having interference problems (our office area is not that big physically)?

    Read the article

  • How to display a graph only for business hours with CACTI?

    - by Blast Raider
    I have noticed that I can only display an uninterrupted period with Cacti. I am wondering whether is possible or not to make a custom graph which displaying only the business hours during a period (a week, a month, etc.). If it is possible, how could I configure it on Cacti ? For example, I would like to be able to display a graph with an average inbound/outbound traffic between 8am and 7pm for 5 business days a week of the last month. I would apreciate any help. Thank you.

    Read the article

  • Please recommend tools for PC, browser, home network performance problems?

    - by mobibob
    My client is experiencing some odd response behavior in their browser for the past few days. Classic, "nothing has changed" so I am starting at ground zero. Browsing a website will timeout or take a ridiculous time to load -- other times, the same site and query is immediately responsive. Once a connection is established, video streams are uninterrupted. The home network hosts a website, but it is not experiencing any activity in Apache's 'access.log' I am using speedtest.net to check if the ISP through the internet is 'OK' -- which looks typical (average +/-). I have to suspect the home network is beaconing or something very abnormal, but I don't know where to start.

    Read the article

  • Process of carrying out a BER Test

    - by data
    I am subscribed to an ISP supplying a 3meg ADSL line. Lately (for the last 4 weeks) speeds have dropped from the usual average downstream speed of ~250kbps to just 0.14Mbps (according to speedtest.net) and employees are complaining about lack of access to the server. I have been calling customer support and logging calls for the last 3 weeks, but they have been unable to determine the source of the problem other than carrying out a few bitstream tests and checking the DHCP renewal times. I am going to call back and suggest carrying out a BER test. What type of equipment is needed to carry out this test? I have access to a wide range of Cisco networking equipment. Other: We don't need a leased line as there are less than ten employees.

    Read the article

  • How much memory should my rails stack be consuming?

    - by Hamish
    I am running my own webserver on a 384MB VPS from Slicehost to serve two Ruby on Rails applications on separate Virtual Hosts. I am running Phusion Passenger with Apache2. The following is the contents of my Passenger.conf <IfModule passenger_module> PassengerRoot /opt/ruby-enterprise-1.8.6-20090610/lib/ruby/gems/1.8/gems/passenger-2.2.11 PassengerLogLevel 0 PassengerRuby /usr/local/bin/ruby PassengerUserSwitching on PassengerDefaultUser nobody PassengerMaxPoolSize 3 PassengerMaxInstancesPerApp 2 PassengerPoolIdleTime 300 # Ruby on Rails Options RailsAutoDetect on RailsSpawnMethod smart NameVirtualHost *:80 If i do a 'top' on my server I have 314MB used on average, this seems like too much? Am I mistaken and if not what possible steps can I take to reduce the Memory usage? Thanks!

    Read the article

  • Torrent download (utorrent and flashget) suddenly got limited speed!! Help, how to fix it?

    - by el
    I was using utorrent and flashget... and usually the flashget download speed or torrent speed gets up to 70 kbps...specifically in flashget it would show that its speed is 70 kbps and while downloading it would shouw another speed (aeverage speed) which is usually about the same with the actual considering it is consistently 70 kbps......but all of a sudden i dunno what happened....im using win 7.........all of my torrent downloads are only upto 20 kbps...and in flashget.....although the speed that it reads gets to 70 or even 80 kbps....its average speed or the downlolad speed its using is only 20..........I am using a laptop and no matter where is the internet coming from its the same thing...i dont know what is limiting it....please help me, thanks.

    Read the article

  • Why does my wifi fail to stay connected?

    - by Mark0978
    We have 5 different laptops in the house, 1 macbook, 1 Compaq, 3 dells, plus numerous little devices like ipods, squeezeboxs, squeezebox radios and controllers, you name it. All of them connect to and stay connected to the "downstairs" WIFI just fine. But my laptop (the most expensive of the lot) refuses to stay connected, preferring the much weaker and slower signal upstairs. My laptop is running Win7 (64 bit) Ultimate with an Intel 3945ABG WIFI contoller (It's a Dell Precision M6300). The WIFI Access Point is a Cisco WRT54G2 V1 with firmware: Firmware: DD-WRT v24-sp2 (10/10/09) micro Time: 11:11:10 up 42 min, load average: 0.06, 0.01, 0.00 WAN: Disabled

    Read the article

  • Does SQLIO lie when run from a Hyper-V guest on a VHD?

    - by ScottStonehouse
    SQLIO seems like a useful tool. I thought it would be interesting to try to measure the speed difference between a physical disk and a VHD. So I ran SQLIO on the Hyper-V host on the physical drive. Results seemed reasonable. Then I ran it from the guest to test the vhd (on the same physical disk). I expected it to be a bit slower. But instead it was way faster - like 0ms average latency. So I'm trying to learn something here. It seems like hyper-v is fooling SQLIO somehow but I don't understand it well enough to figure it out. It's a dynamic vhd, no snapshots or anything, and the vhd is the only file on the disk. The physical disk is actually a two SAS drive RAID 1.

    Read the article

  • Is there anything that can be done to make X forwarding over LAN less painful?

    - by Earlz
    I have two *nix machines that I like to remotely do graphical things sometimes with. One of those machines is my Arch Linux media server, and another is my OpenBSD router/general server. My current task was installing a Windows XP over virtual box on my media server. I preferred to do the installation from my main workstation so that I could continue to watch TV from my media server on it's "actual" display. I soon noticed that X forwarding is quite painful though, even over low-latency LAN connections. Are there any settings I can tweak to make it a bit less painfully slow, jittery, and laggy? My LAN is rather small, 100Mbit, and ping times to the other box average at about 0.15ms. Also, I know Virtualbox has a VNC mode it can use instead of doing X forwarding, but this particular task was just an example

    Read the article

  • When to use Nginx PHP Fast CGI with a TCP socket instead of a UNIX socket?

    - by user64204
    I've followed this guide to setup PHP in FastCGI mode with Nginx. This guide describes 2 ways of doing it: TCP socket and UNIX socket. I've ran some Apache Benchmark on my locale machine and here are the results: Below tests ran multiple times to get better average statistics: $ ab -c 200 -n 100000 http://.... APACHE: 1800 req/sec NGINX (TCP socket): 2500 req/sec NGINX (UNIX socket): 15000 req/sec As far as I understand, there is overhead with using a TCP socket rather than a UNIX socket, hence the better performance with the latter. However I was not expecting such a performance difference given that the TCP socket is on the localhost, and therefore would like to ask the following question: Q: Given the huge performance gain with using a UNIX socket, what are the configuration scenarios where it would make sense to use a TCP socket instead?

    Read the article

  • failure daemon and changing pid number

    - by Alessandra Bilardi
    proftpd, sshd and apache processes run with /etc/init.d/its-script on linux distro. I was monitoring 21, 22 and 80 ports with farm monitoring service: every 5 minutes service check each port and notify only failure. The failures were 5-6 times on 24h. It seems that someone kicks the switch sometimes.. I add monit and collectd monitoring and the monitoring about 21, 22 and 80 ports is every 1 minute. I do not receive farm monitoring service notify. I receive only monit notify about failure and/or succeed/changing pid number of proftpd, sshd or apache process. The failures are still 5-6 times on 24h. collectd monitoing about cpu, load average and each process is regular and there are no peaks. There is nothing kicks the switch but there is something which determines failure monitoring. is it a simple interference or is it indicative of some abnormality? What could cause these failures?

    Read the article

  • How to calculate bandwidth limits per user on WiFi network

    - by Lars
    A typical 802.11g access point can provide around 25 Mbps of bandwidth. How is the bandwidth shared among the users? Furthermore, how many users can be served by a single access point using 802.11g in an environment with low interference, and average web activity from the users? The goal is to use bandwidth limitation to avoid starvation for some users in case some of the users start to download a file or stream HD video or some other bandwidth intensive activity. Can someone break down the math on this?

    Read the article

  • Regarding temperature in our server cabinet

    - by kafka
    We have a floor standing server cabinet, which sits in the corner of our office. There is no AC in the office, and the average temperature is around 21C (70F). The server cabinet currently contains 2 tower servers, a switch and a router. It's well enough ventilated, but does get understandably warm in there. We've recently put a thermometer in there to monitor the temperature, and recently I've noticed the temperature hit 33C (91F). What I'd like to know is if this temperature is too high, and whether there is a maximum temperature we should not be exceeding?

    Read the article

  • negative time for a ping echo reply

    - by Mario
    I was happily pinging in the net when suddenly: Pinging X with 32 bytes of data: Reply from .235: bytes=32 time=1444ms TTL=41 Reply from .235: bytes=32 time=1767ms TTL=41 Reply from .235: bytes=32 time=1531ms TTL=41 Reply from .235: bytes=32 time=-1198187ms TTL=41 Ping statistics for .235: Packets: Sent = 4, Received = 4, Lost = 0 (0% loss), Approximate round trip times in milli-seconds: Minimum = 1444ms, Maximum = -1198187ms, Average = 1073443462ms A negative time for the reply. I checked wireshark and it had the same values: Time delta from previous displayed frame: -1198.187867000 seconds Time since reference or first frame: -1179.935038000 seconds I didn't change the time of the machine while pinging. This was made in my local network at home, from a XP VM in windows 7. So I blame the VM (virtualbox). But I was wondering if this strange behaviour (to me) could have a reason, or if any of you have seen this before. Thank you bye.

    Read the article

  • Maximum MTU size

    - by user192702
    Thought one of the issues I'm experiencing with the following question is due to MTU rightfully so. ESXi 5 VM Putty session hangs, vSphere client timing out However, when I tried testing the maximum MTU size it seems there's just no limit. Thought Ethernet only allows maximum MTU. But I'm up to 54450. ping -l 54450 192.168.10.7 Pinging 192.168.50.7 with 54450 bytes of data: Reply from 192.168.10.7: bytes=54450 time=1081ms TTL=62 Reply from 192.168.10.7: bytes=54450 time=1079ms TTL=62 Reply from 192.168.10.7: bytes=54450 time=1079ms TTL=62 Reply from 192.168.10.7: bytes=54450 time=1079ms TTL=62 Ping statistics for 192.168.10.7: Packets: Sent = 4, Received = 4, Lost = 0 (0% loss), Approximate round trip times in milli-seconds: Minimum = 1079ms, Maximum = 1081ms, Average = 1079ms

    Read the article

  • SQL Server 2008R2 Express: which is the users limit in a real case scenario?

    - by PressPlayOnTape
    I know that sql server express has not a user limit, and every application has a different way to load/stress the server. But let's take "a typical accounting software", where users input some record, retrieve some data and from time to time they make some custom big queries. May someone share its own experience and tell me which is the limit of users that can realistically use a sql server express instance in this scenario? I am looking for an indicative idea, like (as an example): "I had a company with an average of 40 users logged in and the application was working ok on sql server express, but when the users become 60 the application started to seem non repsonsive" (please note this sentence is pure imagination, I just wrote it as an example).

    Read the article

  • How to put 1000 lightweight server applications in the cloud

    - by Dan Bird
    The company I work for sells a commercial desktop/server app that runs on any non dedicated Windows PC or server and uses Tomcat for all interactions with the application. Customers are asking that we host their instance of the application so they don't have to run it locally on their own servers. The app is lightweight and an average server, in theory, could handle 25-50 instances before users would notice a slowdown. However only 1 instance can run per Windows instance (because the application writes to a common registry branch) so we'd need something like VMWare to create 25-50 Windows instances. We know we eventually need to reprogram to make it truly cloud-worthy but what would you recommend for a server farm or whatever for this? We don't have the setup to purchase our own servers so we must use a 3rd party. We have budgeted $500 - $1000 per year per customer for this service. Thanks in advance for your suggestions, experiences and guidance.

    Read the article

< Previous Page | 35 36 37 38 39 40 41 42 43 44 45 46  | Next Page >