Search Results

Search found 18662 results on 747 pages for 'third party tools'.

Page 467/747 | < Previous Page | 463 464 465 466 467 468 469 470 471 472 473 474  | Next Page >

  • Monitoring server room temperature

    - by user640
    I have a small server room with its own AC unit. Recently, the AC died, and the temperature increased from 70 F to 90 F. We rarely go in this room, so I was lucky that someone happened to notice that the fans were running a lot louder than normal as they walked past the door. It looks like I need a way to be notified when the temperature in that room gets too hot. What tools are you using to monitor the temperature in your server room? How does this tool notify you of a problem (email, SNMP, etc). Note: I've read this question on server temperature, but I'm interested in the whole room, not just the inside of a server case. Edit: Thanks for all the great responses so far! Many of these products measure much more than just temperature. What else should I be looking at and why?

    Read the article

  • Git for Application Settings

    - by devians
    I use a lot of tools at work and at home, and im constantly tweaking them in one location or the other. It's somewhat common practice for people to use Git to version their .vim, .vimrc, and other . files, since you can host your config files on github and have the share-ability and all the other advantages that implies. Being able to version and branch my configs sounds like a grand idea, since I'm always messing about with them. I'd like to discuss the best practice for doing this on a slightly wider scope. How would you implement it? Have your configfiles repo in ~/Library/Configs or similar, and symlink the appropriate files? How to handle preference files for Applications, ie iTerm2. These files are recreated every time, so you'd have to symlink 'backwards' and put a link in the repo? rather than symlinking to the repo, since it would just delete the symlink.

    Read the article

  • How to use 3T SAS HDDs which aren't supported by Dell PERC 6i

    - by Sunry
    We bought several 3T SAS hard disks want to use with our Dell R710 server. But we found our R710 can't recognize the 3T hd, and then we dig out the raid card is PERC 6i which doesn't support it. Contacted with Dell that H700 and H800 support 3T hd, thought we can buy a H700 , but the SAS cable is not available in the market right now. And it seems very hard for us to change the raid card, also no suitable tools to tease the special screws. So I want to know if any other method to use these 3T hds instead of putting them in the warehouse? Can I low formatting it to 2TB, and plug it into Dell R710 to use? Or a SAS to USB adapter to use it as USB external disk? I just want to use them by Dell R710 even external.

    Read the article

  • Schedule EC2 instances

    - by mattcodes
    I want to be able to schedule some simple EC2 EBS backed instances (already configured) to start at 8am and stop at 4pm. This is only time I'll be using my integration server. Is there a simple services (paid or not) that I can use to handle this. All I found so far is to buy a cheap VPS at linode or somewhere and install ec2 tools and schedule via crontab, but what a PITA that is to. On the other end is something enterprisey like Rightscale but not my idea of simple.

    Read the article

  • Good documentation tool that is not Latex?

    - by flpgdt
    I am far from being a expert in Latex but I'm ok to document my projects with it. Though I would seldom find people in the corporate word eager to learn latex and going along with the documentation. 99% of the cases they would just ask me the Word version of it. For technical documentation I find less resistance, but still, whenever I start a project with someone not familiar with latex, the starting up is troublesome. That said, latex is a bit of an oversize tool for my needs really. My documents hardly go further from tables, lists, few images and type styles (although I'd love to still be able to produce hyperlinked PDFs). What are other tools there, simpler and with a easier learning curve than Latex, but still PDF worthy and with minimally decent capabilities? It also has to run on windows :( Oh. yeah, MSWord is not an option ;)

    Read the article

  • How to download a url as a file?

    - by Michelle
    A website url has "hidden" some mp3 files by embedding them as shockwave files, as follows: <span class="caption"><!-- Odeo player --><embed src="http://odeo.com/flash/audio_player_tiny_gray.swf"quality="high" name="audio_player_tiny_gray" align="middle" allowScriptAccess="always" wmode="transparent" type="application/x-shockwave-flash" flashvars="valid_sample_rate=true external_url=http://podcast.cbc.ca/mp3/sundayeditionstream_20081125_9524.mp3" pluginspage="http://www.macromedia.com/go/getflashplayer"></embed></span> How can I download the files for off-line listening? I've found two methods: 1. The StackOverflow Method Create a new local html file with just the links eg <a href="http://podcast.cbc.ca/mp3/sundayeditionstream_20081125_9524.mp3">Sunday Edition 25Nov2008</a> Open the file in the browser, right click the link and File Save Link As. 2. The SuperUser Method Install the Firefox addin Iget. (Be sure to use the right version for your Firefox version.) Tools Downloads Enter url in field. Are there any other ways?

    Read the article

  • Delete temporary files from batch script in xp

    - by Keith Bentrup
    I'm looking for a good batch script that would quickly find & clean all the known safe temporary folders/files from Windows (as many variants as possible) machines (e.g. the windows temp folder, all users IE temp folders, etc.). I'm fond of UI tools like CCleaner (over Cleanmgr.exe), but when I'm trying to clean several computers quickly and/or with minimal involvement, it would be nice to have a script. Plus with a script, I could chain several scripts together. Maybe one to then fire up various antivirus and/or malware detectors. Anyone have a good one or can point to a good resource?

    Read the article

  • What's more cost effective, Hosting your web startup on Foss or Windows?

    - by user37899
    Hi, Not coming from the windows world, I'm confused about licensing. I think my knowledge may be out of date. Before we gave up with windows web servers (IIS 2), we used to have to pay Client Access licence's. This worked out quite expensive. Is it cheaper to host 1000's of users on Windows than use Free open source software tools? http://serverfault.com/questions/124329/network-load-balancing-efficience-and-limits. This post suggests I can pay $15 a month, for unlimited users. I certainly hope that this is unbiased view, I am a professional and use the right technology for the right job. I hope i am not feeling the wrath of a windows (or linux for that matter) fanboy. Perhaps a Microsoft certified Licensing person can clear this up. Should i be recommending to startups windows servers and products over lamp? Cheers!

    Read the article

  • Multiple WAPs: Bandwidth, Frequency Considerations

    - by Pete Cresswell
    The router in my LAN closet does 2 and 5 GHz. In the kitchen, I have a single-band 2 GHz WAP, and in the garden shed I have another single-band 2 GHz WAP. All are set to Bandwidth = 40 MHz, Wireless Network Mode = N-Only. The kitchen WAP and the LAN closet router both come up with multiple bars on my smart phone from almost anywhere in the house. The garden shed WAP will register one bar... but only sometimes. The Questions: Are these things in danger of butting heads? Should I re-set them to Bandwidth = 20 MHz? Bandwidth = Auto? Are there any tools that I could use on an Android smart phone, iPod, or WiFi-enabled laptop to make my own analysis?

    Read the article

  • Suggestions? Password & Encrypted Read/Write File like a Mac (.dmg or .SparseBundle) also R/W on Windows, Ubuntu

    - by Jeff Drew
    For years I have used .dmg or .sparsebundle (Encrypted and Password Protected) to safely keep home directory backups on my Mac. Now, I am looking for a similar Full Permissions/Read/Write that maintains an encrypted, and password protected file that it Tri-Platform. I'd like to have the future ability to use it on Mac OS X, Windows 7/8, and Ubuntu (current releases+). I appreciate your recommendations. Thank you. (I like mounting a DMG and having a file directory structure that can be easily maintained and organized. When done, un-mounting the file.) (I've seen Windows tools to open encrypted DMG files? and I will explore these options, but with the desire to also keep the file accessible on on three OSes, someone might have additional suggestions.)

    Read the article

  • A raw dump from a corrupted large file

    - by Masoud M.
    I have a large .rar file inside partion D (Windows7/NTFS). It's corrupted due to bad sector (I think) and when I copy it to another place (External-HDD) the system freezes after 88% of progress. I even tried to copy it with my Ubuntu and same problem occurred. Also I tried chkdsk and it dosen't fix it. I think my last chance is dump that file with a tool which ignores bad sectors and create a raw copy of it. Then I will repair the file with rar tools. But I can not found a tool to raw dump a specific file. (In linux there is dd tool but it dumps all partition and I can not use it) So, Is somebody know a tool to do a raw dump from a file?

    Read the article

  • Why is music tagging software so inconsistent?

    - by Billy ONeal
    Hello :) A few years ago I spent an insane amount of time using the excellent Tag&Rename program. However, I find that for random, inexplicable reasons, some music tools simply disregard my tags, and drop or destroy the album art, or have strange handling around some characters. For example, "AC/DC" is poorly handled by most music players when I use Tag&Rename to write the tags. Is there a piece of software that works like Tag&Rename but is more compatible, or is there a way to ensure Tag&Rename writes more compatible tags?

    Read the article

  • Advice for migrating email server

    - by Chris Adams
    Hi there, I'm planning to migrate a Zimbra server with about 200gb of data from a server hosted in an office into a datacentre, to increase uptime (we've had a couple of outages when our network here started flaking out, and we have people in other countries relying on this server too). However, I'm not sure how best to migrate the data into the data centre without rendering the connection unusable during office hours, because there's far too much to send in over night over the two meg upstream connection we have here. I'm familiar with using tools like nice to stop a long running process degrading machine performance - is there a simple way to throttle a connection between office hours, so the long running transfer doesn't block the pipe, but then opens up outside of office hours to make the most of the bandwidth? I'm aware the alternative here is to simply mail a hard drive to the data centre, but I'd like to avoid doing that if I could. We're using Centos Linux for our servers, in the office and the datacentre, so extra points for an open source linux answer.

    Read the article

  • Laptop loses signal from WiFi router, but mobile phone holds it fine

    - by Anton
    I have an extremely weird issue with my WiFi router. Both Ubuntu & Windows 7 can connect to it fine, but after 5-10 minutes browser (any one) stops opening pages and tells me it cannot resolve host address. But, at the same time, tools like Skype or BitTorrent work without any issues. I can also browse Internet on my mobile phone connected to the very same router. If I reset router it helps, but after 5-10 minutes I see just the same problem... Ubuntu tells me that WiFi signal is lost (mobile sees it), Windows 7 just won't let me browse anywhere. Can anyone give me a suggestion on this, please? Thanks

    Read the article

  • How to determine if a file has been backed up?

    - by Console
    I try to consolidate old drives to new ones of larger capacity. Sometimes files have been renamed, but are otherwise identical. Sometimes an old directory has just a few more files in it than a newer directory with the same name. Sometimes a file has the same name but the size differs. So I often find myself asking the question: Are there any files on this old drive or directory that I haven't already copied to the new drive? I just want to know that I have the files, I don't want to try and sync stuff automatically (Syncing tools tend to just sync, creating duplicate folder structures and other problems, so I prefer to do it by hand). Basically, if an old drive has a file called "foo.bar" ten directories deep, and my new big drive has an identical file called "oldstuff.zip" in the root, I just want a "yes you have it" or "no, unique files exist". Is there a free tool, a script or a quick and easy method (Mac/Unix or Windows) to get the answer?

    Read the article

  • Windows Server 2008 R2 permissions: Users can't write despite being administrators

    - by Matias Nino
    We just set up a new R2 server and created a bunch of local user accounts on it that are part of the administrator's group. We then set permissions on some shares and folders to allow FULL CONTROL to anyone from the administrator's group. However, the users cannot write to these folders when logged on. On some folders on the C:\ they are prompted for consent in order to gain permission to read them. Any ideas? Are there any tools that would help me troubleshoot this? Thanks in advance for any tips.

    Read the article

  • Is memcached pre-installed with Xcode on OS X?

    - by GeneQ
    The answer seems to be yes, according to Apple's documentation. To make sure I'm not hallucinating, I checked Apple's developer tools documentation. Yes, memcached is supported and documented by Apple: Apple Developer Library : memcached(1) Apparently, memcached is installed with Xcode since at least 10.6 onwards. The reason I'm asking this is that on the Web and on SO there are lots people asking how to install memcached on OS X but curiously no one seems to mention that the easiest way to do so is to simply install Xcode via the AppStore (or using a DMG). All the answers given involves using homebrew or some other complicated way to install memcached from source. Is there any compelling reason why the Apple shipped memcached is not good enough? I don't see any advantages of compiling and installing memcached from source since there's an Apple supported version.

    Read the article

  • Where does netstat get the process name?

    - by tjameson
    I am developing a node application and there is an option to set the process title (process name). This only sets it in some tools (like ps and top), but not in htop or netstat. I found this article that explained how most applications do it, but it doesn't change in netstat. That lead me to wonder where those programs are getting the process name. Would they be getting it from /proc/##/cmdline? (## being the PID of the process) I figure messing with things in /proc is a bad idea (and probably not possible), so if this is where those programs are getting it, is there a way to change it?

    Read the article

  • How do I remove office 2008 from OS X when the "Remove Office" tool says no version of office are in

    - by kbyrd
    I have Office 2008 for OS X through Microsoft's HUP (Home Use Program). I want to reinstall, so I need to first remove it. For Office on OS X, Microsoft provides a binary called "/Applications/Microsoft Office 2008/Additional Tools/Remove Office/Remove Office". When I run it, it says it can't find any versions of office on my computer and won't do anything. I know MS-Office for Mac puts many different components in various places all over the system and not just within it's own applications folder, so how to I successfully remove Office so I can re-install?

    Read the article

  • vps like [load] graphs

    - by foober
    I investigated a couple of tools but they were really annoying and not polished. kSar for exampe is supposed to graph sar output, but it doesn't work. There's a perl script around (sar2rrd) that's supposed to convert sar output in rrd format and generate graphs. Doesn't work. (at least it doesn't like the output of "atsar" as per debian/ubuntu package). Tried munin but it wants to mess with http servers, and for some reason it didn't really work, too. It displayed errors in the webpage generated by the http server it put on port 4949. So, is there a simple install and forget tool to generate daily load,cpu,memory,network graphs? It seems strange to me that this problem has not been solved, maybe I'm looking in the wrong places

    Read the article

  • Upload a directory recursively to an FTP server

    - by Nicolas Raoul
    I am writing a Linux shell script to copy a local directory to a remote server (removing any existing files). Local server: ftp and lftp commands are available, no ncftp or any graphical tools. Remote server: only accessible via FTP. No rsync nor SSH nor FXP. I am thinking about listing local and remote files to generate a lftp script and then run it. Is there a better way? Note: Uploading only modified files would be a plus, but not required

    Read the article

  • Ubuntu make symbolic link between new folder in Home to existing folder

    - by Fath
    Hello, To the point. I have Ubuntu Maverick running on my Lenovo G450. Before, it was Windows 7. All my data are inside another partition, its NTFS. FSTAB line to mount that partition : /dev/sda5 /data ntfs auto,users,uid=1000,gid=1000,utf8,dmask=027,fmask=137 0 0 Inside /data there are folder Musics, Graphics, Tools, Cores, etc. If I'm about to create new folder, let see, GFX on /home/apronouva/GFX and make it link or pointing to /data/Graphics, how do I do that ? So when I open /home/apronouva/GFX the content will be the same as inside /data/Graphics .. and whatever changes I made inside GFX, it will also affect /data/Graphics I tried : $ ln -s /data/Graphics /home/apronouva/GFX it resulted : error, cannot make symbolic link between folder Thanks in advance, Fath

    Read the article

  • Server suddenly running out of entropy

    - by Creshal
    Since a reboot yesterday, one of our virtual servers (Debian Lenny, virtualized with Xen) is constantly running out of entropy, leading to timeouts etc. when trying to connect over SSH / TLS-enabled protocols. Is there any way to check which process(es) is(/are) eating up all the entropy? Edit: What I tried: Adding additional entropy sources: time_entropyd, rng-tools feeding urandom back into random, pseudorandom file accesses – netted about 1 MiB additional entropy per second, problems still persisted Checking for unusual activity via lsof, netstat and tcpdump – nothing. No noticeable load or anything Stopping daemons, restarting permanent sessions, rebooting the entire VM – no change in behaviour What in the end worked: Waiting. Since about yesterday noon, there are no connection problems anymore. Entropy is still somewhat low (128 Bytes peak), but TLS/SSH sessions have no noticeable delay anymore. I'm slowly switching our clients back to TLS (all five of them!), but I don't expect any change in behavior now.

    Read the article

  • Intel site says VT-x is supported on my CPU, but tests say otherwise

    - by Anshul
    I have a laptop with an Intel 2nd Gen. i7-2729QM (http://ark.intel.com/products/50067/) and the Intel site says that it supports VT-x, but I've downloaded the Intel Processor Identification Utility (http://www.intel.com/support/processors/tools/piu/sb/CS-014921.htm) and the test says that my CPU does not have VT-x. There is no option in my BIOS that allows me to enable/disable VT-x. I researched my laptop model (SAGER NP8170) and most forums say that it's enabled by default and there's no option in the BIOS. So assuming that's true, what gives? I also downloaded another tool called SecurAble from GRC and it also says that my CPU doesn't support VT-x. VirtualBox also says that my CPU does not support virtualization. My mind is boggled by why it says on the Intel site that my CPU supports VT-x but all other tests show otherwise. Anyone know what's going on? Thanks in advance.

    Read the article

  • PDF to HTML - batch converter - most reliable and accurate free AND paid for software?

    - by Rob
    I'm look for either a free or paid-for (about 50$/40pounds) BATCH PDF to HTML converter to convert several PDF files at once. Needs to be able to handle vectored and bitmap images within the file, outputting both as jpegs referenced by the html pages. I've tried iorigsoft paid-for PDF to HTML - problems it seems to hang or just go idle, and the stuff it actually converts have broken links - the wrong name is used for constituent chapters as html. Also tried application from intrapdf.com but this crashes near the beginning of the conversion, consitently. Looked at opensource tools but they look equally flakey or use old PDF versions. Need it on Windows 7 32bit home. Thoughts?

    Read the article

< Previous Page | 463 464 465 466 467 468 469 470 471 472 473 474  | Next Page >