Search Results

Search found 594 results on 24 pages for 'serverfault'.

Page 12/24 | < Previous Page | 8 9 10 11 12 13 14 15 16 17 18 19  | Next Page >

  • Very high memory usage, but not claimed by any process?

    - by SharkWipf
    While stress-testing LVM on one of our Debian servers, I came across this issue where memory would fill up a lot to the point where it would run the server out of memory, but no process would claim the memory. See http://i.imgur.com/cLn5ZHS.png, and see http://serverfault.com/a/449102/125894 for an explanation on the colors used in htop. Why is this happening? And is there any way to see what process is using the memory? Htop is configured not to hide any processes, so what is it that htop is missing? In this particular case, I can fairly certainly say that it is caused, directly or indirectly, by lvmcreate, lvmremove or dmsetup, as I was stress-testing that. Do note that this question is not about solving the LVM problem, but about why the memory isn't claimed by any process. Stopping all LVM commands does bring the memory back down to <600MB.

    Read the article

  • rsync without password, none of google (server fault) tutorials worked

    - by Jake Armstrong
    I need to use rsync for a daily backup operation and in the past (on different servers) I managed to just use a rsa key etc, but now none of google (serverfault) tutorials work at all. It keeps asking me for a password. I have webmin and ssh/root access to both servers. My steps: create a key on server 1 send key.pub to server 2 add key.pub to .ssh/authorized_keys chmod 700 .ssh/authorized_keys go back to server 1 and try rsync and it keep asking for password... rsync command: rsync -avz -e ssh file.txt root@server2:/root EDIT: well, I cleaned up everything and this time, instead of inserting a custom name to the key I used the standard one on server1. sent the .pub to server2 and it worked as a charm... So the answer is that server1's ssh wasn't even using the right key...

    Read the article

  • Activating Sharepoint Server Publishing Infrastructure returns "Column Limit Exceeded"

    - by Haydar
    I am getting the following error when attempting to activate the Sharepoint Server Publishing Infrastructure. There are only about 140 columns defined in the entire site collection. I had previously activated and subsequently deactivated this feature. I don't know whether this might be causing the problem. But I really have no idea. I am a developer and this is my first question on serverFault. Please be gentle. Error Message: Column Limit Exceeded. There are too many columns of the specified data type. Please delete some other columns first. Note that some column types like numbers and currency use the same data type. Troubleshoot issues with Microsoft SharePoint Foundation. Correlation ID: 2ec9ae3a-5464-4922-9760-874099f51420 Date and Time: 11/8/2010 3:43:44 PMDate and Time: 11/8/2010 3:43:44 PM

    Read the article

  • Remote access and local access same hostname

    - by cpf
    Hi serverfault, I have a server in a clients network, seperated from theirs with a router/firewall, the intention is to have this server available through one hostname (example.com) My idea is to have (at least) a DNS server in the outside, to have outside (out of the clients' network) access the internal server. The problem would at that point be the internal client (PC A) My question: What would I have to do to make something like this work? Is it even possible or already done? The goal is to not have to change anything on either PC A or PC B, while both should access the same "internal server" while surfing to "example.com" Perhaps adding logic to the DNS server would work (Detect the external IP of internal client [PC A] is the same as the IP for example.com - Give the local IP as reply?) Anyhow: Thanks for helping me think on this!

    Read the article

  • Window 7 not recognizing printer for driver (re)installation

    - by user55406
    OK I am having this printer problem. I had a user delete her printer. Please don't ask, I don't know. She is using an HP deskjet 6940 printer, and for her desktop she has an HP dx2400, with Windows 7. When I plug in the USB cable it should install the new drivers, but it does not. I believe you guys don't need to know my network setup. I could be wrong. If you guys here at serverfault could help that will be great. :)

    Read the article

  • Dealing with employees that use Hotspots? [closed]

    - by javasocute
    Recently a consultant revealed to me and my superiors that a particular department in my company had a tremendous decline in productivity over the past year. This department happened to mostly work with laptops and tablets and thus had wireless. Our barracuda did not display any vast amounts of internet entertainment and so I was initially puzzled by his statement. However he explained that more and more employees are using "hotspots" that when not broadcasted, become invisible to our network. So basically these people could connect to hotspots and watch youtube all day and we would never know it. The consultant also said that this was becoming an issue for many companies. So my question is, how can I detect invisible hotspots? I was browsing google and there are several different programs that can scan for such access points. But I am hoping to get a recommendation from the serverfault community. Thank you.

    Read the article

  • Running out of space on a SAS-based workstation

    - by SteveWilkinson
    Hi - first question on serverfault - pls excuse if asked elsewhere. I have a Dell Precision 690 with one 73GB (15000RPM) SAS drive in it. Running Windows 7 (x64) with a bunch of apps means it is running out of space. I have on order a 147GB (15000RPM) SAS drive (same GB/s, same make) which I want to put into the machine. Not knowing anything about SAS, do I need to replace the 73GB with the 147GB and restore from a backup, or can I put the drive in the machine and get the controller to treat the two drives as one large one? (The original paperwork for the machine says 73GB SAS (No RAID) if that is relevant.) Thanks in advance.

    Read the article

  • Running out of space on a SAS-based workstation

    - by SteveWilkinson
    Hi - first question on serverfault - pls excuse if asked elsewhere. I have a Dell Precision 690 with one 73GB (15000RPM) SAS drive in it. Running Windows 7 (x64) with a bunch of apps means it is running out of space. I have on order a 147GB (15000RPM) SAS drive (same GB/s, same make) which I want to put into the machine. Not knowing anything about SAS, do I need to replace the 73GB with the 147GB and restore from a backup, or can I put the drive in the machine and get the controller to treat the two drives as one large one? (The original paperwork for the machine says 73GB SAS (No RAID) if that is relevant.) Thanks in advance.

    Read the article

  • Windows 2008 Enterprise License can't activate Standard Edition

    - by starchx
    We have downloaded and installed the Windows 2008 Standard R2 edition months ago. Now the server is in production. We signed up the Microsoft Partnership Action Pack subscription last week and get a license for Windows 2008 Enterprise edtion. I am trying to activate the standard edition with Enterprise key as advsied here:http://serverfault.com/questions/318968/upgrade-domain-controller-sku-from-server-2008-r2-standard-to-enterprise , but failed. Is it because windows 2008 we have is different? (downloaded from MS site with eval license). Thanks. tim

    Read the article

  • Mail Secure & Stable Open Source Mail Server

    - by Fanar ALHAYALI
    I have asked question on http://stackoverflow.com/questions/9868426/i-need-to-know-which-email-server-i-have-to-use and someone tell me my question would be better on serverfault. I know that this is a common question and asked many times. but there are so many available mail servers that i am not able to decide the one. Kindly tell that which is the Secure, Stable and fast open source mail server for Centos or Redhat Server. Is there any guide which can be used to deploy the mail server with all its components e.g. smtp, pop3, imap, spam, calender server, antivirus, DNS Setting. Currently I'm using sun messaging V6 which installed on Solaris 10 and my boss ask me to make a report for the best mail server today in the marketing? I tried to have a look on Google but I couldn't find interesting information for my report. Any advice would be appreciated.

    Read the article

  • Setup IPv4 local on IPv6 VPS

    - by A.D.
    I have a dedicated server running multiple IPv6 only OpenVZ containers. I want them to be able to communicate with the IPv4 internet, but I realized that isn't going to be possible with IPv6 only. So they need to have an IPv4 address as well, not sure if a local address will work for it, but pretty sure it should. I added 169.254.1.100 in the container .conf file, but when I try to start it, I get this : Adding IP address(es): (the IPv6 address) 169.254.1.100 arpsend: 169.254.1.100 is detected on another computer : 00:04:9b:f2:b0:00 vps-net_add WARNING: arpsend -c 1 -w 1 -D -e 169.254.1.100 eth0 FAILED I did a lot of research, and searched serverfault before posting this, but found nothing relating to this.

    Read the article

  • Updating solr on a ColdFusion 9 install?

    - by Jordan Reiter
    I'm thinking about upgrading the solr install included with ColdFusion 9 to the latest Apache release. This raises a few questions: Is there a compelling reason not to upgrade to 3.6 (is it slower than, more cumbersome than, or backwards-incompatible with 1.4) altogether? The solr install included with CF9 is customized. Is there a way to customize it myself, or to at least fool CF into treating it like its predecessor? Will all of my existing indexes work as-is (are?) with the new version? Has anyone out there on ServerFault done the upgrade? I'm especially interested in hearing about unforeseen or unexpected effects from the upgrade.

    Read the article

  • How do you interpret `strace` on an apache process returning `restart_syscall`?

    - by indiehacker
    We restart an apache server every day because RAM usage reaches its limit. Though of value See this serverfault answer, I dont think lowering the MaxClients in the apache configuration is a solution to the unknown root problem. Can you make sense out of the below data? Below is an extract of what $top with M returns: 20839 www-data 20 0 1008m 359m 22m S 4 4.8 1:52.61 apache2 20844 www-data 20 0 1008m 358m 22m S 1 4.8 1:51.85 apache2 20842 www-data 20 0 1008m 356m 22m S 1 4.8 1:54.60 apache2 20845 www-data 20 0 944m 353m 22m S 0 4.7 1:51.80 apache2 and then investigating a single process with $sudo strace -p 20839 returns only this one line, which is cryptic, for me: restart_syscall(<... resuming interrupted call ...> <unfinished ...> Any insights? Thanks.

    Read the article

  • What are some good asp.net shared hosting pre-sales questions?

    - by P a u l
    I'm not asking for any host recommendations, those are covered in other questions. What are some good pre sales questions for asp.net shared hosting? They never seem to answer all the questions in their feature lists. So far I have a few: dedicated application pool? sql server management studio supported? Is tunneling required? can I reset my application pool in the control panel? are php and perl fully supported as well? are subdomains supported, and will I need a routing script in the root or are they routed automatically? etc. Developers have a critical need for good hosting to stage applications. I think this is absolutely developer related and don't want the question on serverfault.

    Read the article

  • 64-bit Hardware Virtualization on VirtualBox

    - by Cat
    I am trying to set up a SQL Server practice lab using VirtualBox and a trial copy of Windows Server 2K8 R2 in ISO format. I received an error message that states my processor does not support 64-bit hardware virtualization. Although I "Enabled" Intel ardware virtualization in the BIOS that still doesn't work. According to the Intel website, the processor does support VT-x, however the accelerated tab in VirtualBox is greyed out and no references to VT-x are mentioned in the settings options available to me. Any ideas on how I can get around this? I checked around ServerFault and couldn't find anything but if I missed an applicable post a link is great too. Specs below - if additional information is needed please comment and I will provide. Thanks in advance. VirtualBox version - 4.1.18 Hardware - Lenovo B570 1068-A3U i3-2310M

    Read the article

  • postfix: deliver mail to specific emai addresses locally and send the rest to relayhost

    - by sme
    I want to achieve the following with postfix: Almost all outgoing emails shall be sent via a relayhost, configured in main.cf with relayhost = my.mailserver There's one specific email address ([email protected]) for which I want the email to be delivered locally. I set up a local account on the machine to match the email address and added "mydomain.com" to the mydestination property in main.cf. Then of course postfix tries to deliver every email to @mydomain.com locally and bounces them because the respective user doesn't exist. Question http://serverfault.com/questions/149453/hybrid-gmail-mx-postfix-for-local-accounts seems to be somewhat related (though more complicated).

    Read the article

  • How do I know if my disks are being hit with too much IO reads or writes or both?

    - by Mark F
    Hi All, So I know a bit about disk I/O and bottlenecks relating to this especially when relating to databases. But how do I really know what the max IO numbers will be for my disks? What metric might be available to me for working out roughly (but needs to be a good approximation) of how much capacity (if you will) have I got left available in I/O. I've seen it before where things are bubbling along nicely and then all of a sudden, everything screams to a halt, and it ends up being an IO bound problem. Is there a better way to predict when IO is reaching its limits? This article was interesting but not giving the answer I desire. "http://serverfault.com/questions/61510/linux-how-can-i-see-whats-waiting-for-disk-io". So is my best bet surrounding just looking at 'CPU IO WAIT'? There must be a more reactive method for this? Best, M

    Read the article

  • Using Dropbox API instead of a FTP server.

    - by Somebody still uses you MS-DOS
    This is a small aplication scenario. Usually, when you have to do some backups of source code/database on your server, you use a second ftp server, a cronjob to tar.gz your db dumps and source files, and send this file to your ftp server from your application server. Dropbox created an API to use it's infrastrucutre. Since they provide 2gb for free accounts, I thought about being able to upload to it instead of a ftp server. So, if you do some freelance work, you can create a free account for each client and use this approach, maybe encrypting the files you send. You even gain a revision for each sent file, like a revison control system, for free, from the last 30 days. What do you think of this approach? Is it possible? And, more importantly: what are the security risks involved? (That's why I'm asking this on serverfault, since this POV from sysadmins will be more accurate). Thanks!

    Read the article

  • Slow self hosted wordpress website

    - by Integrati Marketing
    Hi All, we have a great site which has been humming along nicely for about 5 months and then in May it went from a page load speed time of 3-5 secs to now an agonising 15+ secs!!! The host has been really helpful and has even shifted the site to a new server which is faster! I guess seeing as though we do not have the insight or your expertise we would ask the Serverfault community and see what this crowd of experts could recommend? Appreciate any insight, thank you. site is here: integrati.com.au Cheers. :)

    Read the article

  • Slow slef hosted wordpress website

    - by Integrati Marketing
    Hi All, we have a great site which has been humming along nicely for about 5 months and then in May it went from a page load speed time of 3-5 secs to now an agonising 15+ secs!!! The host has been really helpful and has even shifted the site to a new server which is faster! I guess seeing as though we do not have the insight or your expertise we would ask the Serverfault community and see what this crowd of experts could recommend? Appreciate any insight, thank you. site is here: integrati.com.au Cheers. :)

    Read the article

  • Nginx config with try_files and rewrite : precedence?

    - by Penegal
    Good morning, everybody. Firstly, this question may have been already asked, but I searched ServerFault during about 15 minutes without finding it, so, if it was already asked, please accept my apologies. I'm trying to rationalize my Nginx server config, but I have a rather dumb question that I couldn't solve, even with extensive Web search, except if I totally f*cked my search. Here is the question : is try_files parsed before or after rewrite ? Asked differently, Do I have to put try_files after all rewrite directives, or is Nginx config parser smart enough to evaluate try_files after all relevant rewrite directives ? The link with the config rationalization is that the answer to this question will change the organisation of the config, ie if config file order of try_files and rewrite changes the config behaviour, it will force me to disperse my includes, some of them containing try_files and other ones containing rewrite, because I also have rewrite directly in nginx.conf. Hoping you can help me, Regards.

    Read the article

  • What would Stack Exchange's yearly expenses be if it were to be using a third party host?

    - by abel
    StackExchange manages it's own servers, as it should, but if SE were to be hosted on a 3rd party "cloud" hosting (like Amazon's), what would it's monthly / yearly expenses be(keeping everything else the same)? A detailed answer comparing it to the bills that Stackexchange boots currently (including power/property/staff) would help. (PS: I know that the blog is a good resource. I also understand that managing your own hosting is almost the same as setting up a hosting company and using it for your own needs. Plus is this a question for meta or does it fit within serverfault's purview?)

    Read the article

  • Override template shell on linux system in Active Directory domain?

    - by benizi
    Is there an easy way to override the Samba "template shell = /bin/bash" setting on a per-user basis? This is for Linux systems joined to an Active Directory domain. Some users want /bin/bash. Others including myself want /bin/zsh. Is there some AD attribute I can set? Anything I've found via googling seems hackish at best (writing a script to replace /bin/sh -- maintenance hassle). A similar serverfault question Override LDAP shell seems OpenLDAP-oriented (but if someone knows how to get it working with AD, please say so).

    Read the article

  • Can I move a copy of Windows 7 from one machine to another?

    - by Chopper3
    This is a pretty basic question sorry but I've looked in the 'related questions' and couldn't find an answer. When Windows 7 came out I bought a copy of Home Premium, the retail box with both 32 and 64 bit versions, to install via Boot Camp on my macbook. I've just got a retail (32 and 64 bit) copy of Windows 7 Ultimate that I'd like to install instead-of/over the Home Premium version and it struck me that I should, at least in principal be able to install the Home Premium version on my son's machine - is this the case and if so is there anything I should do first or once they're installed to let MS know I'm being a good guy about this? Sorry it's such a dim question but I'm a serverfault guy really and know very little about Windows. Thanks.

    Read the article

  • Resolve local subdomain on apache for paths within user dir

    - by MaoPU
    On Apache 2.2.x I've activated mod_userdir. I used the default setup, so that http://localhost/~name/ will be connect with ~name/public_html/ and a path within public_html, e.g. ~name/public_html/mySite can be reached through http://localhost/~name/mySite. How can I achieve, that the same path can be reached through http://mySite.name.localhost/? I don't want a manual approach like it is suggested in other SF questions (such as http://serverfault.com/q/133921/53624), but rather want an automatic mapping of all available paths to the corresponding URL. I think, several steps will need to be taken: Change mod_userdir configuration, so that the subdomain of localhost will be connected with all available user names on the machine. The second step would maybe include the usage of mod_rewrite, so that the subsubdomain could be matched to the path within ~name/public_html... What would be your prefered way?

    Read the article

< Previous Page | 8 9 10 11 12 13 14 15 16 17 18 19  | Next Page >