Search Results

Search found 111890 results on 4476 pages for 'git update server info'.

Page 1955/4476 | < Previous Page | 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962  | Next Page >

  • Ubuntu upstart hangs on interactive start & stop

    - by danorton
    How do I get Ubuntu upstart to not hang on interactive start & stop? I have created many upstart scripts that work fine during init, but often hang when I enter them at the console. If I CTRL+C out, all that happens is that the job changes state. The script is never run. I’m running Ubuntu Lucid on a Xen virtual server with a Linux 2.6.39 kernel. Below is merely a representative example of many scripts that behave this way: description "apache2" start on local-filesystems \ and (net-device-up IFACE=lo) \ and (runlevel [2345]) stop on runlevel [016] respawn respawn limit 10 5 expect daemon script . /etc/apache2/envvars /usr/sbin/apache2ctl start end script

    Read the article

  • Zend Framework 2 loading slow and loss of connection using WAMP

    - by Charlie
    I've been facing an issue with Zend framework running on my local Wamp 2.2 server. I am not sure what I'm doing wrong but ZF2 seems to load really slow when making an http request. Any other request to a php or html file seems to run smoothly. Also, sometimes when the loading time takes longer, I get this message: "The connection to [virtualhostname] was interrupted" I then need to hit refresh to complete the request. I checked apache error log and everything looks fine. Please, I appreciate any type of guide/suggestion to take care of this issue. I followed the starter guide word by word.

    Read the article

  • non-mapped virtual memory & total number of connections

    - by tszming
    We have two MongoDB data nodes (replica set) - Primary & Secondary. I noticed that the non-mapped virtual memory is relatively high and wondering if they are hurting our MongoDB performance (The server usually peaked at around 6-7K queries per sec). In MMS, it was stated: "The most common case of usage of a high amount of memory for non-mapped is that there are very many connections to the database." So we checked the memory usage with db.serverStatus().mem in our Secondary: { "bits" : 64, "resident" : 6846, "virtual" : 416797, "supported" : true, "mapped" : 205549, "mappedWithJournal" : 411098, "note" : "virtual minus mapped is large. could indicate a memory leak" } Note: We are using 2.0.4 and now the default stack size should be 1MB per connection. The current number of connections is around 1.1K, but the non-mapped virtual memory (virtual-mappedWithJournal) is around 5699 MB. The trend is quite stable so I can't say there is a leak here, but where is the memory gone? Any idea?

    Read the article

  • Suggest Cisco Router for a small company of 60-100 Users

    - by user69154
    Trying to find the right router to use - we'd like to go Cisco. We have ~60 programmers that do alot of remote server work. Will grow to 100 programmers by the end of the year. We've talked to companies that sell Cisco routers. While they make recommendations, we feel that the sales guys may be selling us an overkill solution. What kind of company (that is not linked to sales of routers) can provide a service to recommend the right kind of router for our company? There are lots of network engineers out there, but am looking for some company that focuses on advising a network solution. Any suggestions of companies are welcome.

    Read the article

  • Apache Connection vs. Request

    - by user101570
    I apologize in advance if this is a basic question, but I am quite confused after reading the Apache documentation and other tutorials. Does a single Apache prefork process serve all HTTP requests for a given client? That's what I thought, but when I reduce maxclients down to a low number, my page load times go to a crawl. This despite the fact I'm the only client on the server in question. This would suggest each process serves a single HTTP request at a time, rather than serving all requests within the TimeOut window. So if a single webpage requires 15 HTTP requests to load fully, do I require 15 prefork Apache processes to optimally serve it?

    Read the article

  • What are the common linux ( RH ) commands for SAN related activities ? How to check if a LUN is attached to the computer ?

    - by Nishant
    How do I check if a LUN has been presented to my server ? What are the Linux commands for that ? Do the LUNS show up in a fdisk -l command like a normal /dev/sda gets listed ? What are other commands assosicaed with general SAN related checks in Linux ? What is WWN and how does that have any relevance and Also please explain multipathing why if we have LUN's , what is the use of multipathing then ? Bit lenghty but I am not able to get a grasp on the topic . Any help would be appreciated .

    Read the article

  • What is your list of programs to install to Windows after a fresh install?

    - by marked
    I am looking for others' list of program that absolutely must be installed to a fresh install of Windows before going any further. I hope to compile a list here to use as reference for all new Windows installs/restores. Automating this list of programs installation is the next step. Some great lists so far! I will continue to monitor this and then make a list of everything I use. Another thought, has anyone come up with a way to automate this? Possibly going to each programs download page, downloading the most recent version, and installing? I may way on this as well, any ideas would be appreciated! As far as the type of machine, most of mine are hybrids i.e- server, workstation, development machine. So all of the above!

    Read the article

  • correct format for datetime appended to filename

    - by jhayes
    I'm trying to setup a batch file to execute a set of stored procs and dump the output to a timestamped text file. I'm having problems finding the correct format for the timestamp. Here is what I'm using osql.exe -S <server> -E -Q "EXEC <stored procedure> " -o "c:\filename_%date:~-0,10%_%time:~-0,10%.txt" The error I get is: Cannot open output file - x:\filename_Thu 06/25/_16:26:43.1.txt No such file or directory I can't find the documentation and I've played around with it but can't find the correct format.

    Read the article

  • Exchange 2010: Import a PST when Local Move Request fails

    - by gravyface
    So the trail of tears continues with my SBS 2003 to 2011 migration: all the mailboxes have moved mailbox store from OLDSERVER to NEWSERVER, with the Local Move Requests completing successfully, except for one. I've logged into their machine and have exported their mailbox as a PST. I'm about to import it, but it seems to me that because the mailbox is still on OLDSERVER, even with a new Outlook profile pointing to NEWSERVER in Outlook, it'll push the mail into the current mailbox store on the old server. Please tell me I don't have to blow away her existing mailbox, logon, etc. on the old SBSERVER: is there a way to change the state from "Legacy" to "User Mailbox" without actually moving the mailbox store? Create a new mailbox for her user in NEWSERVER?

    Read the article

  • Encryption container for multiple people

    - by Adam M.
    I was just wondering if anyone may have come across a product that would allow for a container based encryption to be used by multiple people, in a Windows Server setup. I wanted to see if there might be something like a truecrypt that could handle being accessed by two accounts? Looking to see if there is a product that would have such properties that would allow only a hand full of users access to the content of the location, but allow for the files to be backed up a normal backup system. That way if a file had to be restored, the container could be redirected to another location for one of the users to get access to it? This would allow for access to be restricted beyond the NTFS and file share permissons

    Read the article

  • Oracle 10g Failover Database - How to fail back?

    - by rrkwells
    I want to know how the failover database concept works after recovery. We have defined our application to connect to a backup database in case the production database fails. If this happens, then all the transactions will be happening on that backup database. Once the production db server is running again, then how do we make sure the changes made in the backup database will be reflected on the production database? We want to make sure that any changes made while failed over are not lost. We are using Oracle 10g.

    Read the article

  • nagios service check

    - by DRH
    I am new to nagios and we have a small issue I need to ask assistance with. Many of the machines that we monitor can go unresponsive for a bit when some very intensive cpu tasks are run. This makes nagios send warnings and alerts while these hosts are busy reporting things like 'ping timeout' or 'zombie processes' and even swap space warnings, but in actuality there is not a problem. Is there a way to configure nagios to not send such alerts, but check x number of times over a period of time and only then send an alert at the end of that time if the server in question has not recovered?. Looking at the commands.cfg file, I see entries like this: define command{ command_name check_local_swap command_line $USER1$/check_swap -w $ARG1$ -c $ARG2$ } How could I modify this example to accomplish what I want above. Thanks

    Read the article

  • Realtime website slowness (PHP/MYSQL)

    - by 3s2ng
    We got a website with realtime transactions. The past few days we experienced some slowness during a period of time. Usually the slowness last for 5 minutes. And also during the slowness we noticed that sometimes we can not connect to SSH and FTP or sometimes very slow to connect to those services. Currently we are trying to identify the issue. We already setup database monitoring tool. Now are about to signup with pingdom.com My question is. If the website is too slow due to the database (table or row lock) will it affect the other services like SSH and FTP? Does the ping correlates the page load and the connection between my PC and server? Thanks, Mark

    Read the article

  • Can't conncet to memcached

    - by DMClark
    We currently have memcached running on CentOS. None of our PHP applications can connect, have tried multiple applications trying to establish access. The most informative PHP error we get is: "Memcache::get() [function.Memcache-get]: Server 127.0.0.1 (tcp 11211) failed with: Permission denied (13) in /var/www/.." memcached 1.4.5 PECL 2.25 We can telnet and it works. IP tables is full access from lo to lo. We've tried this on two different servers with both compiled version and the rpm in CentOS 5.5 and get the same result. Is there anything fairly obvious that we are missing?

    Read the article

  • mongodb eating 48G in 1min

    - by ledy
    In mongodb i work with this collection: Size 55.93g Data Size 39.82g Storage Size 41.08g Extents 53 Indexes 4 Index Size 9.64g It takes few seconds of mongdb being up with this single collection and all 48GB RAM on the dedicated server are gone. That's worse because there is also a mysqld+nginx/fcgi on this machine which should be allowed to use at least 24GB together. I.e. remaining 24GB, enough for the mongod! However, it does not share in a fair way. Everybody says that the memory for mongod is managed by OS and releases unneccessary space for other processes if they demand RAM. On my machine it is not releasing RAM. What's wrong? free total used free shared buffers cached` Mem: 49559136 49403908 155228 0 57284 47247564 -/+ buffers/cache: 2099060 47460076 Swap: 8008392 164 8008228

    Read the article

  • How to setup KVM partitioning?

    - by Richard
    I'm trying to install SolusVM/KVM on a server with one 500 GB HD and one 256 SSD. I am a complete beginner at Linux. SolusVM has a script on the site to download and install automatically, but how do I specify to install the script on the 500 GB HD and put the virtual machines on the SSD? And how should my partitioning look like? I want KVM isos and stuff on HD and virtual machines on SSD. Thanks a lot.

    Read the article

  • CentOS I/O blocking flushing dirtypages

    - by user3662224
    I have an Apache webserver on CentOS. It is not doing anything fantastic or complicated at all but it does at intervals experience a high load (120 for 4 CPUs) for some reason. During this time I see high waits and write durations. I have narrowed this down to flushing of dirtypages to disk to make room for new data. I have tried many things and talked to the developers who are also puzzled by this. I am stumped. Details: vmware virtual guest server 4 cpus 12G RAM Apache PHP I have 40 servers in this cluster that share the same disk structure. Only the webservers are experiencing this issue. Any ideas?

    Read the article

  • nginx location pathing issue

    - by Michael Jefferys
    I've got a pretty much default sites-enabled set up in my nginx on debian squeeze and i'm now trying to get it to serve up my munin graphs on myhost/munin/ Heres the location i've added to the config location /munin { root /var/cache/munin/www/; index index.htm index.html; } And here is the error I recieve: 2012/07/09 23:52:03 [error] 3598#0: *13 "/var/cache/munin/www/munin/index.htm" is not found (2: No such file or directory), client: 93.*.*.*, server: , request: "GET /munin/ HTTP/1.1", host: "" This set up used to 'just work' in apache. I'm new to nginx so a bit lost as to why its adding the extra /munin when looking for the path. Any advice?

    Read the article

  • Network randomly drops

    - by user147661
    I have a VPS, ubuntu server 12.04 and for unknown reasons the network connection drops and I have to access my vps via the console and ifdown eth0 and ifup eth0. This brings the connection back up from about a few seconds to a few hours. I have run tail -F /var/log/syslog to monitor the log to see what happens at the point of drop and there is nothing unusual. The vps is running under virt-manager and the hardware is set to virtio. I have had linux friends look at my vps and they are unsure of what is the cause.

    Read the article

  • Need help sending mass emails [closed]

    - by Jose
    Possible Duplicate: Prevent mail being marked as spam I have an Ubuntu server that I want to be able to send out several thousand emails each containing a generic header and an unique pdf attachment containing an invoice. What I want to know is what would be the best way to accomplish this? Also, are there any other programs I need to install such as anti-virus / verification tools? I would like to be able to know what the procedures for this normally are as I would prefer that the emails are not send to the junk folder of the clients receiving the emails. Any tips would be appreciated. Thanks

    Read the article

  • How does Google manage to serve results so fast?

    - by Quintin Par
    I am building an autocomplete functionality for my site and the Google instant results are my benchmark. When I look at Google, the 50-60 ms response time baffle me. They look insane. In comparison here’s how mine looks like. To give you an idea my results are cached on the load balancer and served from a machine that has httpd slowstart and initcwnd fixed. My site is also behind cloudflare From a server side perspective I don’t think I can do anything more. Can someone help me take this 500 ms response time to 60ms? What more should I be doing to achieve Google level performance?

    Read the article

  • Vagrant and cups port forwarding not working. Not accessible

    - by AAlvz
    I'm working with vagrant and I'm trying to use it as a printing server. I installed cups. Internally everything works just fine. I can even make a quick curl to my localhost:631 (cups port inside my vagrant) and there's everything. The thing is I can't access it in any way I try from the host machine. Obviously I forwarded the port and I've tried with several ports. I've also tried with Debian squeeze and Ubuntu 12.04. Here is my current Vagrantfile Vagrant.configure("2") do |config| config.vm.box = "guruDebian" config.vm.network :forwarded_port, guest: 80, host: 8080 config.vm.network :forwarded_port, guest: 631, host: 6363 ## HERE IS CUPS end Any ideas? ... I'll upload any file if necessary.

    Read the article

  • mongodb segmentation fault(11) macosx

    - by Wish
    I have problem, i cant figure out, how to fix.. So i am on MacOSX machine, running php 5.3.15 version, using mongo 1.3.1 version. When i try to execute php script, in which i try to connect to remote mongodb server, I get segmentation fault(11).. I installed php driver with sudo pecl install mongo I have seen, that this problem is quite popular, but havent found real solution yet.. I dont know if I am asking this question in correct stack site.. If you need anything else, just ask.

    Read the article

  • can't open large rss remote files

    - by anarchOi
    I'm setting up a rss feed in vbulletin to import the entries into the forum. The rss is very large (2000+ entries) and it is hosted on a different server (i have control on it). Problem is that i cannot add this feed into vBulletin, i am getting a weird error. If i edit the feed and remove half of the entries then it will work correctly, which makes me think it is because the file is too large. I suppose i have to change a setting in php.ini or something to allow bigger files to be opened, but what do i need to look for ? Thanks I'm on Debian.

    Read the article

  • When a new user is created on Centos 6, it takes a while (30 mins) before he can access his group folder

    - by Diepseun
    I created a new user and made it part of a certain group which has full access (777) to a folder. Checked the user in Samba, password the same as his Windows (XP) password, rebooted his desktop but he didn't have access to the folder. Checked the Samba group and config file and the user was defined as a member of the group. It didn't make sense and I then did something else for a while. When I tried again, without doing anything further about it, the user had access to the folder. I did restart the Samba server after my original changes. Thanks in advance.

    Read the article

< Previous Page | 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962  | Next Page >