Search Results

Search found 95574 results on 3823 pages for 'mac osx server'.

Page 1844/3823 | < Previous Page | 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851  | Next Page >

  • MySQL 5.0 Unavailable but still working every monday morning

    - by user1578031
    So I have a MySQL server with 4-5 databases and every monday morning, I can't login with PHPmyAdmin or the cmd line tools or options and the funny thing is the connection can be made later in the day around 3:30 - 4PM. I can't replicate the issues on any dev boxes and cant upgrade to 5.5 for a number of reasons. I want to try and poll the sql databases with a simple script and just connect to it and then disconnect. I'm not very good with SQL but would like to know if anyone could help with a script which I can automate to run sunday every 30mins or so, to see if the issue can be stopped by the database being connected to over the very quiet period of sunday... After this im out of ideas and have to wait for the HW/SW refresh to get 5.5 on there...

    Read the article

  • How to make scp copy hidden files?

    - by rascher
    I often use SCP to copy files around - particularly web-related files. The problem is that whenever I do this, I can't get my command to copy hidden files (eg, .htaccess). I typically invoke this: scp -rp src/ user@server:dest/ This doesn't copy hidden files. I don't want to have to invoke this again (by doing something like scp -rp src/.* ... - and that has strange . and .. implications anyway. I didn't see anything in the scp man page about an "include hidden files". How can I accomplish this?

    Read the article

  • How to run nodejs on linux platform

    - by rotem
    How to run node.js on host with linux platform? To run node.js on localhost with windows operation system is simple I download package from nodejs.org/download/ and I execute Windows Installer (.msi) I go to console command line and I type node file.js and everything fine. but in my host with linux platform I have control panel with no option to run type file exe, msi and there is no window with command line, So how can I be able to run nodejs on my host? I call to support of my hosting bluehost.com and they don't know. my Details server and control panel Thanks for any help

    Read the article

  • Synckolab will not start automatically

    - by EBV2010
    We have a Kolab server for e-mail, calendar and contacts with Thunderbird for a client. The add-ons are Lightning and Synckolab. The workstations are Kubuntu, most 10.04, some 11.10. It basically works but for one nagging problem: the automatic sync (that is, the settings that starts Synckolab upon start of Thunderbird and every x minutes as per setting) does not fire. We went through the whole routine: setting it, setting it to zero and back to any number of minutes, stopping/starting Thunderbird or the entire computer to make sure it sees and sets it. In the configuration console is reflects the changes. But still it will not automatically fire Synckolab. Manual syncs work without any problem (none that we've seen - it reflects all the added, changed etc calendar events). In short: Synckolab does not fire automatically with any setting who have thought of.

    Read the article

  • Unix printing a banner page on every print job

    - by yum_tacos4u
    I have a Data General server on unix that is printing a banner page on every print. I originally thought that the banner page was comming from the printer. As this is an HP printer, I used telnet to get to the jetadmin and then proceded to disable the banner page, but this did not solve the issue. I then went into the sysadm program to see if the TCPIP printing was set to print a banner page on print jobs, but I did not see any options to print a banner page. Any help or ideas on how to disable the banner page from printing in unix? Here is a banner page example print

    Read the article

  • Capture the build number for a remote-triggered Hudson job?

    - by EMiller
    I have a very simple inhouse web app from which certain Hudson builds (on another server) can be triggered remotely. I have no problem triggering the builds, but I don't know how to capture the associated build number for later reference. I'm using the buildWithParameters trigger, and the actual result of that call is just a mess of HTML - I don't believe it gives me back the build number. I started down the path of pulling the whole build list for the job (via the api), and then attempting to reconcile that list against my records - but that's much more complicated than I'd like it to be. I also considered sleeping for a few seconds after launching the job, and then grabbing the latestBuild from the Hudson api - but I'm sure that's going to go wrong at some point (someone will fire off two jobs quickly, and I'll get the association wrong).

    Read the article

  • How do you configure recycle bins on roaming profiles?

    - by Zombian
    I copied the following from a post on the Spiceworks forum which remained unanswered: Is there any way to place the Recycle Bin back on the desktop of a Roaming Profile with the Desktop being redirected? I have used Google and can't find a straight forward answer. I am asking for people with experience in this. This is for a Windows XP machine. I saw mention of needing to use a program such as Undelete but I'm hoping that is not the case. Further explanation: I use redirected folders and whenever a user deletes something from their desktop,my documents it doesn't show up in the recycle bin. It doesn't appear in the recycle bin on the server either. Where is this data? I doubt it is permanently deleted. Is there a way to change the recycle bin on the users' desktop to display those files? Thank you!

    Read the article

  • Creating user accounts in Amazon EC2

    - by Tvanover
    I am putting together a test environment using Amazon's EC2 for me and some friends to collaborate on a project. I am not a server guy but I do know my way around a bash prompt and have done some work on ubuntu before. I am using Amazon Linux AMI i386 EBS and have gotten apache and php running. Now I need to create the user accounts my friends and I will use to upload files (sftp) and work on the project (ssh). How should I go about this? Should I just use adduser and configure it like normal? Or should I use the AWS IAM groups?

    Read the article

  • IIS and PHP restrict IO permissions

    - by ULTRA_POROV
    I have php installed trough a fastCGI module. Is there a way to restrict the module (php.exe) read / write permissions to only the directory (+ subdirs) of the IIS site that is calling it? I need this to prevent one IIS PHP site from having access to files outside its own directory. How to do this? Is there a setting in php.ini or in the IIS configuration? I believe such a feature could exist, because when a file on the server is requested the root path of the site is also known, all it would take is that IIS passes this path to the php module, and the php module should on its end allow only IO operations within this path. PS: I know it is possible to achieve this by using a different windows account for each website, this is not an option.

    Read the article

  • How do you synchronise huge sparse files (VM disk images) between machines?

    - by chrisdew
    Is there a command, such as rsync, which can synchronise huge, sparse, files from one linux server to another? It is very important that the destination file remains sparse. It may be longer (but not bigger) than the drive which contains it. Only changed blocks should be sent across the wire. I have tried rsync, but got no joy. groups.google.com/group/mailing.unix.rsync/browse_thread/thread/94f39271980513d3 If I write a programme to do this, am I just reinventing the wheel? http://www.finalcog.com/synchronise-block-devices Thanks, Chris.

    Read the article

  • Are there any tools for monitoring individual Apache virtual hosts in real-time?

    - by Dave Forgac
    I'm looking for a way to monitor and record Apache traffic, separated by virtual host. I am currently using Munin to capture this and other data for the entire server however I can't seem to find a way to do this by vhost. This link describes using a module called mod_watch which is apparently no longer in development: http://www.freshnet.org/wordpress/2007/03/08/monitoring-apaches-virtualhost-with-munin/ The file that is listed as being compatible with Apache 2.x is reported to have problems with missing vhosts an reporting data correctly. Does anyone know of a reliable way to determine real-time traffic per vhost? If I can find this it should be easy enough to write a new Munin plugin.

    Read the article

  • grep only returns help text

    - by Pete Mancini
    Well, I am perplexed. I am working with an Ubuntu server and I type in grep 'bash' *.sh BUT fgrep 'bash' *.sh works like a champ. which grep and which fgrep both point to their respective executables in /bin. I am perplexed as to what I am doing wrong. EXAMPLE output: $ grep -F 'grounding' repl.clj Usage: grep [OPTION]... PATTERN [FILE]... Search for PATTERN in each FILE or standard input. PATTERN is, by default, a basic regular expression (BRE). Example: grep -i 'hello world' menu.h main.c $ fgrep 'grounding' repl.clj (p/concepts-for-grounding-term imp1 "PERSON" "summary") See? grep is failing but fgrep is working fine. That is why I am perplexed.

    Read the article

  • Small footprint on a Laptop

    - by sqldebacle
    I am trying to find a solution where I can do the following: 1) Run a small footprint on my laptop 2) Run virtual instances of OS w/ no primary OS installed. All the OS that I will ever use will be all virtualized. I tried playing around with the VMware Esxi, and got it to boot it from the flash drive, etc. But this just runs the server. I cannot actually run my virtual instances from there. Anyone has done this? Something similar implemented with VMWare products without needing 2 computers will be great. Thanks, -Subhash

    Read the article

  • Can't get ZSH working on CentOS

    - by waveslider
    I've been using zsh for a couple of years now on Ubuntu and really like it a lot. I've installed it on our production server as well, which is running CentOS 5.2 However, I just installed it via yum on a new VM I created to use as a development box, to replicate our production box as closely as possible. Although yum shows that it is definitely installed (/bin/zsh) and that it is set as my shell, it does not appear to be working. Instead of creating the .zshrc and .profile files in my home directory, it created a .tcshrc file. Also, I did not receive the default configuration menu that is always displayed once you begin using ZSH, and none of the features (like advanced tab comple

    Read the article

  • What is the best way to setup a heartbeat agent for failover between two VMs?

    - by EGr
    I have two VMs in VirtualBox that use NAT for their network adapters. They are both getting the same IP address, so I will need to reconfigure that; but knowing that, is it possible to set up a heartbeat agent to failover an apache server if one of the two VMs go down? The way I pictured it would be that the webserver would be able to be accessed externally via :80. No matter what VM was running, I would be able to access the website at that IP/port since failover would be setup. I'm running into trouble setting up IPs when the network adapters are set to NAT, and people have told me that I shouldn't be setting the IPs in this configuration. So what should I do to achieve what I'm looking for? Is it even feasible?

    Read the article

  • Reverse proxy with SSL and IP passthrough?

    - by Paul
    Turns out that the IP of a much-needed new website is blocked from inside our organization's network for reasons that will take weeks to fix. In the meantime, could we set up a reverse proxy on an Internet-based server which will forward SSL traffic and perhaps client IPs to the external site? Load will be light. No need to terminate SSL on the proxy. We may be able to poison DNS so original URL can work. How do I learn if I need URL rewriting? Squid/apache/nginx/something else? Setup would be fastest on Win 2000, but other OSes are OK if that would help. Simple and quick are good since it's a temporary solution. Thanks for your thoughts!

    Read the article

  • Apache Virtual host Subdomains points to same directory

    - by Jakobud
    I have setup subdomains using Apache before and have never really ran into any big problems. But with this (I believe Centos) server that is one of my clients, I'm not understanding what I'm doing wrong. Here is the .conf that apache is loading: Listen 80 NameVirtualHost *:80 <VirtualHost *:80> ServerName www.thedomain.com DocumentRoot /u1/thedomain.com/public RailsEnv production </VirtualHost> <VirtualHost *:80> ServerName subdomain.thedomain.com DocumentRoot /u1/subdomain.thedomain.com/public_html </VirtualHost> When I access either the primary or subdomain addresses, they both point to the primary www.thedomain.com content. Any thoughts? UPDATE: Yes I did a configtest and graceful after making the changes.

    Read the article

  • Migrating Gmail to Office 365

    - by user218699
    Good Morning, I have been setting up Office 365 for my organization. We are currently using Gmail. I have synced our local Active Directory server w/ Office 365, as well as our domains. The problem I am having has to do with migrating mailboxes from Gmail to Office 365. I have been using this article to walk me through the process: http://technet.microsoft.com/en-us/library/dn568114.aspx The issue arises when I begin to sync the mailboxes. Currently I have been trying to sync my own mailbox as a test. The synchronization process has been going on for about 15 hours (for just one mailbox) with no errors or any information given by Office 365, other than the "Syncing" status on the migration page in the Exchange Admin Center. Is syncing a single mailbox supposed to take this long, or have I missed a step? Thanks!

    Read the article

  • What exactly is an invalid HTTP_HOST header

    - by rolling stone
    I've implemented Django's relatively new allowed hosts setting, which is meant to prevent attackers from submitting requests with a fake HTTP Host header. Since adding that setting, I now get anywhere from 20-100 emails a day notifying me of invalid HTTP_HOST headers. I've copied in an example of a typical error message below. I'm hosting my site on EC2, and am relatively new to setting up/maintaining a server, so my question is what exactly is happening here, and what is the best way to manage these invalid and I assume malicious requests? [Django] ERROR: Invalid HTTP_HOST header: 'www.launchastartup.com'.You may need to add u'www.launchastartup.com' to ALLOWED_HOSTS.

    Read the article

  • Apache 2 Symbolic link not allowed or link target not accessible

    - by astropanic
    My apache server runs as user foo. I have some Rails applications in /home/foo/app1 /home/foo/app2. Each of them has an vhost <VirtualHost *:80> ServerName app1.foobar.com ServerAlias www.app1.foobar.com DocumentRoot /var/www/html/app1/current/public RailsEnv production <Directory /var/www/html/app1/current/public> AllowOverride all Options -MultiViews </Directory> </VirtualHost> I have a symlink in /var/www/html/app1 : current -> /home/foo/app1/tmp_20102611 All file permissons are set correctly (user foo group foo), I can go through the filesystem from shell. SELINUX is disabled Distro is CentOs 5.5 Which the above symlink I get an 403 and an error entry in error_log Symbolic link not allowed or link target not accessible:/var/www/html/app1/current When I symlink my app in the subdir of /var/www/html instead of /home/foo it works. How I can avoid this error still placing my app in my /home/foo directory ?

    Read the article

  • Is there a chroot build script somewhere?

    - by Nils
    I am about to develop a little script to gather information for a chroot-jail. In my case this looks (at the first glance) pretty simple: The application has a clean rpm-install and did install almost all files into a sub-directory of /opt. My idea is: Do a find of all binaries Check their library-dependencies Record the results into a list Do a rsync of that list into the chroot-target-directory before startup of the application Now I wonder - ist there any script around that already does such a job (perl/bash/python)? So far I found only specialized solutions for single applications (like sftp-chroot). Update I see three close-votes for the reason "off topic". This is a question that arose because I have to install that ancient piece of software on a server at work. So if you still feel this is off-topic - leave a comment...

    Read the article

  • ftp users configuration in OpenSuse 12

    - by chieroz
    I usually work with MacOSX servers but this time I need to set up a ftp service on a OpenSuse 12.2 server and I am a little lost. I am using the remote YAST2 tool via ssh. I created several users who can connect via ssh and/or ftp, so the basic setup is ok. But when connecting via ftp all my users don't have write permissions. The FTP directory for authenticated users is /srv/www/htdocs, which has permissions root:root. The OpenSuse manual say it's bad practice to change these permissions, but my normal users (even the ones in the sudoers list) cannot upload files. So I am stuck: as a workaround I use rsync, but from time to time I just need to establish a working ftp connection. What's the right approach for users permissions in this scenario? Thanks a lot.

    Read the article

  • What are the common Linux commands for SAN-related activities? How do I check if a LUN is attached to the computer?

    - by Nishant
    How do I check if a LUN has been presented to my server? What are the Linux commands for that? Do the LUNs show up in a fdisk -l command like a normal /dev/sda gets listed? What are other commands associated with general SAN related checks in Linux? What is WWN and how does that have any relevance? If we have LUNs, what is the use of multipathing? Bit lengthy but I am not able to get a grasp on the topic. Any help would be appreciated.

    Read the article

  • Windows 8 blue screen error watchdog violation

    - by Pramodh
    My system is windows 8 pro rtm x64. Recently my system started hanging frequently. Nothing but the restart button works. But now just after hanging it showed me a BSOD error saying watchdog violation. I used bluscreen viewer to see the error in the memory dump and it showed me something about hal.dll, ntoskrnl.exe. This is what windows showed me. Problem signature Problem Event Name: BlueScreen OS Version: 6.2.9200.2.0.0.256.48 Locale ID: 1033 Extra information about the problem BCCode: 133 BCP1: 0000000000000000 BCP2: 0000000000000281 BCP3: 0000000000000280 BCP4: 0000000000000000 OS Version: 6_2_9200 Service Pack: 0_0 Product: 256_1 Bucket ID: 0x133_DPC_NETIO!KfdClassify Server information: c03a9f52-f2b5-483f-9b4a-cbb5be3a72c0 Can anyone please walk me through the steps to get rid of this error please? Thank you.

    Read the article

  • snmptrap and snmptt - authcommunity not found

    - by sabs6488
    I am trying to configure snmptt to translate the snmp traps received and handle them as passive checks in icinga monitoring server as described here . after doing the changes to the snmptrapd.conf . I am trying to restart the service and I see authcommunity : command not found, traphandle : command not found . my understanding is the authcommunity and traphandle are just configuration directives which will tell snmptrapd about the community string to use and the traphandle script to be called. It would be helpful if someone can help me understand better. Thanks, sabs

    Read the article

< Previous Page | 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851  | Next Page >