Search Results

Search found 20099 results on 804 pages for 'virtual host'.

Page 391/804 | < Previous Page | 387 388 389 390 391 392 393 394 395 396 397 398  | Next Page >

  • How do you keep track of all your passwords?

    - by Sam Saffron
    How do you keep track of all your passwords? Personally I host a personal copy of clipperz, I used keepass and passpack in the past. What password manager would you recommend, what features does it have that make it awesome? Now at 70+ "answers" it's a pretty good bet that your favourite program is already mentioned. Upvote that if that's the case. If you can't yet upvote, come back when you've gained enough reputation instead of posting a duplicate answer.

    Read the article

  • nginx can't see MySQL

    - by user135235
    I have a fully working Joomla 2.5.6 install driven by a local MySQL server, but I'd like to test nginx to see if it's a faster web serving experience than Apache. \ PHP 5.4.6 (PHP54w) \ CentOS 6.2 \ Joomla 2.5.6 \ PHP54w-fpm.i386 (FastCGI process manager) \ php -m shows: mysql & mysqli modules loaded Nginx seems to have installed fine via yum, it can process a PHP-info file via FastCGI perfectly OK (http://37.128.190.241/php.php) but when I stop Apache, start nginx instead and visit my site I get: "Database connection error (1): The MySQL adapter 'mysqli' is not available." I've tried adjusting my Joomla configuration.php to use mysql instead of mysqli but I get the same basic error, only this time "Database connection error (1): The MySQL adapter 'mysql' is not available" of course! Can anyone think what the problem might be please? I did try explicitly setting extension = mysqli.so and extension = mysql.so in my php.ini to try and force the issue (despite php -m showing they were both successfully loaded anyway) - no difference. I have a pretty standard nginx default.conf: server { listen 80; server_name www.MYDOMAIN.com; server_name_in_redirect off; access_log /var/log/nginx/localhost.access_log main; error_log /var/log/nginx/localhost.error_log info; root /var/www/html/MYROOT_DIR; index index.php index.html index.htm default.html default.htm; # Support Clean (aka Search Engine Friendly) URLs location / { try_files $uri $uri/ /index.php?q=$uri&$args; } # deny running scripts inside writable directories location ~* /(images|cache|media|logs|tmp)/.*\.(php|pl|py|jsp|asp|sh|cgi)$ { return 403; error_page 403 /403_error.html; } location ~ \.php$ { fastcgi_pass 127.0.0.1:9000; fastcgi_index index.php; include fastcgi_params; fastcgi_param SCRIPT_FILENAME $document_root$fastcgi_script_name; include /etc/nginx/fastcgi.conf; } # caching of files location ~* \.(ico|pdf|flv)$ { expires 1y; } location ~* \.(js|css|png|jpg|jpeg|gif|swf|xml|txt)$ { expires 14d; } } Snip of output from phpinfo under nginx: Server API FPM/FastCGI Virtual Directory Support disabled Configuration File (php.ini) Path /etc Loaded Configuration File /etc/php.ini Scan this dir for additional .ini files /etc/php.d Additional .ini files parsed /etc/php.d/curl.ini, /etc/php.d/fileinfo.ini, /etc/php.d/json.ini, /etc/php.d/phar.ini, /etc/php.d/zip.ini Snip of output from phpinfo under Apache: Server API Apache 2.0 Handler Virtual Directory Support disabled Configuration File (php.ini) Path /etc Loaded Configuration File /etc/php.ini Scan this dir for additional .ini files /etc/php.d Additional .ini files parsed /etc/php.d/curl.ini, /etc/php.d/fileinfo.ini, /etc/php.d/json.ini, /etc/php.d/mysql.ini, /etc/php.d/mysqli.ini, /etc/php.d/pdo.ini, /etc/php.d/pdo_mysql.ini, /etc/php.d/pdo_sqlite.ini, /etc/php.d/phar.ini, /etc/php.d/sqlite3.ini, /etc/php.d/zip.ini Seems that with Apache, PHP is loading substantially more additional .ini files, including ones relating to mysql (mysql.ini, mysqli.ini, pdo_mysql.ini) than nginx. Any ideas how I get nginix to also call these additional .ini's ? Thanks in advance, Steve

    Read the article

  • Intranet Ip - Access from Custom Domain

    - by Alexander Wigmore
    I have setup a local intranet in my office using IIS7 (Windows 7 Machine), currently it can be accessed through the PC's static IP, however I would like it so that internally it can just be accessed through an easier method, e.g typing in http://intranet (or something similar). There are over 60 PC's int he office, so individually updating Host files on the PC's is not really ideal. We don't need it to be accessible from the outside world (I.e, we don't care/want it to be an Extranet). Any tips?

    Read the article

  • open source web application for viewing .ics calendars

    - by aaron
    I need an open source web application that allows users to view .ics calendars I am hosting a calendar with DAViCal, and users can view/edit calendars with their own clients (like Apple iCal or Mozilla Sunbird), but I want a web app that will display the public calendars so people w/o a user account can view the calendar just by visiting a URL. Requirements: Open source web app: something I can host right along with DAViCal Human-readable calendar: i.e. it has a day/week/month view to visualize calendar events

    Read the article

  • Run VirtualBox in background, without a window

    - by Bashman
    Hi! I was wandering if it's possible to run an Ubuntu virtual machine using VirtualBox in background, I mean, without any window open. The idea is to conect via SSH to the linux host, that would be running in background. Have you ever done somenthing like this? Is it possible? Thanks!

    Read the article

  • Make a bootable USB drive that can install both Windows XP and Ubuntu

    - by Utkarsh
    I have ISO images for both Ubuntu and Windows XP. I want to host both of them on a USB drive so that I can install either without needing installation CDs (I don't have a CD drive). How can I do that? SO, I want to have both Windows XP and Ubuntu on my USB Drive so that i could install any one of tem just from a USB. I do not have CD Drive thats why i wanna do that. I have ISO image of both ubuntu and windows xp

    Read the article

  • Any 3rd party tools that integrate IIS management with some sort of application inventory/change man

    - by NoCarrier
    I've got an environment with several IIS 6 web servers hosting hundreds of apps (dozens of sites, hundreds of virtual directories) all with a myriad of different configurations belonging to dozens of different developers (all deploying apps willy nilly). Is there some sort of managed software solution that will centralize management of all my IIS6 environments provide some inventory functionality allow for reporting or querying of said application inventory enforce and automate some sort of deployment process?

    Read the article

  • Mass bulk add domains to web hosting service (possible?)

    - by Scott
    I was wondering if anyone does bulk adding of domains to your web hosting provider (Amazon, Linode, Rackspace, etc). I am thinking of creating a product that allows user to host their site on top of my web hosting and want something that can allow me to bulk add domains (and point DNS to my web hosting DNS) with as little manual work as possible. I am thinking of getting a VPS to do this. Is this possible even? Thanks Scott

    Read the article

  • Which SQL database should I run?

    - by Chris Salij
    I'm creating my first website from hand-coded scratch (just used WordPress and Drupal before) and I'm at the point where I'm setting up my database for the site. However my host has given me 4 options to choose from in the database department. MS SQL Server 2005 MySQL PostgreSQL MySQL 5 Which should I go for? From what I understand PostgreSQL is meant for smaller databases with high access times. Though to be honost I'm not quite sure.

    Read the article

  • where does https on apache get it's configuration?

    - by Matthew
    So originally my host (mediatemple dv) has two default directories for the roots: 1)httpdocs/ 2)httpsdocs/ In the conf directory I changed the vhosts.conf and httpd.include and others to change from httpdocs to custom folders. Now I installed a new ssl certificate and https://example.com goes to the default page located at httpsdocs. I'm just wondering where configurations for apache are stored for ssl. Ideas?

    Read the article

  • What's the best way to rewrite traffic from domainA.com/foo to domainB.com/bar while properly rewrit

    - by Chad DePue
    We have a number of sites that have blogs, like domainA.com/blog domainB.com/blog and we host the blogs on wordpress multi user: our-separate-wordpress-site.com/domaina_blog our-separate-wordpress-site.com/domainb_blog for SEO reasons we really, really want domainA.com/blog to be the blog url, not the other path. But we don't see any examples where this is done, because we need not to just rewrite the traffic, but cookies as well... is this possible with a webserver or a reverse proxy?

    Read the article

  • Deny login from certain hosts if logging in with specific sql credentials

    - by Dave
    I want to stop some of our developers from connecting to the production sql server using a specific sql account. They have rights to connect through windows authentication with lower rights. They claim that changing the password will affect too many other processes running on our processing machine. So I want to deny access if they're connecting from there dev machines for now. Another way this would work is if I could just allow connections from one specific host.

    Read the article

  • Virsh: Different XML with edit than dumpxml. Why?

    - by Dave Vogt
    I'm trying to fetch the VNC access data from a virtual machine managed by libVirt. However, when I run virsh dumpxml $machine, the vnc passwd is missing: <graphics type='vnc' port='-1' autoport='yes'/> Checking the same using virsh edit $machine, I see the password is actually there: <graphics type='vnc' port='-1' autoport='yes' passwd='asdf'/> Why is this? Is this intentional (what reason?), or could this be a bug?

    Read the article

  • TFS 2010 and SCVMM to manage snapshots

    - by ELSheepO
    I have a VM with TFS 2010 and SCVMM 2008 installed on (I know your told not to do that). I have SCVMM linked to the host and thats working good, and have configured the scvmm on the lab management tab on TFS. My question is, how can you take/roll back snapshots automatically? ie install base take snapshot install patch run tests take snapshot go back to snapshot taken at step 2 Thanks in advance.

    Read the article

  • Remote MySQL connection fails (10060)

    - by Stefan Åstrand
    When I am trying to connect to a MySQL database from a remote computer I get a prompt saying: Connection Failed: [HY000] [MySQL][ODBC 5.1 Driver]Can't Connect to MySQL server on 'XXX.XXX.XX.XX' (10060) I have created a user account in the MySQL Administrator and added a host to enable remote access, I have also made an exception for my Windows Firewall on port 3306 but the connection still fails. What is the problem? Thanks!

    Read the article

  • Transient network dropout for Xen DomU's

    - by Stephen C
    We've got a CentOS server running a cluster of virtuals. Occasionally the cluster's internal network drops out for a minute or so ... and then comes back. The problem is somehow related to the actual network traffic, but it is not a simple load issue. (The system is generally lightly loaded, and the problem occurs irrespective of actual load.) The setup: CentOS 5.6 on Dom0, various CentOS on the DomU's Hardware - a Dell R710 with a BroadCom NextXpress 2 NIC (sigh) using the latest drivers for the NIC from BroadCom Xen configured to use network-bridge and vif-bridge Some iptable tweaks to route an unrelated port to one of the virtuals. The system has one externally visible IP address, and Dom0 runs an Apache httpd configured with a number of virtual hosts each of which reverse proxies to web servers running on the virtuals. (The virtuals have to be NAT'ed, primarily because we don't have enough allocated public IP addresses.) The symptoms: Works fine most of the time. When someone tries to UPLOAD a large file to one virtuals, the internal network drops out ... for all virtuals: The Dom0 httpd sees a network timeout talking to the backend server on the virtual and reports a 502. A previously established ssh connection from Dom0 to any of the DomU's freezes. Our monitoring shows ping failures for traffic between virtuals. The Xen consoles to the DomU's do not freeze. No log messages in any log files that I can see, on either Dom0 or the DomU's ... apart from the Dom0 httpd logs. After a minute or so, the problem clears by itself. This is 100% reproducible. What we've tried: Downloading, building and installing the latest BNX2 driver on Dom0 Turning off MSI on the NIC - adding "options bnx2 disable_msi=1" to /etc/modprobe.conf Turning off tcp segmentation offload - "ethtool -K eth0 tso off". Sacrificing a black rooster at midnight. I've exhausted all my options apart from switching to KVM ... or slaughtering more roosters. Any suggestions?

    Read the article

  • Can Office365 (Enterprise) sync from a remote LDAP server?

    - by kez
    I am looking in to the details of a migration to Office365 - primarily for the use of hosted Outlook capability. I see that the Enterprise plans have "Active Directory Sync", but details seem scarce. Is it possible to sync an independent LDAP server (such as Apache Directory) up to O365 (I want to keep it separate), or would I need to host a local instance of Active Directory and use that to sync users up to O365?

    Read the article

  • using tcpdump to display XML API requests without headers or ack packets

    - by Carmageddon
    I need assistance, I am trying to use tcpdump in order to capture API requests and responses between two servers, so far I have the following command: tcpdump -iany -tpnAXs0 host xxx.xxx.xxx.xxx and port 6666 My problem is, that the output is still hard to read, because it sends the Headers, and the ack packets. I would like to remove those and only see the XML bodies. I tried to use grep -v, but apparently this is all one request, so it filters the entire thing... Thanks!

    Read the article

  • Wiki/CMS with synchronization?

    - by Clinton Blackmore
    We're looking into putting up a wiki or CMS for internal use by our IT department. One of the big things we want to use it for is disaster recovery procedures. Given that a disaster, such as a power or network outage, might render the wiki inaccessible, it seems sensible to to host the wiki in two places so that if one is inaccessible, we can fall back to the other. Are there any wikis or CMSes that synchronize (or an alternate way to achieve a similar end)?

    Read the article

< Previous Page | 387 388 389 390 391 392 393 394 395 396 397 398  | Next Page >