Search Results

Search found 10299 results on 412 pages for 'apache'.

Page 332/412 | < Previous Page | 328 329 330 331 332 333 334 335 336 337 338 339  | Next Page >

  • Upgrading Fedora on Amazon to 12 but getting libssl.so.* & libcrypto.so.* are missing

    - by bateman_ap
    I am upgrading to Fedora 12 on a Amazon EC2 using help here: http://www.ioncannon.net/system-administration/894/fedora-12-bootable-root-ebs-on-ec2/ I managed to do a 64 bit instance OK, however facing some problems with a standard one. On the final bit of the install from 11 to 12 I am getting an error: Error: Missing Dependency: libcrypto.so.8 is needed by package httpd-tools-2.2.1.5-1.fc11.1.i586 (installed) Error: Missing Dependency: libssl.so.8 is needed by package httpd-tools-2.2.1.5-1.fc11.1.i586 (installed) This is referenced in the comments from the link above but all it says is: Q: Apache failed, or libssl.so.* & libcrypto.so.* are missing A: These versions are mssing the symlinks they require. Easy fix, go symlink them to the newest versions in /lib However I am afraid I don't know how to do this. If it is any help I tried running the command locate libssl.so and got: /lib/libssl.so.0.9.8b /lib/libssl.so.6

    Read the article

  • How do I get my 192.168.* Linux server accessible via http://hostname/?

    - by rfrankel
    (Sorry if this question isn't worded well and/or is duplicate. I'm not a networking guy and I'm probably not using the right terms...this also makes it hard to see if this has already been answered.) I'm running a CentOS server in VirtualBox, Windows host, and I can see access Apache-hosted pages at http://192.168.1.109/ from machines on my LAN. But what I'd like is for people to be able to type http://hostname/ ...both because it's easier and primarily because I'm not sure that local IP is static. I'm not really sure how to proceed - could someone point me in the right direction? Thanks.

    Read the article

  • 2 virtual domains: one on port 80 one on port 8080, how?

    - by Simone
    I've been struggling with this, basically what I want is this: A virtual domain to run on the 80 another virtual domain to run FROM 80 TO 8080 what I've done so far is trying different combinations with apache, but all i got is having all the virtual domains pointing to port 80 this is my conf: <VirtualHost domain1.com:80> DocumentRoot /var/www/domain1 <Directory "/var/www/domain1"> allow from all Options +Indexes </Directory> ServerName domain1.com </VirtualHost> <VirtualHost server.domain2.com:80> DocumentRoot /var/www <Directory "/var/www"> allow from all Options +Indexes </Directory> ServerName server.domain2.com ServerAlias www.server.domain2.com ProxyPass / http://server.domain2.com:8080/ </VirtualHost> suggestions? :(

    Read the article

  • Unable to sunchronize local and remote directories ("set times: Operation not permitted")

    - by Tom Auger
    I'm running into FTP errors using software like NetBeans or WinSCP: whenever I attempt to perform a synchronization or update of files from local -- server I get errors on the client saying "set times: Operation not permitted". This is clearly an issue with the way I've configured my Fedora installation. The user that I'm logging in with cannot touch -t any of these files, though he IS part of a group that has r/w access on the files. I do have root / sudo access to this server. What I would like to know is: a) is it likely that this problem would be solved by allowing my FTP user to "touch -t" these files b) how do I enable a certain user to be able to set timestamps on files without giving them ownership of the files (certain of these files need to be owned by Apache, for instance, so I don't want to chown them). Thanks in advance.

    Read the article

  • Can you have multiple PHP 5.x versions with cPanel 11.3x?

    - by atomicguava
    I've been Googling around for a while on this one but I haven't found a good answer yet! Is it possible to set up cPanel 11.3x so that it can run different versions of PHP 5.x (e.g. 5.2, 5.3 and later on 5.4) for each of the configured apache vhosts / domains? It would be great to do this either using .htaccess, php.ini or a setting within cPanel itself. I've seen EasyApache 3 mentioned in the documentation but even after reading through several times I haven't seen a definitive yes or no for whether this is possible - please let me know if you need any more info. This was the documentation I found for EA3: http://goo.gl/IH1sP

    Read the article

  • Asking for brief explanation of reverse proxying and recommended software

    - by 80skeys
    I need to set up a reverse proxy where the backend web server is serving https pages only. I've never set up a reverse proxy and would like a brief overview of how it works. One of my questions is whether the proxy needs to run https also, or does simple http suffice? Second question would be whether to use Apache, varnish, nginx, or squid. This is for an internal site for a small company, so not a lot of traffic expected. Maybe a few dozen users each day.

    Read the article

  • After deleting log files, Ubuntu server still saying there is no space

    - by Mark
    My Ubuntu server has stopped due to a lack of disk space. I deleted some log files which has grown huge very quickly. But df -h still shows I have no space left. When I run du -sh /* I can see that I should have plenty of disk space left after deleting the logs. I ran lsof +L1 and it brought up two files: /var/log/mail.log and /var/log/mail.err. These are two logs I had deleted. I restarted apache, postfix and mysql (mysql wont restart because of lack of disk space, it think) but still df -h shows no space.

    Read the article

  • Nginx and low-speed connections: request terminates after 253 seconds

    - by meze
    I'm trying to make nginx to handle static files. All is working fine except that when I throttle my connection speed to 8kbit/s, the loading process of a file just stops after 253-255 seconds (4.2 min according to chrome). No error in the log, the status code is 200 but the response is received partially. If I disable nginx and make apache to send the same file - it loads successfully after 10 minutes. The config I use for debugging is: client_header_buffer_size 16k; large_client_header_buffers 4 8k; client_max_body_size 50m; client_body_buffer_size 16k; client_header_timeout 20m; client_body_timeout 20m; send_timeout 20m; Did i miss some configurations?

    Read the article

  • migration of physical server to a virtual solution, what i have to do?

    - by bibarse
    Hello I'm new in this forum, so i would like that you forgive me for my blissfully and my low English level. I'm a trainee in company one month ago, and my mission is to migrate 3 physicals servers to a virtualization technology. The company edit softwares for E-learning so there are lots of data like videos, flash and compressed (zip). This is some inventory of the servers: OS: Debian, 2 redhat, apache, php/mysql, sendMail/Dovecot, webmin with virtualmin template to create dynamically the web sites because there is no sysadmin ... The future provider will be responsible of to secure, update and create the virtual machines (outsourcing) and with a RedHat OS's. So i want that you help me to choose a virtualisation technologie (for the i prefer KVM of Redhat RHEV, VMWare is expensive), how evaluate the hardware needs (this for evolution of 4 or 5 years) and to elaborate a good planing to don't forget any think. Thank you for your responses.

    Read the article

  • Migrating from Desktop PC to real Server

    - by tevlon84
    i am a student, working as a part-time Administrator at a startup. I never ever used a real Server ( only a Desktop Pc with apache ) The Company i am working for is growing and they want to switch to a real Server. My idea would be to use the Ubuntu build-in Backup function and use this Backup file as Base for the Rack-Server, but i don't know, which problems i would run into. Is it a good idea ? So basically my question is : *What is the easiest way to migrate from a Desktop PC to a real Rack-Server? ( on an Ubuntu Server) *

    Read the article

  • Per-user vhost logging

    - by kojiro
    I have a working per-user virtual host configuration with Apache, but I would like each user to have access to the logs for his virtual hosts. Obviously the ErrorLog and CustomLog directives don't accept the wildcard syntax that VirtualDocumentRoot does, but is there a way to achieve logs in each user's directory? <VirtualHost *:80> ServerName *.example.com ServerAdmin [email protected] VirtualDocumentRoot /home/%2/projects/%1 <Directory /home/*/projects/> Options FollowSymlinks Indexes IndexOptions FancyIndexing FoldersFirst AllowOverride All Order Allow,Deny Allow From All Satisfy Any </Directory> Alias /favicon.ico /var/www/default/favicon.ico Alias /robots.txt /var/www/default/robots.txt LogLevel warn # ErrorLog /home/%2/logs/%1.error.log # CustomLog /home/%2/logs/%1.access.log combined </VirtualHost>

    Read the article

  • Web application access different between Domain Name and IP

    - by h82
    in our office, we have a web application running. When we access the application by the domain name, http://server.domain.com/application/name it will display the current version of application. However, when we go by the IP address, http://192.168.1.111/application/name it will display the old version of that application. One thing is that we can access that application either by http://server.domain.com/ (it will be redirected to the long URL automatically) or http://server.domain.com/application/name when we are using domain name. But only accessible via the exact URL when we use IP address. Why is it showing the old version and how can it be corrected? It is running JRun4, Apache on Red hat. I've checked in httpd.conf a bit but could not find any. Please advice what should be done to display the same (updated version) when we access using domain name or IP address. Thank you.

    Read the article

  • Tunneling through SSH for 1521 port access?

    - by A T
    I am developing locally on my computer, using my own Apache server with PHP configured. My database however is remotely located on an Oracle 11g Database Server. We were also given a separate remote server for hosting our .html and .php files, however only FTP access has been provided there. Development is far too slow waiting for the FTP connection to push. So I decided to develop locally, but still use the remote DB server. Unfortunately that gives me an error. Not sure how—or where—to integrate tunnelling. Do I add something to the oci_connect HOST in my PHP file, or do I encapsulate my whole environment over SSH?

    Read the article

  • Setting up VPS hosting

    - by RobinFTW
    I'm trying to set up my own vps hosting. It wont be a paid service, just an experiment for me and some nerdy friends. What I'd like to be able to do is this: Run multiple virtual servers on 1 external IP. These servers can run anything from Minecraft servers to simple http servers. They will also need to be accessible thru SSH. What I don't get is how I can address these servers using domain names. I've done some research and found out that I could use Vhosts with Apache. However this only applies to http servers. It was also suggested I'd use a reverse proxy(squid) but this also only applies to http requests. I could just use different ports for different servers, but thats not ideal and not what I want. Can someone suggest a setup? Maybe some tutorials or anything.

    Read the article

  • Port forwarding using IP Tables

    - by Darider
    I have a server with a LAN facing address of 192.168.5.100 (eth0) and another LAN facing address of 192.168.6.6 (eth1). On this server I have a Virtualbox instance of fedora running an apache webserver (on port 8080) with a bridged interface to eth1 on the host server with address of 192.168.6.7 Users on the 192.168.6.x network can access the Webserver on the Vbox instance (192.168.6.7) with no problems. My question is what kind of iptables entries or commands should be made/executed so as to allow for users on the 192.168.5.x network to access the webserver on the vbox instance. (I'm hoping their url can be something like: http://192.168.5.100:8080)

    Read the article

  • Rejecting new HTTP requests when server reaches a certain throughput

    - by user56221
    I have a requirement to run an HTTP server that rejects new HTTP requests (with a 503, or similar) when the global transfer rate of current HTTP responses exceeds a certain level. For example, if the web server is transferring at 98Mbps, and a new HTTP request arrives, we would want to reject this (as we couldn't guarantee a good speed). I've had a look at mod_cband for Apache, limit_req for nginx, and lighttpd's rate limiting features, but none of them seem to handle my (rather contrived, granted) use case. I should add that I'm open to using pretty much any web server, and am open to implementing this in iptables rules if someone can craft such a rule! (Refusing the TCP connection is fine, it doesn't have to respond with an HTTP 503). Any suggestions?

    Read the article

  • Performance Drop Lingers after Load [closed]

    - by Charles
    Possible Duplicate: How do you do Load Testing and Capacity Planning for Databases I'm noticing a drop in performance after subsequent load tests. Although our cpu and ram numbers look fine, performance seems to degrade over time as sustained load is applied to the system. If we allow more time between the load tests, the performance gets back to about 1,000 ms, but if you apply load every 3 minutes or so, it starts to degrade to a point where it takes 12,000 ms. None of the application servers are showing lingering apache processes and the number of database connections cools down to about 3 (from a sustained 20). Is there anything else I should be looking out for here?

    Read the article

  • Need to configure multiple default gateways for four seperate physical network ports for a FreeBSD Webserver

    - by user20010
    I need to configure default gateways for four separate physical network interfaces for a FreeBSD Webserver. Basically, this is a web server that needs to be accessed by multiple WANS. I've been using various online resources, and a combination of setfib, pf, and ipfw. This web server will be deployed in multiple sites where access to next hop router info is not available, so we can't use static routes. We've used setfib to successfully create multiple routing tables and can ping beyond every default gateway we've created. Using setfib # ping ip.addr.what.ever we can ping anything available on a wan and beyond the router. The problem is we can't get Apache web server (port 80) traffic to route out when external users access the server(box). Multiple people have examples of binding setfib to ipfw commands, but none of them seem to work.

    Read the article

  • All HTTPS, or is it OK to accept HTTP and redirect (secure vs. user friendly)

    - by tharrison
    Our site currently redirects requests sent to http://example.com to https://example.com -- everything beyond this is served over SSL. For now, the redirect is done with an Apache rewrite rule. Our site is dealing with money, however, so security is pretty important. Does allowing HTTP in this way pose any greater security risk than just not opening or listening on port 80? Ideally, it's a little more user-friendly to redirect. (I am aware that SSL is only one of a large set of security considerations, so please make the generous assumption that we have done at least a "very good" job of covering various security bases.)

    Read the article

  • Running Debian as a Virtual Machine Host

    - by Tone
    I have a need to run a linux server to host subversion, bugzilla over Apache and also act as a file and print server. I also have a need to host Windows Server 2008 virtual machines for development purposes (I'm a .NET guy). The machine I have is a dual core AMD 2.5 Ghz w/2GB RAM. So here are my questions: 1) Should I run Debian (or some other distro) as the base AND create a Debian VM to host subversion, bugzilla and be my file/print server? or is it ok to use Debian as both my VM host and for those other reasons? 2) VMWare has a free server edition, Virtual Box also is another free option. Which one of these is better suited for what I need to do? Are there any other free (or inexpensive) alternatives out there? 3) Will I need a GUI with Debian in order to manage my VMs? 4) Can I run VMs without a GUI to conserve system resources?

    Read the article

  • I can't connect to my WAMP server from a client on the same network

    - by F4r-20
    Usually I would manually set up Apache and PHP but this time I thought I would go for the WAMP alternative. I've set up WAMP and it's working perfectly fine on the host machine. If I try and access the WAMP server from another machine on the same network (same subnet too: 192.168.1.*) I get this message: Forbidden You don't have permission to access / on this server. I have been online and gone through steps. I have changed Allow from 127.0.0.1 to Allow from all and it's still not working. Is there anything else I can try? Thanks.

    Read the article

  • nginx block URI request but allow internal directory

    - by Mike Anders
    I'm new to nginx from apache. I'm trying to simply block the URIs: /_mydir/* = / (redirect) But, I want to rewrite: /ex/(.*)$ = /_mydir/$1 I have tried: location /ex/ { rewrite ^/ex/(.*)$ /_mydir/$1 last; } location /_mydir { rewrite ^/_mydir/(.*)$ http://$http_host/ redirect; } But what always happens is once I block the '/_mydir' directory the rewrite is also blocked. I have also tried: location /_mydir/ { internal; } This also ends up blocking the rewrite. All help is greatly appreciated, thanks. UPDATE: I fixed this problem using: rewrite ^/ex/(.*)$ /_mydir/$1 break;

    Read the article

  • safe to setup NAS in virtual image with web server?

    - by Erik
    My current setup is Ubuntu Desktop, running Apache + MySQL + PHP stack. I use this image to host my small website to the outside world. I want to now run a virtual machine on the box, that will act as the FreeNAS box for my internal network. This sounds like a bad idea if the website gets hacked/attacked. Should I just give up and grab a dedicated machine instead? Can I instead virtualize the web server and NAS side by side on this machine?

    Read the article

  • How much ram to be able to convert large (5-6MB) jpegs? [closed]

    - by cosmicbdog
    I've got a project where we want to be processing large jpegs (5-6MB) with apache and php (using GD library). My understanding is that the server converts the image into a BMP making it quite ram heavy and currently we're unable to do it with our 1gb of memory. Here's the error we get: Fatal error: Allowed memory size of 67108864 bytes exhausted (tried to allocate 17408 bytes) How much ram should we be looking at running with to process images of this size? Edit: As Chris S the purist highlighted below, my post is apparently vague. I am doing the most basic and common manipulation of an image, say turning it from a 4352px x 3264px jpg of 5mb in size, to a 900px x 675px file.

    Read the article

  • Squid gives always tcp_miss reverse proxy

    - by JaakL
    I added installed latest squid3 in front of apache as reverse proxy. The problem is that it gives always tcp_miss, in fact I have not yet found a single TCP_HIT message in the log file, and most of the content is static. Relevant config values for cache_dir and refresh_pattern are default ones, directory /var/spool/squid3 exists and has some files/folders. I have 100+G free storage, but reconfigure gives warning "WARNING cache_mem is larger than total disk cache space!", which does not make any sense to me. I have googled a lot and seen with similar problems, but none of them has helped.

    Read the article

< Previous Page | 328 329 330 331 332 333 334 335 336 337 338 339  | Next Page >