Search Results

Search found 7082 results on 284 pages for 'trs 80'.

Page 43/284 | < Previous Page | 39 40 41 42 43 44 45 46 47 48 49 50  | Next Page >

  • Troubles doing transparent proxy for virtual machines

    - by Dan H
    Hi iptables gurus. First here is the basic topology: Internet | Gateway | Workstation---eth0---virbr0 | +-----+-----+ | | | vm1 vm2 vm3 I need to test a traffic analyzer running on my workstation, listening on some port (say 8990) on eth0. The rule [I think] I want is "any packets leaving virbr0 going anywhere to port 80 must instead go to port 8990 on eth0". My software running on port 8990 does its own check of the NAT packet mangling to push the packets through after it inspects them. I've been banging my head on this, with different variants of: iptables -t nat -A PREROUTING -i virbr0 -p tcp --dport 80 -j DNAT \ --to 10.0.0.10:8990 And I've tried the more generic method of using the mangle table with --set-mark and ip rule add fwmark, but I'm not getting it. I guess what's confusing me is that everything runs on the same box. Thanks for any guidance.

    Read the article

  • VPN pre-shared key problems

    - by Owl
    I have two vpns set up on a Symantec Gateway Security 320. VPN 1 goes to a Symantec Firewall/VPN 100 to another clinic of ours and every hour they lose connectivity and the error log on the Firewall/VPN100 shows an invalid pre-shared key error, although, both devices show the same pre-shared key entered. VPN 2 goes to our software vendor to use an additional part of our program. I am unable to ping the remote address and so is the other company, but my VPN status shows it is connected. They have told me the pre-shared key seemed to be automatically trying to resubmit itself as if it were incorrect, about every hour even though it is correct. They also told me port80 traffic was closed but I show the HTTP service using 80 redirected to 80 in my firewall settings. Please help.

    Read the article

  • Apache VirtualHost, multiple sites. 1 ssl with redirect and 1 regular http

    - by pedalpete
    I've got a server with one site which I am redirecting to https via <VirtualHost *:80> DocumentRoot /var/www/html/secure ServerName secure.com Redirect / https://secure.com </VirtualHost> That works no problem. Now I'm trying to add another non-secure site <VirtualHost *:80> DocumentRoot /var/www/html/notsecure ServerName notsecure.com </VirtualHost> of course, because the redirect is on '/', all sites are getting redicted. I've tried changing the Redirect to the full document root, but no luck.

    Read the article

  • Mixed IP and Name Based Virtual Hosts with nginx

    - by nerkn
    I set up many domains but I dont know how to configure if only ip address is given. say foo.com I have a setup to go web/foo.com/htdocs, I want to 88.99.66.55 ip address like a domain to web/fook.com/htdocs server { listen 80; server_name 85.99.66.55; location / { root /home/web/fook.com/htdocs; } location ~ \.(php|php3|php4|php5)$ { root /home/web/fook.com/htdocs; include fastcgi_params; fastcgi_pass 127.0.0.1:9000; } } resulted [warn]: conflicting server name "85.105.65.219" on 0.0.0.0:80, ignored

    Read the article

  • ssh can't connect after server ip changed

    - by Kery
    I have a server with ubuntu installed. After I change the network configuration and restart server, ssh client can't connect server any more. But in the server I can use ssh client to connect itself and the netstat command shows that sshd is listening port 22. And in my computer (win7) ping command is OK to server's new IP. The configuration in /etc/network/interfaces is: auto eth0 iface eth0 inet static address 10.80.x.x netmask 255.255.255.0 gateway 10.80.x.1 I'm very confused about this. Hope somebody can give me some idea. Thank you in advance!!!

    Read the article

  • output redirection not happening completely

    - by Gaurav
    I am running the following command on ubuntu 12.04 bash shell: hping3 -S -p 80 -c 30 www.google.com >> outputfile.txt But the problem is that this leaves some lines in the end and doesn't copy in the file. I even tried hping3 -S -p 80 -c 30 www.google.com >> outputfile.txt 2>&1 It also gives same problem. However this works perfect: ping -c 30 www.google.com >> outputfile.txt Any idea about what I might be missing?

    Read the article

  • Starting Apache through XAMPP is not working

    - by nogamo
    I've checked the error logs for this particular problem and it seems no access can be made to port 80. (OS 10013)An attempt was made to access a socket in a way forbidden by its access permissions. : make_sock: could not bind to address [::]:80 Turning the Windows firewall off didn't work, so I'm a bit lost. All of this is running on my laptop so I don't have my own "server" per se but I was under the impression that this would work without the need for one independent of my own personal computer. How would I solve this?

    Read the article

  • Outlook '10 hangs often, IMAP sync

    - by user23150
    We have 3 employees using IMAP to sync with their Desktop and Android Phones. Two are using 70% of their accounts storage, another is using 80%. They all have similar counts on folder structure, etc. The employee with the 80% storage is constantly having Outlook freeze on them for up to 5 minutes at a time. I realize this is Outlook connecting and doing activity on the server, but no one else has this problem. In fact, one user with the 70% storage used uses a very slow laptop, and doesn't have freezing issues. The network is the same, the settings are the same - at a loss how to proceed? Obviously "Outlook is a crummy IMAP client" doesn't help management...

    Read the article

  • Lighttpd based server issues crop up when port forwarding

    - by michael
    I have four host computers running lighttpd webservers. they are sitting behind a hspa modem, which each occupying a http port between [81 - 84]. 80 is taken by the modem itself. The port forwarding is setup correctly, however, only a portion of any webpage I request from any of the hosts comes through (they all fails after %20 of the page). If I put the host on port 81 into the dmz, it serves pages fine. The others do not respond to the dmz treatment. Is it possible the web content on the hosts somehow require ports aside from their respective http port? Or is it possible that even though the server.port in the lighttpd_ssl.conf file is set, the individual hosts are still expecting to serve on port 80? I am not familiar with lighttpd, nor did i set them up. they are running on video encoders i purchased. I can grab any files from them required for further information on the problem.

    Read the article

  • ZendServer uninstall and xampp install ports conflict

    - by BlackFire27
    I uninstalled zend, cleaned it from the registry (it doesnt exist there).. when I go to localhost , I can see its favicon. Also my xampp port is blocked by the previous installation of zend that uses the port 80.. so I swapped xampps port to 8080.. it works..but I would rather to use port of 80.. is there a way to achieve so? perhaps changing zends port to something else? or how can I find who is listening to the same port.. By what I can see is that zend is still using the localhost..how i sthat possible

    Read the article

  • Proxy between data centers [closed]

    - by dstarh
    Possible Duplicate: Can IIS be configure to forward request to another web server? We are switching data centers (actually datac-center to EC2 to be specific) and some customers have not yet made DNS changes to point the domains to the new load balancers. We are thinking of leaving the existing servers up and just using a proxy server to forward the request to the new load balancer. Can anyone recommend a good proxy server for doing this I've got squid installed but it seems it's fairly easy to just have a fairly wide open proxy server and we don't want this. I want all requests coming in on port 80 to be sent to port 80 at a specific domain (elastic load balancer) the data center env is windows 2k3 and the EC2 env will all be linux but the ec2 env should be irrelevant

    Read the article

  • wamp remote access

    - by user1589779
    I have a wamp server (Version 2.2 on Windows 7) running on my computer that works perfectly well for localhost access. But now, I need to grant access to a collegue on the same subnetwork. How to achieve that? My collegue receives a 403 error from the browser. I have 2 vhost : <VirtualHost *:80> ServerName localhost DocumentRoot "D:/Workspace" </VirtualHost> <VirtualHost *:80> ServerName oscar DocumentRoot "D:/Workspace/SMACS3/web/htdocs" </VirtualHost>

    Read the article

  • HAproxy to web host sub directory?

    - by daemonza
    Hi for reasons outside my control, I need to load balance two servers, that run a non-virtual host enabled app on IIS. Normally in HAProxy I would load balance servers(apache, tomcat, etc) like this : acl is_www_example_com hdr_end(host) -i www.example.com use_backend www_example_com if is_www_example_com backend www_example_com balance roundrobin cookie SERVERID insert nocache indirect option httpchk HEAD / HTTP/1.0 option httpclose option forwardfor server node1 192.168.1.1:80 cookie node1 server node1 192.168.1.2:80 cookie node1 Which will route to the node 1 and node 2 server and serve up the virtual host site. if I need to route to www.example.com/application/data How would I be able to do it, with the above example, if at all even possible?

    Read the article

  • EC2: map multiple applications to different domains

    - by EsseTi
    i'm playing with EC2 and i've been able to create my instance that has a django appliacation on port 80, and a tomcat on 8080. now, with elastic IP i can manage to redirect my domain to django application. now i would like to map subdomains to each tomact applications. for example django app (ec2...:80) --> mydomain.com tomcat (ec2...:8080) --> tomcat.mydomain.com webbapp1 (ec2...:8080/webapp1/) --> webapp1.mydomain.com is this possible with the free account? ciao

    Read the article

  • Installing CURL on Ubuntu Karmic

    - by Racertim
    Trying to get this up and running: https://github.com/cloudnull/massupload I have everything except CURL installed and when I attempt to, it fails with the following: Reading package lists... Done Building dependency tree Reading state information... Done The following NEW packages will be installed: curl 0 upgraded, 1 newly installed, 0 to remove and 3 not upgraded. Need to get 196kB of archives. After this operation, 311kB of additional disk space will be used. WARNING: The following packages cannot be authenticated! curl Install these packages without verification [y/N]? y Err http://us.archive.ubuntu.com karmic/main curl 7.19.5-1ubuntu2 404 Not Found [IP: 91.189.91.13 80] Failed to fetch http://us.archive.ubuntu.com/ubuntu/pool/main/c/curl/curl_7.19.5-1ubuntu2_i386.deb 404 Not Found [IP: 91.189.91.13 80] Thank you!

    Read the article

  • problem with accessing a php page

    - by EquinoX
    So I have a info.php page which is located on the folder /var/www/nginx-default, however when I go to my ip address/info.php, it always redirects me to this site: http://www.iana.org/domains/example/ is this because I have a virtual host that I called example? Here is my config for the example website: server { listen 80; server_name www.example.com; rewrite ^/(.*) http://example.com/$1 permanent; } server { listen 80; server_name example.com; access_log /var/www/example.com/logs/access.log; error_log /var/www/example.com/logs/error.log; location / { root /var/www/example.com/public/; index index.html; } } The way I access this site is by changing my /var/hosts in my macbook so that example.com is mapped to my server IP address... however now when I do xxx.xxx.xxx.xxx/info.php.. it redirects me to that site I posted above

    Read the article

  • How to make your apache application accessible within network

    - by guest
    I have a Windows XP machine where I have installed WAMP and made a PHP based web application. I can access the web application from within this machine by using the browser and pointing to: http://localhost/myApp/ --- and the page loads fine. Now I want this site (http://localhost/myApp) to be accessible to all machines within the network (and may be later, to the general public as well). I am quite new to this, how do I make my site accessible to all machines within the network and to the general public in the internet? I tried modifying the httpd.conf file in Apache (WAMP) by changing Listen 80 to Listen 10.10.10.10:80 (where I replaced 10.10.10.10 with the actual IP of this windows xp machine). I also tried "Put Online" feature in WAMP. None seem to work though. How do I make it accessible?

    Read the article

  • Easiest method of creating multiple Word documents with incremental number?

    - by DakotaDusty
    I need to create about 80 Word documents that are identical except for a numerical identifier in the document. The identifier is the same as the document filename, eg. the document named "SN-100.doc" must have the string "SN-100" inside the document text. Each unique document will reside in a folder location given by its unique numerical identifier.For example, the file directory hierarchy will look something like this: /SN001/SN-1.doc /SN002/SN-2.doc . . . /SN080/SN-80.doc What is the easiest and fastest method of doing this?

    Read the article

  • domain works randomly

    - by mthenw
    Hi, I have problem with configuring bind9 on debian (lenny) server. Generaly speaking everything is working ok but sometimes I get 404 on few domains (eg. 4stopnie.com but after few refreshes in browser site loads) or I can't validate site with validator.w3.org (error '500 Can't connect to 4stopnie.com:80 (Bad hostname '4stopnie.com')'). Domains were moved from other server. After moving I changed serial number in zone file. $ttl 600 @ IN SOA ns.wpoznaniu.info. xxx.4stopnie.com. ( 2011011601 3600 600 86400 600) @ IN NS ns.wpoznaniu.info. @ IN A 80.82.21.196 www IN CNAME @

    Read the article

  • Security against IP spoofing [on hold]

    - by user1369975
    I am pursuing a college project, in which I am running three fake services on three ports to protect the main service (say running at port 80). The concept is that if the user is malicious, he'll try to bring the services down and access the fake services. These ports adopt a blocking process of a connection request and record the IP and port of the client. These are logged and aren't granted access on service on port 80. But what to do if the client spoofs his IP? How can I modify my system?

    Read the article

  • radvd won't accept non-/64 subnets

    - by Dolda2000
    I'm trying to set up radvd on a 6RD configuration (on Linux), where I have a /64 subnet, and I'm trying to use it on two distinct physical networks, so I'm trying to give each of them a /80 subnet. However, when I start radvd on these interfaces, it says this: radvd[3987]: prefix length should be 64 for int (int being the name of one of the interfaces.) I know that stateless autoconfiguration doesn't work on subnets that are larger than /64, like a /48 subnet, but AFAIK it's supposed to work on anything that is smaller than /64, so /80 shouldn't be a problem. The final effect, now, is that radvd simply advertises the wrong subnet prefix, effectively "removing" the first 16 bits of it. What is this? Am I all wrong on that, is radvd buggy, or is there some configuration option that I'm missing?

    Read the article

  • apache 2.2: how can I set up a VirtualHost inside the RootDirectory?

    - by redraw
    I want to set up a VirtualHost inside the RootDirectory. For example, My project is in C:/myproject and I want to access with http://localhost/myproject EDIT: I've made an alias inside the httpd-vhosts.conf, however I don't have permissions. <VirtualHost *:80> DocumentRoot "C:/apache-2.2/htdocs" ServerName localhost Alias /test "D:\arbol\documentos\test" </VirtualHost> Is this code below the proper way to give permissions? <VirtualHost *:80> DocumentRoot "C:/apache-2.2/htdocs" ServerName localhost Alias /test "D:\arbol\documentos\test" <Directory "D:\arbol\documentos\test"> allow from all order allow,deny AllowOverride All </Directory> </VirtualHost>

    Read the article

  • Need help with setting up comet code

    - by Saif Bechan
    Does anyone know off a way or maybe think its possible to connect Node.js with Nginx http push module to maintain a persistent connection between client and browser. I am new to comet so just don't understand the publishing etc maybe someone can help me with this. What i have set up so far is the following. I downloaded the jQuery.comet plugin and set up the following basic code: Client JavaScript <script type="text/javascript"> function updateFeed(data) { $('#time').text(data); } function catchAll(data, type) { console.log(data); console.log(type); } $.comet.connect('/broadcast/sub?channel=getIt'); $.comet.bind(updateFeed, 'feed'); $.comet.bind(catchAll); $('#kill-button').click(function() { $.comet.unbind(updateFeed, 'feed'); }); </script> What I can understand from this is that the client will keep on listening to the url followed by /broadcast/sub=getIt. When there is a message it will fire updateFeed. Pretty basic and understandable IMO. Nginx http push module config default_type application/octet-stream; sendfile on; keepalive_timeout 65; push_authorized_channels_only off; server { listen 80; location /broadcast { location = /broadcast/sub { set $push_channel_id $arg_channel; push_subscriber; push_subscriber_concurrency broadcast; push_channel_group broadcast; } location = /broadcast/pub { set $push_channel_id $arg_channel; push_publisher; push_min_message_buffer_length 5; push_max_message_buffer_length 20; push_message_timeout 5s; push_channel_group broadcast; } } } Ok now this tells nginx to listen at port 80 for any calls to /broadcast/sub and it will give back any responses sent to /broadcast/pub. Pretty basic also. This part is not so hard to understand, and is well documented over the internet. Most of the time there is a ruby or a php file behind this that does the broadcasting. My idea is to have node.js broadcasting /broadcast/pub. I think this will let me have persistent streaming data from the server to the client without breaking the connection. I tried the long-polling approach with looping the request but I think this will be more efficient. Or is this not going to work. Node.js file Now to create the Node.js i'm lost. First off all I don't know how to have node.js to work in this way. The setup I used for long polling is as follows: var sys = require('sys'), http = require('http'); http.createServer(function (req, res) { res.writeHead(200, {'Content-Type': 'text/html'}); res.write(new Date()); res.close(); seTimeout('',1000); }).listen(8000); This listens to port 8000 and just writes on the response variable. For long polling my nginx.config looked something like this: server { listen 80; server_name _; location / { proxy_pass http://mydomain.com:8080$request_uri; include /etc/nginx/proxy.conf; } } This just redirected the port 80 to 8000 and this worked fine. Does anyone have an idea on how to have Node.js act in a way Comet understands it. Would be really nice and you will help me out a lot. Recources used An example where this is done with ruby instead of Node.js jQuery.comet Nginx HTTP push module homepage Faye: a Comet client and server for Node.js and Rack To use faye I have to install the comet client, but I want to use the one supplied with Nginx. Thats why I don't just use faye. The one nginx uses is much more optimzed. extra Persistant connections Going evented with Node.js

    Read the article

< Previous Page | 39 40 41 42 43 44 45 46 47 48 49 50  | Next Page >