Search Results

Search found 16467 results on 659 pages for 'request filtering'.

Page 425/659 | < Previous Page | 421 422 423 424 425 426 427 428 429 430 431 432  | Next Page >

  • Firewall blocks outgoing email

    - by Martin Trigaux
    On my Debian server running a Django website, I have an error when I need to send an email. The error received is Exception Type: gaierror Exception Value: [Errno -2] Name or service not known Exception Location: /usr/lib/python2.6/socket.py in create_connection, line 547 You can see the full error log here. After testing, it seems it is my firewall that blocks the request. You can see my iptable file (/etc/init.d/firewall). I think the problem comes from the two commented lines that were supposed to accepts all established connections. When I uncomment them, I have an error iptables: No chain/target/match by that name. Thank you

    Read the article

  • node.js server not running

    - by CMDadabo
    I am trying to learn node.js, but I'm having trouble getting the simple server to run on localhost:8888. Here is the code for server.js: var http = require("http"); http.createServer(function(request, response) { response.writeHead(200, {"Content-Type": "text/plain"}); response.write("Hello World"); response.end(); }).listen(8888); server.js runs without errors, and trying netstat -an | grep 8888 from terminal returns tcp4 0 0 *.8888 *.* LISTEN However, when I go to localhost:8888 in a browser, it says that it cannot be found. I've looked at all the related questions, and nothing has worked so far. I've tried different ports, etc. I know that my router blocks incoming traffic on port 8888, but shouldn't that not matter if I'm trying to access it locally? I've run tomcat servers on this port before, for example. Thanks so much for your help! node.js version: v0.6.15 OS: Mac OS 10.6.8

    Read the article

  • Moving cpanel backup of magento site to VPS

    - by user2564024
    I was having my site in shared hosting, I took the entire backup, its structure is like addons homedir mysql resellerpackages suspendinfo bandwidth homedir_paths mysql.sql sds userconfig counters httpfiles mysql-timestamps sds2 userdata cp locale nobodyfiles shadow va cron logaholic pds shell vad digestshadow logs proftpdpasswd ssl version dnszones meta psql sslcerts vf domainkeys mm quota ssldomain fp mma resellerconfig sslkeys has_sslstorage mms resellerfeatures suspended Now I have subscribed to vps, I have copied the files inside homedir/public_html to var/www/html of my new hosting, but am seeing the following error when I view it browser, There has been an error processing your request Exception printing is disabled by default for security reasons. Error log record number: 259343920016 I have just created database with name magenhto inside mysql. Previously I had cpanel and used one click installer. Hence am not aware of how to use that data inside mysql to this new system and are there any more changes.

    Read the article

  • solving network connection problem on xp

    - by doug
    Hi there I feel so stupid, because I have to fix my dad computer(which was mine years ago). So it is an old computer on which I have installed windows xp and the drivers for main-board. The device manager says that I have no more problems with the hardware. That computer has to connect to a router from my room but it doesn't. On ping 192.168.2.1 I get "Request time out"., The other computers are connecting to that router. Only the old computer doesn't connect to the LAN and later to the internet. Also it doesn't get a good IP(from the same class). After running ipconfig command I know that i have an IP with 169.... What can I do? Do you have any tips for me?

    Read the article

  • How do I set up my home server to go directly to a port other than 80

    - by Kevin
    I'm using dyndns, a lynksis wt54g router, and tomcat 7 with spring to set up a web server. This is my first time to attempt this. I'm sure this is a very common question, but I don't know enough to find the answer after quite a bit searching. Dyndns is successfully forwarding to my ip. The main problem is, the router admin login is coming up when my url is used. I'm hosting my site on port 8080. I have port forwarding set up for port 8080 but my request times out when I attempt to use my url like this www.myurl1234.com:8080. I don't want users to have to type the port anyway. I also tried changing the management port to 82 and hosting on port 80, but I still get the router admin login when I use my url. Where am I going wrong? Can I set it up so that www.myurl1234.com goes straight to port 8080?

    Read the article

  • How can I get a windows server to collate and email 4 daily reports/graphs for server performance

    - by Glyn Darkin
    I run a windows 2008 webserver and would like to setup the most basic performance monitoring in the world. What i would like is: a plot of ASP.Net request time for each w3wp process for a 24hour period a plot of CPU% utilisation for each w3wp process for a 24hour period a plot of Memory utilisation for each w3wp process for a 24hour period a plot of network utilisation for each w3wp process for a 24hour period a plot of disk utilisation for each w3wp process for a 24hour period I would these plots to be emailed to me each morning. Anybody know what is the simplest way to set this up? Thanks for your help in advance. Glyn

    Read the article

  • ConfigurationErrorsException when serving images via UNC on IIS6

    - by Mark Richman
    I have a virtual directory in my web app which connects to a Samba share via UNC. I can browse the files via Windows Explorer without issue, but my web app throws a yellow screen with the following message: Description: An error occurred during the processing of a configuration file required to service this request. Please review the specific error details below and modify your configuration file appropriately. Parser Error Message: An error occurred loading a configuration file: Could not find file '\cluster\cms\qa-images\120400\web.config'. What makes no sense to me is why it's looking for a web.config in that location. I know it's not an authentication issue because the virtual directory can serve images from its root (i.e. \cluster\cms\qa-images\test.jpg serves as http://myserver/upload/test.jpg just fine).

    Read the article

  • Windows 8.1 keeps prompting for Network Share Credentials after every log on or restart

    - by Peret del Trunfa
    I have a Network drive Shared in a Workgroup with 3 clients. Two clients with Windows 7 have persistent connections to the Share. No issues with those two. My windows 8.1 client keeps prompting for credentials at every restart / log on. I spent hours looking around for a solution: I have stored cred in cred manager, and tried every possible combination (WORKGROUP\user , COMPUTERNAME\user, user, .. and so on). I have changed NT and NTLM negotiation in policy manager. I've compared the settings under GPO network security with a working win 7 computer, everything is pretty much the same. -I've captured Wireshark to see SMB negotiation process, honestly I see the messages flowing around, and the share sending AUTH DENIED.. which means is how the 8.1 client formats the request.... that makes the share reject it.. Now I still don't really know why. Any ideas would be appreciated.

    Read the article

  • HAProxy overload protection

    - by user2050516
    using the HAProxy, would it be possible to configure an overload protection, to limit the amount of requests sent to the backing http server(s) to a given rate (z.B 100 Request per second ). If the threshold is exceeded requests should be answered with a default response. I am interested in requests per second not connections per second as a connection can have many requests. And yes to improve the servers is not an option here. If yes a configuration example to achieve that would be excellent. Thank you in advance.

    Read the article

  • iptables block everything except http

    - by arminb
    I'm trying to configure my iptables to block any network traffic except HTTP: iptables -P INPUT DROP #set policy of INPUT to DROP iptables -P OUTPUT DROP #set policy of OUTPUT to DROP iptables -A INPUT -p tcp --sport 80 -m state --state ESTABLISHED,RELATED -j ACCEPT iptables -A OUTPUT -p tcp --dport 80 -m state --state NEW,ESTABLISHED -j ACCEPT The iptables output (iptables -L -v) gives me: Chain INPUT (policy DROP 0 packets, 0 bytes) pkts bytes target prot opt in out source destination 4 745 ACCEPT tcp -- any any anywhere anywhere tcp spt:http state RELATED,ESTABLISHED Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) pkts bytes target prot opt in out source destination Chain OUTPUT (policy DROP 0 packets, 0 bytes) pkts bytes target prot opt in out source destination 2 330 ACCEPT tcp -- any any anywhere anywhere tcp dpt:http state NEW,ESTABLISHED When I try to wget 127.0.0.1 (yes i do have a web server and it works fine) i get: --2012-11-14 16:29:01-- http://127.0.0.1/ Connecting to 127.0.0.1:80... The request never finishes. What am I doing wrong? I'm setting iptables to DROP everything by default and add a rule to ACCEPT HTTP.

    Read the article

  • Is this bad practice?

    - by gekod
    I just wanted to ask for your opinion on a situation that occurs sometimes and which I don't know what would be the most elegant way to solve it. Here it goes: We have module A which reads an entry from a database and sends a request to module B containing ONLY the information from the entry module B would need to accomplish it's job (to keep modularity I just give it the information it needs - module B has nothing to do with the rest of the information from the read DB entry). Now after finishing it's job, module B has to reply to a module C if it succeeded or failed. To do this module B replies with the information it has gotten from module A and some variable meaning success or fail. Now here comes the problem: module C needs to find that entry again BUT the information it has gotten from module B is not enough to uniquely find the exact same entry again. I don't think that module A giving more information to module B which it doesn't need to do it's job but which it could then give back to module C would be a good practice because this would mean giving some module information it doesn't really need. What do you think?

    Read the article

  • How to setup VPN on home network

    - by Tone
    I am a software developer. I travel and sometimes have a need to access my files at home and tweak other family members computers. I would like to connect to my home network via VPN and then RDP into whatever machine i need to. Currently i have a Windows Server 08 machine, which is my file server, database server, web server (for development work), source control repository, etc. (and also somewhat of a workstation when i need it to be). I want to use this same machine to run my VPN through. I have a linksys WRTG54 router. My ISP is AT&T DSL, with a dynamic IP address - so i'm assuming I'll either need to request a static IP or sign up with one of those static ip services.. where it keeps your dynamic ip synced up with a static one. While I do understand software engineering I am no expert in networking. What do i need to do to setup my VPN?

    Read the article

  • Unable to open websites that use HTTPS on linux

    - by negai
    I have the following network configuration: My PC 192.168.1.20/24 uses 192.168.1.1/24 as a gateway. Dlink-2760U router with Local address 192.168.1.1/24 has a VPN connection open with the provider using PPTP. Whenever I'm trying to open some web-sites that has some authorization (e.g. gmail.com, coursera.org), I'm getting a request timeout. This problem is observed mostly on linux (Ubuntu 12.04 and Debian 6.0), while most of such websites work correctly on windows XP. Could you please help me diagnose the problem? Could it be related to NAT + HTTPS? Thanks

    Read the article

  • Hide the ports in the url of a tomcat application

    - by S.M.09
    I have setup tomcat 6 on a server with several different instances which run on different ports For Eg: http://10.XXX.XXX.XXX:8080/myAPP1 http://10.XXX.XXX.XXX:8088/myAPP2 http://10.XXX.XXX.XXX:8082/myAPP3 and so on... These urls are available two the users using these and other applications Now I want to hide the ports on which each of the applications are running... So that the user could see only something like this http://10.XXX.XXX.XXX/myAPP1 http://10.XXX.XXX.XXX/myAPP2 http://10.XXX.XXX.XXX/myAPP3 What is needed to be done for this I have achieved this for one url (i.e. one port) using the firewall rules and redirect the request on HTTP 80 port to the port that tomcat is listening.. But I want such patter to appear for all the urls... How can I go about this? Specs HOST M/C : 64 bit OS: Linux Redhat Tomcat : 6.0

    Read the article

  • Crontab + .sh + php

    - by Kristaps Karlsons
    Hi. I'm trying to call a shell script every 5 minutes, witch executes php file under root. # crontab -l */5 * * * * /home/regularuser/call.sh permissions: -rw-rw-rw- 1 root root 162 Jun 6 23:40 call.php -rwxr-xr-x 1 root root 66 Jun 7 01:20 call.sh call.sh contents: #!/bin/bash php -q /home/regularuser/call.php echo "request processed" My problem is that my php file doesn't get executed via crontab. However, if I call call.sh - everything works perfectly. I'm new to crontab and shell scripting, so any advice/resources are welcome.

    Read the article

  • Architecture for dashboard showing aggregated stats [on hold]

    - by soulnafein
    I'd like to know what are common architectural pattern for the following problem. Web application A has information on sales, users, responsiveness score, etc. Some of this information are computationally intensive and or have a complex business logic (e.g. responsiveness score). I'm building a separate application (B) for internal admin tasks that modifies data in web application A and report on data from web application A. For writing I'm planning to use a restful api. E.g. create a new entity, update entity, etc. In application B I'd like to show some graphs and other aggregate data for the previous 12 months. I'm planning to store the aggregate data for each month in redis. Some data should update more often, e.g every 10 minutes. I can think of 3 ways of doing this. A scheduled task in app B that connects to an api of app A that provides some aggregated data. Then app B stores it in Redis and use that to visualise pages. Cons: it makes complex calculation within a web request, requires lot's of work e.g. api server and client, storing, etc., pros: business logic still lives in app A. A scheduled task in app A that aggregates data in an non-web process and stores it directly in Redis to be accessed by app B. A scheduled task in app A that aggregates data in a non-web process and uses an api in app B to save it. I'd like to know if there is a well known architectural solution to this type of problems and if not what are other pros/cons for the solution I've suggested?

    Read the article

  • Sun Directory Server 5.2 performance

    - by tmow
    Hi all, I'm using logconv.pl (provided by Sun), to measure performance on my server. These two metrics results, are worrying me a bit: Binds: 192164 Unbinds: 111569 In fact the difference between the two it's quite big, how can I determine which are the unbound requests? As stated by Lodovic: Many applications just close the connections without sending an Unbind request. This simply can explain the difference. But the logconv.pl doesn't show details about the unbound requests, do you know any other tools or can you suggest some queries or whatever that can help me find out the root cause? Do you think anyway that the performances may improve fixing the issue?

    Read the article

  • Any recommendations on a NAS for a home-super-user?

    - by marc_s
    Can anyone recommend a good NAS for use in a home-server environment? I would request at least 2, preferably 4 disks, and I am most interested in good to excellent throughput for file-server and backup purposes - don't need any of the fancy media-streaming or -sharing features, that's not of interest to me. For a 4 or more disk solution, support for the various RAID levels (0, 1, 1+0, 5) would be a plus - especially if supported in hardware (rather than just a software emulation). I just need a place to put my collection of data, ISO images, and so forth - and since several external disks (self-built and off-the-shelf) have failed so far, I'm looking into a more reliable solution. Marc

    Read the article

  • Redirecting traffic to a https site

    - by Saif
    In our mail server there are multiple virtual email domain hosted. User can use webmail.example.com (this is the first apache virtualhost) to check mail or they can use mail.THEIR-DOMAIN.com. If the put mail.THEIR-DOMAIN.com apache shows the webmail.example.com as it is the first virtualhost and mail.THEIR-DOMAIN.com doesn't exist. Recently we imposed https for webmail.example.com and added this mod_rewrite rule: RewriteEngine On RewriteCond %{HTTPS} off RewriteRule (.*) https://%{HTTP_HOST}%{REQUEST_URI} But now users not getting the default webmail page as they were getting before. How can we redirect all request coming to the URL "mail.ANY-DOMAIN.com" to "https://webmail.example.com"? I've tried the following but it didn't work: RewriteEngine On RewriteCond %{HTTPS} off RewriteRule (mail.*) https://webmail.example.com Thanks for your help in advance.

    Read the article

  • Rejecting new HTTP requests when server reaches a certain throughput

    - by user56221
    I have a requirement to run an HTTP server that rejects new HTTP requests (with a 503, or similar) when the global transfer rate of current HTTP responses exceeds a certain level. For example, if the web server is transferring at 98Mbps, and a new HTTP request arrives, we would want to reject this (as we couldn't guarantee a good speed). I've had a look at mod_cband for Apache, limit_req for nginx, and lighttpd's rate limiting features, but none of them seem to handle my (rather contrived, granted) use case. I should add that I'm open to using pretty much any web server, and am open to implementing this in iptables rules if someone can craft such a rule! (Refusing the TCP connection is fine, it doesn't have to respond with an HTTP 503). Any suggestions?

    Read the article

  • wamp server does not start: Windows 7, 64Bit

    - by Ram
    I am trying to install wamp server (following is the exact setup name) on windows 7, 64Bit OS. But it never starts, icon stays in orange color meaning some services did not start. wampserver2.2e-php5.3.13-httpd2.2.22-mysql5.5.24-x64 I have been searching from last 3hours but did not find any solution. Port 80 is not in use. In windows services, when I try to start wampapache service manually, it throws following error: Windows could not start the wampapache service on Local Computer. Error 1053: The service did not respond ot the start or control request in a timely fashion. apache_error.log is empty. Things use to work fine in windows XP. May be this is a repeated thread, but I did go through similar posts. But nothing worked! Please help!!

    Read the article

  • Make nginx avoid cache if response contains Vary Accept-Language

    - by gioele
    The cache module of nginx version 1.1.19 does not take the Vary header into account. This means that nginx will serve the same request even if the content of one of the fields specified in the Vary header has changed. In my case I only care about the Accept-Language header, all the others have been taken care of. How can I make nginx cache everything except responses that have a Vary header that contains Accept-Language? I suppose I should have something like location / { proxy_cache cache; proxy_cache_valid 10m; proxy_cache_valid 404 1m; if ($some_header ~ "Accept-Language") { # WHAT IS THE HEADER TO USE? set $contains_accept_language # HOW SHOULD THIS VARIABLE BE SET? } proxy_no_cache $contains_accept_language proxy_http_version 1.1; proxy_pass http://localhost:8001; } but I do not know what is the variable name for "the Vary header received from the backend".

    Read the article

  • Do I really need to reboot for AD changes to be applied?

    - by stimms
    Every time I request a permission change the IT group at my company instructs me to wait 20 minutes and reboot the computer. I cannot believe that in this day and age you still need to reboot the computer to clear whatever cache stores the permissions locally. It feels like something out of the NT 4 days. Do you actually still need to reboot the computer? Is a logout/login sufficient? Is there still a long time(20 minutes) for the changes to propagate through the AD tree?

    Read the article

  • Cannot Ping PS3 Using a Wired Connection

    - by Post
    PS3 is using the inbuilt wireless network adapter (I cannot change this) When I try to ping the PS3 from ANY computer which is on a wired ethernet connection, I get Request Timed Out errors. Whenever I ping from a computer with a wireless connection, it works just fine. To be clear: Pinging from Wireless PC to Wireless PS3 works Pinging from Wired PC to Wireless PS3 fails I have tried this on several PC's and Laptops all with the same results. As an attempted solution I have set up static IP's on all related devices. More information: Default Gateway = 192.168.2.1 PS3(wireless) = 192.168.2.100 PC(wired) = 192.168.2.99 Subnet Mask(for both devices, I have made sure) = 255.255.255.0 Thanks

    Read the article

  • Customer Reviews on Company Listings [closed]

    - by GSTAR
    I'm not sure if this is the right place but I am after some general advice on a feature I am looking to implement on my website. The website I have is a Wedding Directory. Here, companies can advertise in the form of a directory listing. This listing contains the company details - such as what they do and how you can contact them. There are three packages available - Basic package is free and Silver and Gold packages are charged for. Now in order to further enhance the directory, I want to enable customer reviews for each listing. This is basically whereby customers who have used a particular company's service can write a review (and rate out of 5) of their experience. The dilemma I face is that if customers are to be paying for their listings then surely they will expect to not have content on their listing that would tarnish their reputation (i.e. negative reviews). But at the same time I want to be impartial and help my site visitors to make informed choices on which companies to book when arranging their weddings. Of course I will exercise my ability to remove offensive or fake reviews but I do not want to be removing reviews just to satisfy my clients. Suppose a client pays me a premium fee to have their listing on my front page and at the top of the listings. Now suppose that client gets lost of negative reviews - the fact that they have increased visibility means that they will also get increased bad publicity. This in turn means they won't renew their package and will most likely request the listing to be taken down. So, how can I get the right balance here and keep everyone happy?

    Read the article

< Previous Page | 421 422 423 424 425 426 427 428 429 430 431 432  | Next Page >