Search Results

Search found 57596 results on 2304 pages for 'arraku gmail com'.

Page 341/2304 | < Previous Page | 337 338 339 340 341 342 343 344 345 346 347 348  | Next Page >

  • Netbeans mySQL Connection error

    - by user125469
    Hi everybody I'm new and I'm having some trouble with a connection with netbeans and mysql, I need a database online I created one at 000webhost.com but I think that something's wrong .... if any have some idea or other server I really need to connect a database from a server online with an app in netbeans... Please I'm new in this all ideas you have come on.. here it's a pic http://i.stack.imgur.com/KA8iS.png http://s2.subirimagenes.com/imagen/7786594fail.png

    Read the article

  • Cant checkout from WebSVN

    - by yoasqie
    I am trying to download a script from SVN. The url looks like this: http://websvn.myurl.com/list.php?repname=name_1 I am trying to download using: svn checkout --username myuser --password mypass "http://websvn.myurl.com/list.php?repname=name_1" I got "URL is not properly URI-encoded", then I changed the "?" for "%3F", and the output shows: svn: OPTIONS of 'http://websvn.myurl.com/list.php%3Frepname=name_1': 200 OK (http://websvn.myurl.com/) But nothing gets downloaded. Sorry I can't post the real URL, but it's a private project and I don't have anyone else online to ask. What am I doing wrong?

    Read the article

  • How to forbid postfix to send to external domains [closed]

    - by elhoim
    I have a local postfix server, and i want it to only relay emails to the only local domain (localdomain.be): myhostname = localdomain.be mydomain = localdomain.be alias_maps = hash:/etc/aliases alias_database = hash:/etc/aliases myorigin = $myhostname mydestination = $myhostname relay_domains = $mydomain default_transport = smtp relayhost = mynetworks = 127.0.0.0/8 [::ffff:127.0.0.0]/104 [::1]/128 10.0.0.0/24 mailbox_size_limit = 64000000 message_size_limit = 1000000 recipient_delimiter = + inet_interfaces = all inet_protocols = all smtp_host_lookup = native This configuration works fine to allow relay mail locally and on external destination domains, but i would like it to be an impossibility to send to other domains (ie: gmail.com). relay_domains is supposed to ensure that but it does not seem to really filter since i can still send to my gmail address.

    Read the article

  • If email not received then do X (outlook 2013 on Exchange 2010)

    - by Brad
    I receive notification emails daily and would like to automate an easier way to manage all of those notifications. For example: Notification 1 from asd@qwerty.com is received daily between 10pm-1am Notification 2 from asd@qwerty.com is received daily between 12am-3am Notification 3 from qwe@asdfg.com is received daily between 1am-4am I am looking for a way to page myself at 5555551212@vzwtext.com on my cellphone if any of these messages are not received within the defined time frame of when the email should have arrived. I would like to basically email a page like: ATTENTION Notification 2 not received within the allowed range. This way I would be notified instead of having to check the email manually and see that I only received 2 of the three alerts. Is there a way to do this in Outlook? Our exchange server is a hosted exchange server on GoDaddy if that info is needed.

    Read the article

  • nginx root directory not forwarding correctly

    - by user66700
    The server files are store in /var/www/ Everything was working perfectly, then I've been getting the following errors 2011/01/28 17:20:05 [error] 15415#0: *1117703 "/var/www/https:/secure.domain.com/index.html" is not found (2: No such file or directory), client: 119.110.28.211, server: secure.domain.com, request: "HEAD /https://secure.domain.com/ HTTP/1.1", host: "secure.domain.com" Heres my config: server { server_name secure.domain.com; listen 443; listen [::]:443 default ipv6only=on; gzip on; gzip_comp_level 1; gzip_types text/plain text/html text/css application/x-javascript text/xml text/javascript; error_log logs/ssl.error.log; gzip_static on; gzip_http_version 1.1; gzip_proxied any; gzip_disable "msie6"; gzip_vary on; ssl on; ssl_ciphers RC4:ALL:-LOW:-EXPORT:!ADH:!MD5; keepalive_timeout 0; ssl_certificate /root/server.pem; ssl_certificate_key /root/ssl.key; location / { root /var/www; index index.html index.htm index.php; } }

    Read the article

  • IIS doesn't respond to 127.0.0.1 (external IP works fine)

    - by Jordan
    I have an AWS web server - call it box.company.com. It's running IIS and if I visit http://box.company.com in a web browser (from any machine, including box.company.com), it responds correctly with our site. However, if I visit localhost/ or 127.0.0.1/ when I'm logged into box.company.com, I get a "couldn't connect to host" message. The hosts file has only one entry - the standard "127.0.0.1 localhost" line. Pinging 127.0.0.1 works fine. Pinging localhost correctly resolves to 127.0.0.1 and works fine. I've tried restarting IIS and restarting the DNS Cache. I had this problem once before, and restarting the server fixed it, but I'd like to know what's going on in case this happens again in the future.

    Read the article

  • Do emails from ShoppyBag contain virus or malware ?

    - by Sysadmin Evstar
    I am curious -- when the ShoppyBag virus gets sent to you from a compromised "friend", inside the message is a secret one-pixel IMG unique to your email address --- and when your GMail message pulls up the message and the IMG is loaded from their server and displayed, their server knows you have read the message. At that moment, does it then grab your Gmail address book, the Flash Cookies, and all the Local Shared Objects it can find, i.e. at the instant you READ the message with the ShoppyBag virus, is it already too late? Do you have to Delete it to the Trash , then Delete Forever it , without reading it to be safe?

    Read the article

  • Apache: multiple domains handling

    - by cache
    So I use following schema to handle multiple sites on my apache: <VirtualHost 192.168.1.100:80> # get the server name from the Host: header UseCanonicalName Off VirtualDocumentRoot /var/www/%0/docs VirtualScriptAlias /var/www/%0/cgi-bin </VirtualHost> Therefore, if a client go to www.example.com, it will actually point to /var/www/www.example.com/doc/, which is good. However, what if the client go to example.com? It will point to /var/www/example.com/doc, which is not what we want. So my question is: is there any better schema for that? Or what should I do to fix the issue? Thanks!

    Read the article

  • Multiple Session using port 1081 in one box using SSH

    - by regmaster
    Hi Guru's, I am setting Linux Hopping Station to another different servers. My current config to connect to another servers is using different port to connect. e.g ssh -D 1080 -p 22 username@server1.com ssh -D 1081 -p 22 username@server2.com Now what I would like to have to share the same port from the same box. ssh -D 1080 -p 22 username@server1.com ssh -D 1080 -p 22 username@server2.com But when I share it, I will get below error: bind: Address already in use channel_setup_fwd_listener: cannot listen to port: 1080 Could not request local forwarding. How could I configure the same port? help. thank you. I want to share the same port because this is needed when configuring firewall in Citrix Firewall on other machine, not needed to many many ports and keep changing when changing connection. thank you.

    Read the article

  • Tweaking "Most visited sites" button in Firefox

    - by Mehper C. Palavuzlar
    It is sometimes practical to use "Most Visited" sites which stands on the left hand side of Firefox window. When I click on it, I can see maximum 10 URLs. At that point, I have 2 questions: Is it possible to increase the number of maximum most visited sites (say, 30)? Let's say example.com is one of the most visited domains. In the most visited list, there are other pages from this domain, like example.com/intro, example.com/info, example.com/help etc. So those sub-addresses are also in the list, but I just want to see max 1 (or maybe 2) pages from the same domain in the list. Is it possible to arrange the list this way?

    Read the article

  • postfix: deliver mail to specific emai addresses locally and send the rest to relayhost

    - by sme
    I want to achieve the following with postfix: Almost all outgoing emails shall be sent via a relayhost, configured in main.cf with relayhost = my.mailserver There's one specific email address ([email protected]) for which I want the email to be delivered locally. I set up a local account on the machine to match the email address and added "mydomain.com" to the mydestination property in main.cf. Then of course postfix tries to deliver every email to @mydomain.com locally and bounces them because the respective user doesn't exist. Question http://serverfault.com/questions/149453/hybrid-gmail-mx-postfix-for-local-accounts seems to be somewhat related (though more complicated).

    Read the article

  • Varnish with multiple hosts/subdomains

    - by jerhinesmith
    I'm new to Varnish, and I'm hoping it already does this "out of the box", but I'd like to clarify before I consider using it in production: Here's my setup: I have multiple sites running off of the same machine that vary by subdomain (i.e. user1.example.com, user2.example.com, etc.) Each "site" has a profile picture that has the same name (i.e. user1.example.com/profile.png, user2.example.com/profile.png) Will Varnish recognize these as separate resources and cache them accordingly? Or will I need to change something in the VCL to tell it include the full host url when looking up cache hits?

    Read the article

  • Google Sites (via Apps) setup questions

    - by Dave
    I thought that it would be a piece of cake to set up a Google site via Google Apps, but perhaps my previous (limited) experience with web development has given me unrealistic expectations. I have actually had a really tough time finding help with the exact question that I have, which is: How do I change the home page contents??? You see, I'm used to having hosting with someone like GoDaddy, where I can just ftp in and drop my HTML files in the www folder. From research I have found that this is simply not possible with any flavor of Google Sites. That's fine, I can live with it. So let's say I have www.mydomain.com. When I hit that URL, it redirects me to a very long URL (unfortunately) like https://sites.google.com/a/mydomain.com/sites/system/app/pages/meta/domainWelcome, which just says: Google Apps Welcome to mydomain.com If you are the domain administrator get started creating your home page with Google Sites Great! I want to do that. So I click on the "If you are the..." link and end up at a screen where I can choose a template, a name, and some visibility options. If I click on My Sites, there isn't a "default" site, i.e. the one that www.mydomain.com displays. I figured that maybe I just have to create a site first, so I went ahead and did that. My first test was to create a site that was publicly accessible. I thought that maybe if I did that, the Google would decide that this must be my home page since it's the only one. But it doesn't, and I still get the "Welcome to" page. Under "More Actions", I didn't see anything interesting except for "Manage site". I went in there and had a peek around, and didn't see anything about using this as the default home page. Am I looking for something that just doesn't exist? I can't believe there isn't a way to modify the "domain welcome to" page...

    Read the article

  • let CRON send emails through SMTP ( debian squeeze )

    - by supernova
    i would like to send emails whenever a cronjob has completed, i read that this is possible with exim4. in /etc/alias i added the line myuser: mygmail@gmail.com in /etc/exim4/update-exim4.conf.conf i set dc_smarthost='smtp.myserver.ip::25 and in /etc/exim4/passwd.client i set smtp.myserver.ip:[email protected]:mypassword my problem is that i can't see any login at my mailserver, and in the exim logs i saw a few lines with 2012-10-13 09:17:01 1TMvy1-0001fp-F2 ** mygmail@gmail.com R=nonlocal: Mailing to remote domains not supported 2012-10-13 09:17:01 1TMvy1-0001fr-JE <= <> R=1TMvy1-0001fp-F2 U=Debian-exim P=local S=17426 are there any additional config settings i have to set? edit : i solved the prev. error by running dpkg-reconfigure, but now i'm facing the following error <root@debian> R=dnslookup T=remote_smtp defer (-53): retry time not reached for any host

    Read the article

  • .htaccess help required for appache server

    - by mathew
    I searching for a redirection code for my url: what I want is when some one search in my site it should redirect example: if some one search google.com on mysite then in address line it should look like www.mydomain.com/google.com can be in $_POST method or $_GET how do I do that??

    Read the article

  • Facebook, Twitter, Yahoo doesn't work

    - by Toktik
    Some sites doesn't work normally, they are open, without css, images and javascript errors... Facebook stucks on static.ak.fbcdn.net Twitter stucks on a1.twimg.com Yahoo stucks on l.yimg.com On firefox I'm receiving Waiting for ...(any of those). I can access facebook only with SSL. Like https://facebook.com I ping them, only receive request timed out.

    Read the article

  • Using Redirect for an SSL CDN to have a Custom Domain

    - by bendytree
    We're looking to move our website/app assets to a CDN. The problem is, our CDN doesn't offer custom domain names with SSL. In other words, for SSL they offer https://1234.cdn.hostingcompany.com but not https://assets.mysite.com. So this seems like a huge problem since I don't want to re-publish my app with their domain hard coded. So I read somewhere about a method where you send people to https://assets.mysite.com then redirect to https://1234.cdn.hostingcompany.com. Is there merit to that solution or would that completely defeat the purpose of the CDN.

    Read the article

  • sites now not responding on port 80 [closed]

    - by JohnMerlino
    Possible Duplicate: unable to connect site to different port I was trying to resolve an issue with getting a site running on a different port: unable to connect site to different port But somehow it took out all my other sites. Now even the ones that were responding on port 80 are no longe responding, even though I did not touch the virtual hosts for them. I get this message now: Oops! Google Chrome could not connect to mysite.com However, ping responds: ping mysite.com PING mysite.com (64.135.12.134): 56 data bytes 64 bytes from 64.135.12.134: icmp_seq=0 ttl=49 time=20.839 ms 64 bytes from 64.135.12.134: icmp_seq=1 ttl=49 time=20.489 ms The result of telnet: $ telnet guarddoggps.com 80 Trying 64.135.12.134... telnet: connect to address 64.135.12.134: Connection refused telnet: Unable to connect to remote host

    Read the article

  • Should I create a second WAN Interface for a new SSLVPN setup on my Sonicwall 2400?

    - by TheSuperman
    Sonic Wall 2400 I'm setting up a new SSLVPN on our Sonicwall, very new to this. I currently have an exchange server as well, so mail.company.com is directed to our mail server. I'd like to setup a clean link for my low end users, remote.company.com to be for the SSLVPN, but I'm not sure how to do this within the Sonicwall? I setup SSLVPN on port 443. Only 1 WAN setup, on the X1 Interface. We have an A record setup with the same static IP used on the WAN Interface, and is configured for mail.company.com. Should I use a new static IP from our block of usable IP's to create the new Remote.Company.com? If so, I have no idea where to start on this on the Sonicwall? Any suggestions?

    Read the article

  • How to download images using wget from a txt file that contains links

    - by SwanC
    I can download images using wget if I download from a website. But I have several links and I have saved them in a text file. For example: wget -r -A.jpg -np www.fragrancenet.com There are so many pictures on this website. I have saved the links for the particular pictures I want: www.fragrancenet.com/images1 www.fragrancenet.com/images2 www.fragrancenet.com/images3 The links are saved in a text file named images.txt in my computer. How can I download the links in the images.txt text file using wget?

    Read the article

  • Access remote server with Nautilus through double SSH tunnel

    - by D W
    I'm trying to access my work computer from home. We are supposed to SSH into a server, say ssh.company.com and from there ssh into an xhost to to work, say xhost04.company.com. xhost04 is not directly accessible. How can I browse files on xhost04 from my home computer using nautilus (in ubuntu). To access the SSH host I would use: sftp://ssh.company.com/

    Read the article

  • How to run multiple Nginx instances on different port...

    - by Edvinas
    Hi, I would like to have several Nginx instances running on my server on separate ports (for example one user runs Nginx on port 2345, and another user on port 2346). So far, I have been successful in compiling and running the server on their designated ports. However, I am running into a weird issue: If i visit domain1.com:2345 or domain2.com:2346 I get the correct (200) response but if I visit domain1.com or domain2.com (without specifying the port) I get no response at all. Any help/clues in figuring out how to set this up would be highly appreciated.

    Read the article

  • Errors after Google Chrome downgrade

    - by user1933153
    I'm downloaded version 37.0 of Google Chrome but, few plugins and extensions didn't worked. I'm uninstalled that version and install 35.0 again, but now i have some errors. After every Chrome start i get these windows: http://i.imgur.com/rjB1SWJ.png , http://i.imgur.com/gfBa5Ju.png , http:(slash,slash)i(dot)imgur(dot)com/5opKTPE.png And every time i start chrome, i get these errors, it's really annoying, do you have idea how to fix that, or at least hide?

    Read the article

  • Setup SSH key per user for Git access

    - by ThatGuyJJ
    I'm setting up a site that will have multiple development instances running on the same server. Essentially, we'd have dev-a.whatever.com, dev-b.whatever.com, etc.. all running off a single server. I want to give each user some bit of SSH access in order to update and check in code from our Git repository and to manage files via SFTP. However, I want to restrict each user to their own site as well. So if you have access to dev-a.whatever.com, you don't also have access to dev-b.whatever.com and so on. The restriction is already in place if I login via FTP as a certain user, I can't navigate outside my own site -- but if I grant SSH access to that account I can immediately navigate to any file on the server in SFTP. Is RSSH part of the solution? And how can I assign the correct SSH pub key to the corresponding user? We're using BeanStalk for our Git repository management if that makes any impact.

    Read the article

< Previous Page | 337 338 339 340 341 342 343 344 345 346 347 348  | Next Page >