Search Results

Search found 54052 results on 2163 pages for 'run configuration'.

Page 111/2163 | < Previous Page | 107 108 109 110 111 112 113 114 115 116 117 118  | Next Page >

  • Get into the router setting without knowing the password [on hold]

    - by Aron
    So my father has changed the router passwords, and programed the router to turn my wifi connection off automaticly at some times, for example at night. I wanted to know if there is any way I can trick this feature? Can i change my ip in some way, the router won't recognize my ip, so my new ip would have internet connection? Is it any way i can get into the router settings and turn that feature off without knowing the router password? Would be happy if I could get replies as soon as possible!

    Read the article

  • Make nginx config like apache2 virtualhosts

    - by user2104070
    I have web server with apache2 with many subdomains on it like, domain.com, abc.domain.com, def.domain.com etc. etc. Now I got a new nginx server and want to set it up like apache2, so to test I created configs (2 files in /etc/nginx/sites-available/ and link to them from sites-enabled/) as shown, domain.config: server { listen 80 default_server; listen [::]:80 default_server ipv6only=on; root /srv/www/; index index.html index.htm; # Make site accessible from http://localhost/ server_name domain.com; location / { # First attempt to serve request as file, then # as directory, then fall back to displaying a 404. try_files $uri $uri/ =404; } } abc-domain config: server { listen 80; listen [::]:80; root /srv/www/tmp1/; index index.html index.htm; # Make site accessible from http://localhost/ server_name abc.domain.com; location / { # First attempt to serve request as file, then # as directory, then fall back to displaying a 404. try_files $uri $uri/ =404; } } but when I access with domain.com I am getting index.html from /var/www/tmp1 only. Is there something I'm doing wrong in the nginx config?

    Read the article

  • How to make spamassasin reject mail based on score?

    - by Kaizoku
    I am running linux with postfix, dovcot, postgrey and spamassasin/spamd. This is my main.cf smtpd_recipient_restrictions = permit_sasl_authenticated, permit_mynetworks, reject_unauth_destination, reject_invalid_hostname, reject_unauth_pipelining, reject_non_fqdn_sender, reject_unknown_sender_domain, reject_non_fqdn_recipient, reject_unknown_recipient_domain, check_policy_service inet:127.0.0.1:10030 I got spamassasin to add spam header to other emails that postfix didn't catch, but how to make it reject?

    Read the article

  • How does one enable --write-mostly with Linux RAID?

    - by user76871
    Unfortunately the mdadm and mdadm.conf man pages are not quite up to par. I would like to enable the --write-mostly flag for my RAID, but neither the man pages nor the internet will tell me how. I am not aware of any place to put default arguments for mdadm, nor aware of when it would be launched and by what. It seems the logical place to add this information is mdadm.conf, but the flag is unmentioned in man mdadm.conf. Where and how can I enable --write-mostly? Thank you.

    Read the article

  • Changed folder icon in WinXP and now clicking a folder opens in a new window.

    - by widgisoft
    I got bored and decided to change my folder icon from the boring yellow default. Somewhere in my experiments, I discovered that anytime I open a folder it opens in a "search" window. This happened because when I go to the "(file folder)" type, the only option in there is "find" that is not supposed to be the default; upon saving this form XP tries to do me a favour and sets this as the new default. Now whenever I try to click a folder it opens the search window. How do I fix this?

    Read the article

  • Server OS: put it on a separate drive? Yes, no, or depends on the situation?

    - by captainentropy
    Hi, I would like opinions, or facts, both preferably, on whether it's ok to install a server's OS on the RAID array or not. I would predict installation on separate drives is the best but I'm interested in the performance. The server in question will have 8 cores (2.4GHz ea.), 24GB RAM, and ~16TB of usable space of server-class drives in RAID10. There is also a subsytem of an ~equivalent size for backup. I will be running CPU/memory intesive applications on this server in addition to it being file storage for my work (research lab). IF I install the OS (haven't decided which one, probably Ubuntu or Fedora or some other good linux distro) on separate drives will there be any performance problems if they aren't configured in RAID10? IF it is better to have the OS on separate drives should I go for 150GB velociraptors in RAID1 or smallish SSD drives in RAID1? Money is unfortunately a factor as I think I'm close to maxing my budget as it is. Thanks!

    Read the article

  • Change the starting day in Vista's Calendar Gadget

    - by Michael
    I have a problem with the calendar gadget in Vista. The starting day is Sunday, but I would like to change it to Monday. I have checked some forums which points me in the direction that: Change the Regional Settings in ControlPanel to your local region Change the register key HKEY_CURRENT_USER/Control Panel/International/iFirstDayOfWeek to 0 Both of these are already correct, so it didn't help me.. Does anyone else have a suggestions?

    Read the article

  • BIND9 Forwarding by view

    - by Triztian
    Hi I think this is a simple issue, I'd like to forward only to certain IPs in the LAN network, for example I have 2 acl lists: acl "office1" { 192.168.1.15; // With internet access }; acl "production" { 192.168.1.101; // No internet access }; I know that there probably should be more efficient ways to restrict internet access, but at the moment this is what I'd like to try.Here's what I've tried in named.conf.local // Inlcude my acl definitions include "/etc/bind/acls.conf"; view "no-internet" { match-clients { production; }; include "/etc/bind/named.conf.default-zones"; zone "localdomain.com" { type master; file "/etc/bind/db.localdomain.com"; }; zone "1.168.192.in-addr.arpa" { type master; file "/etc/bind/db.192.168.1"; }; } view "internet" { match-clients { office1; }; include "/etc/bind/named.conf.default-zones"; forwarders { 201.56.59.14; // Made Up 201.56.59.15; // Made Up }; zone "localdomain.com" { type master; file "/etc/bind/db.localdomain.com"; }; zone "1.168.192.in-addr.arpa" { type master; file "/etc/bind/db.192.168.1"; }; }; As you can see I want a localdomain.com defined for every computer in my network and forward internet access to the computers in the office but not to the ones on the production floor. I've modified my conf file, however the IP in the "no-internet" acl is able to resolve the domains, even though I've rebooted the computer, flushed the DNS using ipconfig /flushdns and set my DNS Server as the only one, why is this still happening? Thanks in advance.

    Read the article

  • Configuring a monitor's constract/brightess/colours/etc

    - by DMA57361
    I have recently bought myself a new monitor, now I'd had my previous screen well over 5 years now so I'm looking at this one thinking the picture doesn't quite look right (not bad, just different). Rather than just wait until I'm used to the new picture I'd rather get it fine tuned, then get used to it - so I can reap the maximum benefit. So, can you please suggest reliable ways of configuring an LCD monitor's brightness/contrast/colour/etc to provide the optimum possibly quality image?

    Read the article

  • Apache routing vhosts to /var/www

    - by FHannes
    One user at my site has reported that he reaches the content at /var/www when browsing to any of the vhosts at my server. As far as I’m aware, my Apache server does not contain a document root that references this folder. On top of that, this user seems to be the only one experiencing the issue. According to his ISP, the issue isn’t caused by them, yet, on his mobile connection, he can access the site. When browsing to my server’s IP, he also receives the correct content from the default vhost. What could be the possible causes of this issue and how can I get it to stop? I’ve explored pretty much every option I could think of.

    Read the article

  • How to prevent MediaWiki removing non default signatures?

    - by WikiSpeedia At Area51
    We recently upgraded MediaWiki from 1.13.2 to 1.15.4. One of the side effects is that people's signatures are automatically deleted. That is, a signature added under MY PREFERENCES gets changed to match the Real Name field after a couple of minutes. We set $wgCleanSignatures = false in LocalSettings.php but this does not change the behaviour. Does anyone know what is going on and how to prevent this?

    Read the article

  • Strategy to isolate multiple nginx ssl apps with single domain via suburi's?

    - by icpu
    Warning: so far I have only learnt how to use nginx to serve apps with their own domain and server block. But I think its time to dive a little deeper. To mitigate the need for multiple SSL certificates or expensive wildcard certificates I would like to serve multiple apps (e.g. rails apps, php apps, node.js apps) from one nginx server_name. e.g. rooturl/railsapp rooturl/nodejsapp rooturl/phpshop rooturl/phpblog I am unsure on ideal strategy. Some examples I have seen and or thought about: Multiple location rules, this seems to cause conflicts between the individual app config requirements, e.g. differing rewrite and access requirements Isolated apps by backend internal port, is this possible? Each port routing to its own config? So config is isolated and can be bespoke to app requirements. Reverse proxy, I am little ignorant of how this works, is this what I need to research? is this actually 2 above? Help online seems to always proxy to another server e.g apache What is an effective way to isolate config requirements for apps served from a single domain via sub uris?

    Read the article

  • What is the `/etc/hostname` used/required for?

    - by static
    I found in the /etc/hostname my IP-address, than I deleted it and each time I use sudo - I get a message and a system email sudo: unable to resolve host (none) or if in the /etc/hostname is saved myhostname than sudo: unable to resolve host (myhostname). I know it is used to set the system's hostname via /etc/init.d/hostname.sh while booting process, but what is this setting required for (programs, services, daemons ...)? What if i set to localhost (so it doesn't happen any sudo: unable to resolve host (none) anymore, but is it still ok?)? UPD1: I found some information here: http://jblevins.org/log/hostname, but it is all about how to use it and not about - why it is required.

    Read the article

  • different user group can not upload file in the server

    - by Dallal
    I have a CentOS server running in Thailand, and I'm in Canada. The guy at the computer center who set up the server for me doesn't really understand much about linux and left me off an issue to solve myself. I just moved from Mac Server to Linux server, and the first thing I'm facing a problem now is `file name` has failed to upload due to an error The uploaded file could not be moved to `location name` So what happen is that I knew from my experiences of these problem is all about permissions. So I go ahead and checked on my whole folder and found that everything in the folder permission is like myusername mygroupname then I checked the httpd file in the server and it is default to apache apache. My question is that how can I make my user to be in the same group with apache group so that I don't have to have any problem about uploading, changing data in my file....? But without having to affect other user in the same server. I'm holding Administrator account, but not root account, but I can change stuff on the server root no problem. When I was with godaddy.com there never been any problem about the permission and I wish I know how they configure that :(

    Read the article

  • Nginx config with try_files and rewrite : precedence?

    - by Penegal
    Good morning, everybody. Firstly, this question may have been already asked, but I searched ServerFault during about 15 minutes without finding it, so, if it was already asked, please accept my apologies. I'm trying to rationalize my Nginx server config, but I have a rather dumb question that I couldn't solve, even with extensive Web search, except if I totally f*cked my search. Here is the question : is try_files parsed before or after rewrite ? Asked differently, Do I have to put try_files after all rewrite directives, or is Nginx config parser smart enough to evaluate try_files after all relevant rewrite directives ? The link with the config rationalization is that the answer to this question will change the organisation of the config, ie if config file order of try_files and rewrite changes the config behaviour, it will force me to disperse my includes, some of them containing try_files and other ones containing rewrite, because I also have rewrite directly in nginx.conf. Hoping you can help me, Regards.

    Read the article

  • 20 1TB drives vs. 10 2TB drives in RAID5/6 server

    - by Hunter
    Hi everyone, I will be setting up a server at work and I need some advice on some details. The setup will be one blade-type server (8-core, 16GB RAM) with two subsystems - one for the main storage the other to back it up. I'm shooting for a 20TB array (I know it'll be less after formatting and parity drives). So is there any advantage one way or the other with either 20 1TB drives or 10 2TB drives? I'm not sure right now how many controllers I should have either (in the quote I have is a dual-port controller). I would think two controllers for a server of this size would be a better choice than the dual-port controller (but I really don't know). And would an array of this size have any performance issues in RAID 5 or 6 (I know RAID 5 or 6 are "slower" because of all the parity calculations). Also, these will be either WD RE3 (1TB) or the RE4 (2TB). Oh, also, for the backup array would it be ok to use the WD 2TB green drives (also in RAID5 or 6)?

    Read the article

  • problem with accessing a php page

    - by EquinoX
    So I have a info.php page which is located on the folder /var/www/nginx-default, however when I go to my ip address/info.php, it always redirects me to this site: http://www.iana.org/domains/example/ is this because I have a virtual host that I called example? Here is my config for the example website: server { listen 80; server_name www.example.com; rewrite ^/(.*) http://example.com/$1 permanent; } server { listen 80; server_name example.com; access_log /var/www/example.com/logs/access.log; error_log /var/www/example.com/logs/error.log; location / { root /var/www/example.com/public/; index index.html; } } The way I access this site is by changing my /var/hosts in my macbook so that example.com is mapped to my server IP address... however now when I do xxx.xxx.xxx.xxx/info.php.. it redirects me to that site I posted above

    Read the article

  • Upgrade PHP v5.3.3 to v5.3.4

    - by Ty01
    I currently have PHP v5.3.3 installed and configured. Everything is working perfectly, but I would like to keep PHP up to date and upgrade to v5.3.4. Can someone please describe the usual upgrade process for PHP when compiling manually? For example: is it just as easy as downloading the newest source, uncompressing it, compiling it using the same (or comparable) options that were used on the last version? Is there anything that has to be removed or changed from the previous version installed? I'm clueless. Please help!

    Read the article

  • Internet connection fails in Ubuntu on VirtualBox when virtual machine is created from "Import appli

    - by Sanoj
    I have installed Ubuntu Server 9.10 in a virtual machine in VirtualBox, then I made a cope/clone and exported it with "Export appliance" so I can create many cloned virtual machines. But when I try to import an appliance, everthing seams to be fine with the Ubuntu except that it can't connect to Internet and doesn't get an IP-address. The machine is used in Bridged mode. And it doesn't help to change to NAT-mode either. The machine that I cloned seams to work fine, and get an IP address. How to fix this? Where am I doing wrong?

    Read the article

  • Emacs doesn't use ~/.ssh/config when accessing files on a remote machine

    - by Yotam
    I have a fresh install of arch Linux. I've installed Emacs from the rpos, and my home directory is mounted from a separate partition. I have old settings I've used on my ~/.ssh/config along with authentication keys I've regularly used before. Now, when I try to connect to a remote machine using Emacs, Emacs asks for my password and uses the wrong username. Clearly, Emacs doesn't access my config file. When I try to ssh or scp directly to the machine, things work fine. What do I need to update?

    Read the article

< Previous Page | 107 108 109 110 111 112 113 114 115 116 117 118  | Next Page >