Search Results

Search found 29686 results on 1188 pages for 'ubuntu minimal'.

Page 403/1188 | < Previous Page | 399 400 401 402 403 404 405 406 407 408 409 410  | Next Page >

  • Good set of web hosting permissions?

    - by Jorge Israel Peña
    Hey guys, I just got a linode and I'm in the process of configuring it. It's running nginx with php-fpm and passenger. nginx was compiled and is running as user nginx. php-fpm (php with fastcgi process manager) is running as www-data (in group www-data). My sites are currently in /var/www, so for example /var/www/test.com I'm just wondering what the general 'flow' of things is. So for example, /var/www is owned by root, should I chown of /var/www/test.com to nginx or www-data? Or should I put nginx in the www-data group? How should site uploading work, I just transfer files to the /var/www/test.com directory as root (sudo) and then chown -R www-data:www-data .? Thanks. I'm capable of figuring things out on my own, I'm just wondering what the typical/general way of handling users/groups/permissions/site-files is on linux with a webserver.

    Read the article

  • disable .recycle feature for samba shares

    - by Crash893
    I had a pretty big scare when my company file server filled up. after tacking down the source I discovered that there is a .recycle folder that keeps ALL the files ever deleted (which is also hidden) Is there anyway to disable this feature? or periodically run a command that will delete all the junk?

    Read the article

  • Devices renamed with udev not working, original device node still there

    - by skerit
    I'm trying to rename certain video devices using udev. Here's a rule (broken down for readability): ACTION=="add", SUBSYSTEM=="usb", ATTR{busnum}=="1", ATTR{devpath}=="2", ATTR{idVendor}=="05e1", NAME="video97" It works. /dev/video97 is created But the original /dev/video1 node is STILL there. The even worse, when I try to view the /dev/video97 node in guvcview it doesn't work, while the original node DOES work.

    Read the article

  • Nagios only create warning for a http service

    - by MeinAccount
    I would like to also monitor non-crucial services with nagios like for example our GitLab-server or phpMyAdmin instance. Is there any way to just create warnings instead of circuital errors for some services? At the moment I'm using the following: define service { host_name localhost use generic-service service_description HTTP GitLab check_command check_www!git.example.com!'/users/sign_in' } define command { command_name check_www command_line /usr/lib/nagios/plugins/check_http -H '$ARG1$' -I '$HOSTADDRESS$' -e 'HTTP/1.1 200 OK' -u '$ARG2$' }

    Read the article

  • What actions does Ubuntu trigger when battery is low?

    - by blueyed
    When the battery is low, the screen gets dimmed after a few seconds already. This appears to be some special power-saving mode, and might be related to the time in org.gnome.settings-daemon.plugins.power.time-low (1200 seconds (20 minutes) the default). While this seems to get triggered by gnome-settings-daemon, I wonder what else Ubuntu does when this happens (e.g. via DBus listeners), or other event listeners that look for a "low battery" state. It seems like something in this regard causes Ubuntu / X / the system to behave more sluggish afterwards (when the laptop is on AC again), and I would like to look into what might be causing this. I could not find anything related via dconf-editor, e.g. in org.gnome.settings-daemon.plugins.power. It appears to get setup via idle_configure in plugins/power/gsd-power-manager.c, but it's probably something more related to something that listens on the DBus interface, which gets notified via e.g.: if (!g_dbus_connection_emit_signal (manager->priv->connection, NULL, GSD_POWER_DBUS_PATH, "org.freedesktop.DBus.Properties", "PropertiesChanged", props_changed, &error)) I could imagine that some "power saving" property gets set, but not unset when AC is available anymore and/or the battery is not low anymore. I have looked at the CPU governor setting (/sys/devices/system/cpu/cpu*/cpufreq/scaling_governor), but it was ondemand. I am using gnome-settings-daemon with awesomeWM on Ubuntu 14.04. gnome-settings-daemon=3.8.6.1-0ubuntu11.1 I've also compared gsd's plugins/power/gsd-power-manager.c with the one from Debian's gnome-settings-daemon-3.12.1, but could not find anything obvious that might have been fixed/changed in this regard. I have managed to trigger the gnome-power-manager's gnome-settings plugin (which dims the screen etc), by patching upower and use it after killing the system's upower daemon. (note that it's probably only energy that is being used by gpm to calculate it by itself). It does not make the system become sluggish.. OTOH I have not heard the speaker's beeping, which might come from the BIOS, which might be involved here, too - or other programs using the kernel's interface on /sys/class/power_supply/BAT0/. --- src/linux/up-device-supply.c.orig 2014-06-07 16:48:32.735920661 +0200 +++ src/linux/up-device-supply.c 2014-06-07 16:48:39.391920525 +0200 @@ -821,6 +821,9 @@ supply->priv->energy_old_first = 0; } + percentage = 3.1f; + time_to_empty = 3*60; + energy = 5; g_object_set (device, "energy", energy, "energy-full", energy_full,

    Read the article

  • Strange ssh key issue

    - by user55714
    Scenario 1. I am doing this from /home/deploy directory I am trying to set up ssh with github for capistrano deployment. this has been an absolute nightmare. when I do ssh [email protected] as the deploy account I get Permission denied (publickey). so may be the key is not being found, so If I do a ssh-add /home/deploy/.ssh/id_rsa Could not open a connection to your authentication agent. (i did verify that the ssh-agent was running) If I do exec ssh-agent bash and then repeat the ssh-add then the key does get added and I can ssh into github. Now I exit from the ssh connection to my server and ssh back in and I can't ssh into github anymore! Scenario 2 if I login to my remote server and then cd into my .ssh directory and ssh into github then it all works fine I guess there is a problem with locating the key and for some reason the agent isn't funcitoning correctly. Any ideas? Her is a pastie with more details..my .bashrc, permissions etc. http://pastie.org/pastes/1190557/

    Read the article

  • how do I set up a virtual host (it's not working, and I've done everything right)

    - by piratepartypumpkin
    My router redirects port 80 to port 8080. My router works fine and my domain name is routed properly. This is my virtual hosts file: NameVirtualHost *:80 <VirtualHost *:80> DocumentRoot /home/admins/lampstack-5.3.16-0/apps/wordpress ServerName example.com ServerAlias www.example.com </VirtualHost> I can access my website by entering "mywebsite.com:8080" but I cannot access it by entering "mywebsite.com" For further information, this is a part of my httpd.conf: Listen 8080 Servername localhost:8080 DocumentRoot "/home/admins/lampstack-5.3.16-0/apache2/htdocs <Directory /> Options FollowSymLinks AllowOverride None Order deny, allow deny from all </Directory> <Directory "/home/admins/lampstack-5.3.16-0/apache2/htdocs"> Options FollowSymLinks AllowOverride None Order allow, deny allow from all </Directory>

    Read the article

  • Ubuntu's New Web Office Integration

    <b>LinuxUK:</b> "Take for instance a low powered, possibly mobile/embedded system with limited processing power and memory. A cloud based service for these devices could allow resource intensive tasks to be offloaded to an online server somewhere, greatly improving the UX"

    Read the article

  • What are the system requirements for each flavor of Ubuntu Desktop?

    - by Braiam
    I'm thinking about installing Ubuntu Desktop, but I don't know what flavor is the better for my system. What are the minimum and recommended hardware requirements? What kind of CPU? How much memory? Should I have Hardware Acceleration? What flavor should I use? This is an attempt of a canonical answer. My answers have the "official minimal requirements", the recommended are a mix of official sources and opinion based (along with the answer it's told the source). You can comment or edit if you feel that the information is obsolete or incomplete. Is a good rule of thumb that any system capable to run Windows Vista, 7, 8 x86 OS X will almost always be a lot faster with any Ubuntu flavor even if they are lower-spec than described below.

    Read the article

  • Xorg: How can I map AltGr to the CapsLock Key (to toggle 3rd level symbols)

    - by basweber
    Hi, as many others I don't need Capslock. I want to reassign it to have the function of AltGr. I use Kubuntu 9.10 but I think there must be a solution which is distribution independent. I already tried to use setxkbmap or xmodmap. Using xmodmap at least I managed that the CapsLock key to behaves like the Delete key by following this description. But I could not achieve assigning the AltGr behavior to CapsLock.

    Read the article

  • pdns-recursor allocates resources to non-existing queries

    - by azzid
    I've got a lab-server running pdns-recursor. I set it up to experiment with rate limiting, so it has been resolving requests openly from the whole internet for weeks. My idea was that sooner or later it would get abused, giving me a real user case to experiment with. To keep track of the usage I set up nagios to monitor the number of concurrent-queries to the server. Today I got notice from nagios that my specified limit had been reached. I logged in to start trimming away the malicious questions I was expecting, however, when I started looking at it I couldn't see the expected traffic. What I found is that even though I have over 20 concurrent-queries registered by the server I see no requests in the logs. The following command describes the situation well: $ sudo rec_control get concurrent-queries; sudo rec_control top-remotes 22 Over last 0 queries: How can there be 22 concurrent-queries when the server has 0 queries registered? EDIT: Figured it out! To get top-remotes working I needed to set ################################# # remotes-ringbuffer-entries maximum number of packets to store statistics for # remotes-ringbuffer-entries=100000 It defaults to 0 storing no information to base top-remotes statistics on.

    Read the article

  • Shouldn't emesene themes change the emesene indicator icon?

    - by D Connors
    As it stands right now, I can't get the emesene icon in the indicator applet (or notification area, or whatever it's named) to change. No matter what icon theme I choose, even installing new themes, the icon is always the default one. Is that a standard behavior, or am I getting a bug? I'm running the latest version of emesene from their ppa (1.6-dev), on Lucid Lynx.

    Read the article

  • How do I set up my Ubuntu Server to let my Mac connect with FTP?

    - by Dolphin
    I have a server running Ubuntu Server 10.04, and I installed vsftpd on it. Now, I am able to connect to it with only one program which is for windows only (bleh) The program is WinSCP. I cannot connect to the server with any other application, such as the Ubuntu connect to server, Mac connect to server, FileZilla, Web Browser, or anything that is not WinSCP. I need to be able to connect with other applications because I mainly use a Mac to manage the server. I don't want to install wine on my Mac.

    Read the article

  • Add entire 300 GB filesystem to Git Annex repository?

    - by Ryan Lester
    By default, I get an error that I have too many open files from the process. If I lift the limit manually, I get an error that I'm out of memory. For whatever reason, it seems that Git Annex in its current state is not optimised for this sort of task (adding thousands of files to a repository at once). As a possible solution, my next thought was to do something like: cd / find . -type d | git annex add --$NONRECURSIVELY find . -type f | git annex add # Need to add parent directories of each file first or adding files fails The problem with this solution is that there doesn't seem from the documentation to be a way to non-recursively add a directory in Git Annex. Is there something I'm missing or a workaround for this? If my proposed solution is a dead end, are there other ways that people have solved this problem?

    Read the article

  • /usr/bin/env: ruby1.8: No such file or directory

    - by sandstrom
    Problem Trying to setup CI Joe (https://github.com/defunkt/cijoe) and I get the following error: /usr/bin/env: ruby1.8: No such file or directory What I'm doing CI Joe is run as www-data, and the command I'm running is RAILS_ENV=test bundle exec rake -s test When I log in as www-data (sudo su www-data) and execute the same command it works. So I'm trying to figure out what may be wrong. Debug Output $ which ruby1.8 /usr/bin/ruby1.8 This problem is similar to: http://stackoverflow.com/questions/6126165/usr-bin-env-ruby-1-8-no-such-file-or

    Read the article

  • how to install/compile CORSIKA/FLUKA for Ubuntu x32 12.04?

    - by Pantea Davoudifar
    I want to use some programs (CORSIKA/FLUKA) which are essentially designed for 32 bit systems. so I installed Ubuntu 12.04 32 bit on my system (Intel® Core™ i7-2700K CPU @ 3.50GHz × 8). Before this I had installed Ubuntu 9.10 (32-bit) on an older system and installed g77 from hardy repositories, compiled those programs without any problem. But this time when changing the repositories, g77 could not be installed even i removed all the things that i thought make this installation impossible, for example I need gcc-3.4 and removed all newer versions and tried to install them from hardy repositories. but the problem is that, whenever I have g77, corsika does not compile, and whenever I remove it, fluka does not compile, and also i received a error messages like this: crt1.o not found in /usr/bin/lb. In fact these .o files does not exist on my system user/bin/lb I have no directory lb there? I do not know how to link it? Or do i need to reinstall everything?

    Read the article

  • How do I limit concurrent sftp / port forwarding logins

    - by Kyoku
    I have ssh set up so my users can only access sftp and port forwarding, how can I limit the number of concurrent logins on a per user basis? In my sshd_config I have UsePAM set to yes and in /etc/security/limits.conf I have: username - maxlogins 1 I also tried: username hard maxlogins 1 Neither of these works and the users can still log in multiple times.

    Read the article

  • Echo 404 directly from nginx to improve performance

    - by user64204
    I am in charge of production servers serving static content for a website. Those servers are constantly being crawled by bots looking for potential exploits (which isn't that much of a problem security-wise because no application can be reached behind the web server) but generates thousands of 404 per day, sometimes per hour. I am looking into ways of blocking those requests but it's tricky (you want to make sure you don't block legitimate traffic and these bots are becoming more and more clever at looking like they're legit) and is going to take me a while to find an acceptable solution. In the meantime I would like to reduce the performance impact of serving those 404 pages. Indeed we're using nginx which by default is configured to serve it's 404 page from the disk (This can be changed using the error_page directive but in the end the 404 will either have to be served from disk or from another external source (e.g. upstream application which would be worst)) which isn't ideal. I ran a test with ab on my local machine with a basic configuration: in one case I echo a message directly from nginx so the disk isn't touched at all, in the other case I hit a missing page and nginx serves its 404 from disk. server { # [...] the default nginx stuff location / { } location /this_page_exists { echo "this page was found"; } } Here are the test results (my laptop has Intel(R) Core(TM) i7-2670QM + SSD in case you're wondering why they are so high): $ ab -n 500000 -c 1000 http://localhost/this_page_exists Requests per second: 25609.16 [#/sec] (mean) $ ab -n 500000 -c 1000 http://localhost/this_page_doesnt_exists Requests per second: 22905.72 [#/sec] (mean) As you can see, returning a value with echo is 11% ((25609-22905)÷22905×100) faster than serving the 404 page from disk. Accordingly I would like to echo a simple 404 Page not Found string from nginx. I tried many things so far but they all failed, essentially the idea was this: location / { try_files $uri @not_found; } location @not_found { echo "404 - Page not found"; } The problem is that as soon as the echo directive is used, the http response code is set to 200. I tried changing that by doing error_page 200 = 400 but that breaks the configuration. How can I serve a 404 page directly from nginx? (without hacking the source which may be might next step)

    Read the article

  • Is it safe to have NVidia graphics always on on a Linux laptop, or do I risk overheating?

    - by codeape
    I'm getting a Lenovo T520 with two graphics cards: Integrated Intel HD 3000 Discrete NVidia NVS 4200M In BIOS, I can adjust which card(s) to use: Integrated only Discrete only Both (NVidia optimus) Since optimus is not well supported under Linux, I wonder if it is OK to set up the system to use the NVidia card all the time. I have read somewhere that a laptop risks overheating if using a discrete graphics card all the time. Is this true? Does someone have any experience to share?

    Read the article

  • How to set up a WACOM Cintiq 24 HD in Ubuntu 11.10/12.04?

    - by isphording90
    I'm new to Ubuntu and the Linux world (switched from mac to linux) and i have a problem: I'm studyin industrial-design and a lot of my work depends on my wacom cintiq 24 HD. My problem is that ubuntu doesn't find my cintiq. I googled for wacom drivers and found the linux wacom project. The problem for me is that i hardly understand anything of what i have to do... Is there anyone who can tell me how i can set up my 24HD or is there an easy way for people who are new to linux like me. I really really like linux and want to stay with it but that isn't possible for me without my cintiq. I would be very greatefull for any help!

    Read the article

  • How can I access old KMail emails after upgrading to Ubuntu 11.10?

    - by Gary Kleppe
    I initially installed Ubuntu 11.04 and used KMail for my email. All well and good. Then I upgraded to Ubuntu 11.10. Presumably an upgrade of KMail took place as part of this. Now KMail won't even run; when I try, it tells me "Failed to fetch the resource collection" and crashes. I don't mind switching to another email client, but I'd very much like to be able to recover all of the emails I have stored in KMail. Any suggestions on how to do this? Thanks for any help anyone can provide.

    Read the article

  • Best linux USB DVB-T hardware

    - by Roalt
    What is the best supported USB-based DVB-T (free-to-air) tuner for Linux? I've seen a similar question, but this one is PCI-based, not USB: http://superuser.com/questions/104663 I've tried the Conceptronic CTVDIGRCU v3.0 but it does not pick up any channels (although it does work on windows).

    Read the article

  • Grub error 18, gparted not showing anything

    - by Montecristo
    Some week ago I started having some problems with my pc, sometimes it just freezed not allowing me to do anything. I had to turn it off and on and sometimes do it a couple of time even at startup. Now it does not start at all, grub is giving me error 18. I have found that a solution is to create a bootable partition in the first sector of the disk. gparted does not recognize any partition, the window in which there would be my partitions is empty. sudo fdisk -l does not output anything. If I type sudo mount /dev/sda and then tab tab to autocomplete these are the devices coming out: sda sda1 sda2 sda5. If I launch sudo mount -t ext3 /dev/sda1 disk I get the following error: mount: wrong fs type, bad option, bad superblock on /dev/sda1, missing codepage or helper program, or other error In some cases useful info is found in syslog - try dmesg | tail or so dmesg outputs [ 1831.974847] EXT3-fs: unable to read superblock Do you know how to solve this issue? I'm not completely sure this is a software problem, should I try with a new hard disk?

    Read the article

  • Can't re-open "Group My Tabs" page after killing Firefox

    - by isomorphismes
    Due to memory problems I closed Firefox whilst the Group My Tabs (Ctrl+e) feature was displaying the multiple tab groups. After 20 minutes the process still hadn't finished so killall firefoxed it. When I restart Firefox, there is a Javascript process called tabview.js that hangs unless I click Stop at the warning screen. If I do hit Stop then I can no longer open the tabview, so I can't get to any of the tabs except in the subgroup I was last looking at. Any suggestions?

    Read the article

< Previous Page | 399 400 401 402 403 404 405 406 407 408 409 410  | Next Page >