Search Results

Search found 16797 results on 672 pages for 'directory traversal'.

Page 493/672 | < Previous Page | 489 490 491 492 493 494 495 496 497 498 499 500  | Next Page >

  • How can I set clean urls (enable rewrite) if I don't have a domain ?

    - by Patrick
    In order to enable clean urls in Drupal, I add the lines below to the lighttpd configuration file. However I'm now working on a local server and I don't have a domain available. So I need to work with this address http://local.ip/Sites/mywebsite I've tried to replace ["host"] with ["socket"] and replace the domain with ip and subfolders (see address above), but unsuccessfully. How can I set the configuration file to set clean urls even if I don't have a domain ? thanks $HTTP["host"] =~ "(^|\.)mywebsite\.com" { server.document-root = "/var/www/sites/mywebsite" server.errorlog = "/var/log/lighttpd/mywebsite/error.log" server.name = "mywebsite.com" accesslog.filename = "/var/log/lighttpd/mywebsite/access.log" include_shell "./drupal-lua-conf.sh mywebsite.com" url.access-deny += ( "~", ".inc", ".engine", ".install", ".info", ".module", ".sh", "sql", ".theme", ".tpl.php", ".xtmpl", "Entries", "Repository", "Root" ) # "Fix" for Drupal SA-2006-006, requires lighttpd 1.4.13 or above # Only serve .php files of the drupal base directory $HTTP["url"] =~ "^/.*/.*\.php$" { fastcgi.server = () url.access-deny = ("") } magnet.attract-physical-path-to = ("/etc/lighttpd/drupal-lua-scripts/p-.lua") }

    Read the article

  • Hiding samba share from browse list for unauthorised users

    - by karlbright
    Hey Guys, I have been trying to setup my samba shares correctly. The setup i was looking for was having a couple of shares available publicly, guest accounts are ok and can browse these shares all ok. I have this setup correctly. The problem comes when setting up a share that only certain users can view, although i have setup a share that will only allow certain users to access. I havent been able to hide this share from guests. I had a look into the browseable = yes option but this hides it from everyone, including the users that have logged in. Any idea on how to tackle this? The setup i have for this private share is follows: [private] comment = private share for certain users path = /media/drive/private create mask = 0777 directory mask = 0777 writable = yes public = no users = admin

    Read the article

  • Why can't my Perl script in ~/bin find relative file paths?

    - by sid_com
    #!/usr/bin/env perl use warnings; use strict; use XML::LibXML; my $parser = XML::LibXML->new; my $file = './example.xml'; my $doc = $parser->parse_file( $file ); print ref( $doc ), "\n"; When I move this script and the example.xml-file to /home/me/ then the script works. When I move the script and the example.xml-file to /home/me/bin/ then the script doesn't find the example.xml-file. Is this some special-feature of the bin-directory?

    Read the article

  • Saving backup files automatically in (g)Vim after saving a file.

    - by Somebody still uses you MS-DOS
    I had a problem with my gVim. I lost some important modifications after I plugged on my machine after a hibernating process. To avoid this kind of problem, I would like to know if it's possible to add something in my .vimrc (or a plugin) that automatically backups all saving made to my files. Disk space is not an issue, I can delete these files after. I'm already using set backup set backupdir=~/.backup/vim set directory=~/.swap/vim This creates a myfile.extension~ in my .backup/vim. ...but I would like this configuration to add ~ to first save, ~0 to second, ~1 to third, ~2 to fourth, and so on - something that keeps copies from all modifications I made to a file. Is this possible? Do you know if there's a plugin for this?

    Read the article

  • Access Domain over Wireless

    - by William
    I have an active directory server and all hard wired clients connect perfectly fine. However any wireless clients do not recognize that they are connected to the domain, despite the dns server being appropriately set to the domain controller for both wired and wireless connections. Also - both wired and wireless clients are on the same subnet. I am using a Verizon food actiontec router. What am I missing? For instance, this is the same computer connected first via an Ethernet cable and second via the wireless.

    Read the article

  • Uninstall file from Apache Tomcat Web Server

    - by sttaq
    I have a website deployed on an apache web server. I am using a windows installer to deploy and un-deploy this website from the server. During the uninstall I am just removing the .war from the webapp directory of the tomcat. Now when I do this the tomcat automatically removes all the folders created when it unpacked the war file. But it takes some time to remove those folders. I want to know if there is a way to force tomcat to remove the folder immediately by communicating to it with some external tool eg a script, xml-rpc etc?

    Read the article

  • sticky bit on NFS file system

    - by Kris_R
    I have a system where to the main server (homes, nfs, ntp, queue...) can log-in only root – all the other users use front-end host with NFS-mounted home directories (RW) and all other software directories (read-only). My problem is, that time to time, if root or normal user with sudo makes some administrative works on front-end some homes of normal users getting sticky bits (drwsr-sr-x). If it happens usually the user can't log-in (as long as permission for his home are not changed to drwxr-xr-x). The last time I saw it after compiling some new software (normal user configure;make) and installation from the same directory as root (su and make install or direct as normal user sudo make install). Can somebody explain me why it happens and what should I do to get rid of this problem? p.s. I'm using CentOS 5.7

    Read the article

  • Is it possible to analyze the size of a SubVersion repository?

    - by BrianH
    Is it possible to know how much disk space each project in a SubVersion repository is using? I can check out a working copy of each project and look at the size each project takes up, but I don't think that encompasses the total size of the project (all revisions). I can look under the "db" directory of the repository, but none of the files in there make sense - I don't think it is possible to use them to figure out how much space each project occupies. I tried the svn ls --verbose command, but the size that it gives me is just the size of the actual files in the head revision, I don't think it includes all revisions. Maybe this isn't possible, but I thought I would ask. Thanks in advance!

    Read the article

  • How to perform a nested mount when using chroot?

    - by user55542
    Note that this question is prompted by the circumstances detailed by me (as Xl1NntniNH7F) in http://www.linuxquestions.org/questions/linux-desktop-74/boot-failure-upon-updating-e2fsprogs-in-ubuntu-10-10-a-947328/. Thus if you could address the underlying cause of the boot failure, I would very much appreciate it. I'm trying to replicate the environment in my ubuntu installation (where the home folder is on a separate partition) in order to run make uninstall. I'm using a live cd. How to mount a dir in one partition (sda2, mounted in ubuntu as the home folder) into a directory on another mounted partition (sda3)? I did chroot /mnt/sda2 but I don't know how to mount sda3 to /home, and my various attempts didn't work. As I am unfamiliar with chroot, my approach could be wrong, so it would be great if you could suggest what I should do, given my circumstances.

    Read the article

  • NGINX Document Location

    - by GLaDOS
    I want to be able to access a given url, example.com/str. The problem is that the php file that I want to connect to is in a directory of /str/public/. In my nginx logs, I see that it is trying to connect to /str/public/str/index.php. Is there any way to remove that last 'str' in the document request? Below is my location directive in sites-available/default: location /str { root /usr/share/nginx/html/str/public/; index index.php index.html index.htm; location ~ ^/str/(.+\.php)$ { try_files $uri = 404; root /usr/share/nginx/html/str/public/; fastcgi_pass unix:/var/run/php5-fpm.sock; fastcgi_index index.php; fastcgi_param SCRIPT_FILENAME $document_root$fastcgi_script_name; include fastcgi_params; } } Thank you all so much in advance.

    Read the article

  • Unix / linux permissions setup for shared hosting with Apache

    - by weiyin
    I'm in the process of setting up a server from a clean CentOS 5 install. What is the best permission structure (users, groups, unix permissions) for running a single instance of apache for multiple users? Ideally, it should satisfy these requirements: Each user's websites are stored in a subdirectory of their home directory. Users can edit files and permissions. Apache can read the websites of all users. No user can read the website files of other users. Bonus question: how to add PHP and/or Perl and/or Ruby to Apache without allowing any users to access any other user's files?

    Read the article

  • Horrible time with WAMP, Joomla and Windows 7 (Permissions)

    - by jax
    I have WAMP and Joomla installed on Windows 7. I have made a virtual host outside of the www directory that houses a test joomla site. I have also made a joomla component that I want to install and test on this virtual host. I am having constant problems with permissions. I have given all users full access to the folder containing the files under the security Tab in Win7. It seems to me that when (Apache/PHP) creates a file it put restrictions on the permissions which does not allow it to delete or edit the file anymore so I have to manually go and delete the files when I want to uninstall my component. Also, I cannot install the .zip - instead I have to unzip manually to the tmp folder because when I upload the zip in the Joomla Installer I get the same permissions problems as above. How can I get rid of these issues once an for all?

    Read the article

  • (How) does deleting open files on Linux and a FAT file system work?

    - by lxgr
    It's clear to me how deleting open files works on filesystems that use inodes - unlink() just decreases the link count to zero, and when the last file handle to the file is closed, the inode will be removed. But how does it work when using a file system that doesn't use inodes, like FAT32, with Linux? Some experiments suggest that deleting open files is still possible (unlike on Windows, where the unlink call wouldn't succeed), but what happens when the file system is uncleanly unmounted? How does Linux mark the files as unlinked, when the file system itself doesn't support such an operation? Is the directory entry just deleted, but retained in memory (that would guarantee deletion after unmounting in any case, but would leave the file system in an inconsistent state), or will the deletion only be marked in memory, and written at the time the last file handle is closed, avoiding possible corruption, but restoring the deleted files after an unclean unmount?

    Read the article

  • merge pieces of file together

    - by veilig
    I have a directory tree of files I need to loop through and reformat the lines. can I do this with sed with some combination of find? for each file, I'll need to adjust the text in the file from ... * @category Foo * @package Bar ... into just this (Remove the @category line, but prepend the value to the package annotation value) ... * @package Foo/Bar ... what is the most efficient way of doing this? its a few hundred files that will be modified.

    Read the article

  • Easiest method of creating multiple Word documents with incremental number?

    - by DakotaDusty
    I need to create about 80 Word documents that are identical except for a numerical identifier in the document. The identifier is the same as the document filename, eg. the document named "SN-100.doc" must have the string "SN-100" inside the document text. Each unique document will reside in a folder location given by its unique numerical identifier.For example, the file directory hierarchy will look something like this: /SN001/SN-1.doc /SN002/SN-2.doc . . . /SN080/SN-80.doc What is the easiest and fastest method of doing this?

    Read the article

  • htaccess with wildcard SSL

    - by Ericko
    We have a Wildcard SSL Certificate that is supposed to work on any subdomain of a given domain. So in this server we have this file structure: /home/DOMAIN/public_html/subdomainx /home/DOMAIN/public_html/subdomainy etc... Now, the Certificate is installed, but when you visit any subdomain over https (example: hxxps://subdomainx.domain.com ) it points to /home/DOMAIN/public_html/index.php We need that when you visit a subdomain via https hxxps://subdomainx.domain.com That it points to the the same directory that it's http equivalent: /home/DOMAIN/public_html/subdomainx Our provider tells us that this is not possible, that the current behaviour is correct, and that to achieve this we need to do it with htaccess. I've tried a few things, incluiding this solution, that seems to be what I need: http://stackoverflow.com/questions/5365612/advice-on-configuring-htaccess-file-to-redirect-http-subdomain-to-https-equival But can't get it to work. Any tips? Thanks. Added: The server is Apache.

    Read the article

  • How to rsync just the current folder?

    - by Michael
    I want to sync some files from a remote server to my local computer. How can I make rsync to just copy the files with a certain file extension in the directory but no subdirectories? I assumed this to be an easy task, but embarassingly I'm not getting it for nearly 2 hours. So could someone give me an example? I did various experiments with something like the following command: rsync -a --include=what? --exclude=what? -e ssh [email protected]:/test /test

    Read the article

  • rsync stuck with the --checksum option

    - by billc.cn
    I use back-in-time to backup my Linux installation. It serves as an advanced wrapper for the rsync command. Today I tried to add /var/log to the list of folders to be backed up and it caused some serious performance problems. The job seems to stuck on a particular file and the CPU usage of the rsync parent process reaches 100%. I then used lsof to see which file caused the problem and it seems to be the /var/log directory. I did some googling and some experiments with the different rsync options and found --checksum to be the offender. Without the parameter, an incremental backup finishes properly in minutes. With it, the process will stuck when rsync tries to sync a constantly changing log file. This kind of make sense, but it still seems to be a bug to me. Am I using the option correctly? Is there a workaround for this?

    Read the article

  • Hyper-V share a folder between host and instance

    - by Fly_Trap
    I have a hyper-v server and several VM's (Virtual Machines). All the VM's are connected to an external network. I have tried to share a folder on the host and connect via the VM, I can do this but I'm prompted for a user name and password (as you would expect). I do not want to enable the "Everyone" group permissions as the physical host server is on a network of other servers. I have created a new virtual internal network in Hyper-V and given it's adapter a static ip of 33.0.0.100. I have added the virtual adapter to one of the VM's and set to IP to 33.0.0.2 (as advised here). Again this seems to work but I'm still prompted for a user name and password. Am I on the right lines here? I just want to share a directory from the host to the vm's without exposing the share to other servers on the network.

    Read the article

  • How do I pull a backup from a Linux server to my Windows PC using rsync?

    - by Nogwater
    I'm currently using sftp to download nightly backups (.tar.gz) from my web host to my desktop computer. I think I'd like to switch to rsync to minimize the bandwidth (and time). I have cygwin installed on my PC, but don't use it for much. I have shell access to my web host via ssh (PuTTY). Let's say my source directory is myserver.com:/home/username/backups/, I want to grab all of the .tar.gz files from there, and I want to save them to C:\Backups\ locally.

    Read the article

  • Search desktop files using a list of keywords stored in a text file

    - by Tod1d
    I have a list of 1285 keywords (database object names) that I have compiled into a TXT file; one keyword per line. I would like to search a directory of files (most have a .aspx or .cs extension) using this list of keywords. My main goal is find out which of the 1285 database objects are being referenced in these files. Can anyone recommend a tool that could accomplish this? Otherwise, I will just create my own. Thanks.

    Read the article

  • Localization of Database Strings in .Net

    - by Aligned
    I have several database tables that have a description column that I need to display in the UI. .Net has .resx files that will help with the translation of the strings, when the Thread.CurrentCulture.UICulture is set, but I needed a custom approach for the strings that are stored in the database and not in the .resx files. Here’s my approach: 1. Create a resource file for each database table and put them in the /Resources/Database/ directory. 2. Create a method in LocalizationHelpers (GetLocalizedString) that will get the string from the table for English (which should be cached to avoid unneeded service/database calls) or the resx when not English. 3. All database tables need to have a ResxKey field that matches the key in the resx file. 4. By convention the resx file will have the same name as the database table, and the key the same as the database ResxKey.   - if there are multiple columns that need translation, then one ResxKey will be used and Name or Description appended. Here’s the method I’m using to pull the string: public static string GetLocalizedString(string resourceName, string resourceKey) { if (executingAssembly == null) { executingAssembly = Assembly.GetExecutingAssembly(); } ResourceManager manager = new ResourceManager(resourceName, executingAssembly); return manager.GetString(resourceKey); }

    Read the article

  • Emacs doesn't use ~/.ssh/config when accessing files on a remote machine

    - by Yotam
    I have a fresh install of arch Linux. I've installed Emacs from the rpos, and my home directory is mounted from a separate partition. I have old settings I've used on my ~/.ssh/config along with authentication keys I've regularly used before. Now, when I try to connect to a remote machine using Emacs, Emacs asks for my password and uses the wrong username. Clearly, Emacs doesn't access my config file. When I try to ssh or scp directly to the machine, things work fine. What do I need to update?

    Read the article

  • Apache Alias - Chiliproject

    - by asdz
    I'm trying to setup Chiliproject (a ruby application for project management) I have setup my Apache already. However I want the Chiliproject to be like http://abc.com/Chiliproject as I want the abc.com to be used for other application. Following is my Chiliproject vhost setting: ServerName abc.com DocumentRoot /var/www/chiliproject/public Alias /chiliproject /var/www/chiliproject/public Options -MultiViews AllowOverride all When I go to abc.com, the Chiliproject page will appear but when I go to abc.com/chiliproject, I will reach the 404 page not found instead. If I change the DocumentRoot to /var/www, the page abc.com will be what I want, but the abc.com/chiliproject will comes to the 'Directory view' of my page.

    Read the article

  • Run BGInfo At Startup For All Users

    - by slickboy
    I have a Windows 7 image which I intend to deploy across a business. For simplicity I intend to install BGInfo on each machine and have it update each time a user logs in. From what I can see, when BGInfo creates a configuration file, the file contains variables which are local to each account - and therefore the configuration file will only work on the user account that created it. Has anyone any idea as to how make these configuration files 'generic' so that BGInfo will work for all accounts when they log in? At present I have the BGInfor application and a BGInfo configuration file saved on the C drive and I have written a batch file which is stored in the 'All Users/Start Menu/Startup' directory (which executes every time any user on the computer logs in), however this only works for the account which created the configuration file. Thanks for any help.

    Read the article

< Previous Page | 489 490 491 492 493 494 495 496 497 498 499 500  | Next Page >