Search Results

Search found 5711 results on 229 pages for 'jim m somewhere'.

Page 99/229 | < Previous Page | 95 96 97 98 99 100 101 102 103 104 105 106  | Next Page >

  • Home and End keys in Emacs don't work when run from Tmux

    - by Jan Stolarek
    When I run Emacs from Tmux, the Home and End keys do not work (Home key runs the Search command as if C-s was pressed). The problem started when I added this in my ~/.bashrc file: TERM="xterm" export TERM I've read somewhere that TERM variable should not be set manually but this was the only way I was able to solve problems with colors. Without this setting I got different colors in Emacs when running directly from the terminal and different when running from Tmux. This option caused some of the keys not to work in Emacs when it was run from Tmux, so I added this line to my ~/.tmux.conf: set-window-option -g xterm-keys on This solved problem with all keys except for Home and End. Any ideas how to make these keys work again?

    Read the article

  • New Mac OS X Server setup, when i send mail to gmail it goes straight to Spam. Why is that?

    - by basilmir
    New Mac OS X Server setup, when i send mail to gmail it goes straight to Spam. Why is that? My setup: DNS - done (A records PTR are ok) Mail Setup - done Webmail - done Also there seems to be a naming problem. They all come from [email protected] instead of [email protected]. I must be missing an alias somewhere. I've read an entire book on setting this up so don't throw stones :) The GUI is masking a lot of this up for me, so explanations via GUI is appreciated.

    Read the article

  • OpenSSH SFTP server with chroot() + File Permissions

    - by HannesFostie
    I am currently setting up an SFTP server but there is one detail I can't seem to figure out. When I add a user, I would like him to connect using his client and be able to write in his "root dir" right away. My Match case for the SFTP-users group currently has ChrootDirectory set as "/home/%u", and inside that directory I have to have a subdirectory owned by the user, while /home/%u itself is owned by root. Next to that, the "root dir" also has a couple files, .bashrc to name one. Is it possible to put these files somewhere else, remove them, or at least make them invisible to the user? Thanks EDIT: One more little thing I'd like to implement is for one account to have read (or rw, not sure yet) access to all other users' home directories. What is the easiest way to implement this?

    Read the article

  • Unexpected network traffic?

    - by robwalker
    My internet connection is via a fixed wireless connection using a 900MHz Motorola Canopy module. The router reports a fairly consistent 32-64Kbps of incoming traffic on the WAN port. When I attached a PC directly to the port and run Wireshark, I get a dump showing a lot of chatter from other machines that I presume are connected to the same tower. This didn't include end-to-end traffic, but was there were a lot of ARP requests, SSDP traffic, ICMP and other network discovery type stuff. Is this 'normal' or does it suggest a misconfiguration somewhere? As far as I can tell there is no need for my modem to be receiving any of this traffic (other than wanting to know what the names of my neighbours machines and printers are!) Since the internet connection is slow at the best of time, having this amount of background noise seems very wasteful.

    Read the article

  • PHP + IIS Application Pool Identity Windows\Temp permissions

    - by Matt Boothman
    I am currently running PHP (5.3) on IIS 7.5 on a Win2k8 R2 Web Edition Server and would like to know what, if any, problems or security vulnerabilities I may introduct into a system by assigning Read, Write, Modify & Execute permissions to either IUSR account or the IIS_USERS group for %SystemRoot%\Temp? Should I be altering permissions to that folder at all (as Windows reminds me I probably shouldn't when i attempt to change them)? Should I create a temp folder somewhere else and set permissions accordingly? The problem is when i set Anonymous Authentication (I'm guessing is a more secure option???) to use the App Pool identity, when starting sessions PHP gets stuck in a loop because it's unable to create session files in the %SystemRoot%\Temp folder due to lack of permission on the application pool user or IIS_USERS group. Another problem being ImageMagick (PHP Extension) is being denied access to %SystemRoot%\Temp to write temporary files so is throwing exceptions. I have tried searching Google however have not found anything that touches upon this subject specifically. Any help greatly appreciated.

    Read the article

  • SQL Server 2008 R2 Replication log reader could not execute sp_replcmds

    - by user49352
    This log reader agent worked perfectly for several months until the user referenced in the error was removed from the domain. After that time the error 'The process could not execute 'sp_replcmds' on 'SERVER'' was received with more detail 'Could not obtain information about Windows NT group/user' that referenced said user. This user was referenced nowhere in the the log reader agent other than the Publication Access List from which it was subsequenctly removed. The agent would still not successfully start up. The simple problem here, I believe, is that the log reader agent was created under that user and that no longer exists in the domain. Is there an 'owner' somewhere that needs to be changed? Every other function on the database continues to execute successfully. Any other help or thought would be appreciated.

    Read the article

  • Decentralized synchronized secure data storage

    - by Alberich
    Introduction Hi, I am going to ask a question which seems utopic for me, but I need to know if there is a way to achieve what I need. And if not, I need to know why not. The idea Suppose I have a database structure, in MySql. I want to create some solution to allow anyone (no matter who, no matter where) to have a synchronized copy (updated clone) of this database (with its content) Well, and it is not going to be just one synchronized copy, it could (and should) be a multiple replication (supposing the basic, this means, for example, ten copies all over the world) And, the most important thing: It must be secure. By secure I mean only real-accepted transactions will be synchronized with all the others (no matter how many) database copies/clones. Note: Since it would be quite difficult to make the synchronization in real-time, I will design everything to make this feature dispensable. So it is not required. My auto-suggestion This is how I am thinking to manage it: Time identifiers and Updates checking: Every action (insert, update, delete...) will be stored as the action instruction itself, associated to the time identifier. [I think better than a DATETIME field, it'll be an INT one, with the number of miliseconds passed from 1st january 2013 on, for example]. So each copy is going to ask to the "neighbour copy" for new actions done since last update, and execute them after checking they are allowed. Problem 1: the "neighbour copy" could be outdated too. Solution 1: do not ask just one neighbour, create a random list with some of the copies/clones and ask them for news (I could avoid the list and ask ALL the clones for updates, but this will be inefficient if clones number ascends too much). Problem 2: Real-time global synchronization is not active. What if... Someone at CLONE_ENTERPRISING inserts a row into TABLE. ... this row goes to every clone ... Someone at CLONE_FIXEMALL deletes this row. ... and at the same time, somewhere in an outdated clone ... Someone at CLONE_DROPOUT edits this row (now inexistent at the other clones) Solution 2: easy stuff, force a GLOBAL synchronization before doing any new "depending-on-third-data action" (edit, for example). This global synch. will be unnecessary when making an INSERT, for instance. Note: Well, someone could have some fun, and make the same insert in two clones... since they're not getting updated in real-time, this row will exist twice. But, it's the same as when we have one single database, in some needed cases we check if there is an existing same-row before doing the final action. Not a problem. Problem 3: It is possible to edit the code and do not filter actions, so someone could spread instructions to delete everything, or just make some trolling activity. This is not a problem, since good clones will always be somewhere. Those who got bad won't interest anymore. I really appreciate if you read. I know this is not the perfect solution, it has possibly hundred of holes, but it is my basic start. I will now appreciate anything you can teach me now. Thanks a lot. PS.: It could be that all this I am trying already exists and has its own name. Sorry for asking then (I'd anyway thank this name, if it exists)

    Read the article

  • Personal wiki on usb / the cloud?

    - by drby
    I'm looking for a personal wiki that can be installed on an usb stick and (more importantly) somewhere on the cloud (dropbox). I've looked at the Wiki Matrix, but I really don't care that much about any of the options, so I end up with a choice between ~50 wikis at the end. Tried out TiddlyWiki but there are some things that really annoy me like the fact that all pages get opened on the same page. It really looks like it'd turn into a giant mess pretty quickly. I'd like to have something that's pretty close in terms of appearance and usability to wikipedia. Hierarchical categories for organization would be really nice. And accessible storage (in case I ever want to convert it to something else).

    Read the article

  • Building boost 1.42 on FreeBSD 6.3

    - by Ivan Perekluyev
    Hi, i need to build mapnik on freebsd 6.3, but port marked as 'broken', so i forced to build it from source. With boost 1.41 (which is in ports) mapnik doesn't build. somewhere in internet, i found that mapnik successfully builded with boost 1.42. So, i download patch from wiki.freebsd.org/BoostPortingProject andd apply it: wget http://alexanderchuranov.com/boost-port/boost-from-1.41-to-1.42-2010-02-16-17-11.diff cd /usr/ports patch -p0 -i ~/boost-from-1.41-to-1.42-2010-02-16-17-11.diff after that, i trying to install boost-all metaport, but its failed. cd devel/boost-all make install 2>&1 | tee build.log tail -n 100 build.log > short_build.log Build.log (attention, 5m !): dl.dropbox.com/u/7365614/build.log Short build log: http://paste.pocoo.org/show/224474/ Thanks!

    Read the article

  • PHP short_open_tag Won't Enable (CentOS 6)

    - by brack
    I'm setting up an Apache2, PHP 5.3.3 server running on CentOS 6. My web application uses the short tags <? and <?=. I can't seem to get the short tags enabled. When I run phpinfo() I see short_open_tag = off, however in /etc/php.ini I have this: short_open_tag = on (and yes I've restarted the server). I've also tried using <?php ini_set('short_open_tag','1'); ?> at the start of a page and it still doesn't parse code in short tags. The only thing I can think of is there's another php.ini file somewhere that is being used instead of, or overriding the one at /etc/php.ini. Any advice?

    Read the article

  • my linux problems and solutions [closed]

    - by Delirium tremens
    I read somewhere in StackOverflow or StackOverflow Meta that if I had a problem, then solved it myself, I can share the problem and solution with you. How do I? in Linux: remove unneeded packages using apt-get play spc and psf update the system using apt-get in Mint: install lamp install and configure xdebug enable xdebug for cakephp install bazaar colo rename a repository directory when bazaar explorer fails init a repository when bazaar explorer fails use ssh key with launchpad uninstall firefox 3 when synaptic fails install minefield make pearltrees load when flash fails edit clojure documents install compojure create a new compojure project in Kubuntu enable phpmyadmin after installing lamp stop MySQLdb module error in webpy in Ubuntu stop the mouse pointer from disappearing fix the color stop sync read-only filesystem error stop download prompt instead of site enable phpmyadmin after installing lamp enable mod_rewrite after installing lamp

    Read the article

  • suPHP permission requirements for all files or only PHP scripts?

    - by puk
    Generally with PHP, files/folders have a permission of 0777 when we want to write to them. suPHP forbids a permission of 0777. Instead, files are supposed to have a permission of 0644 and folders have a permission 0755. However, this is always worded differently SuPHP won't allow chmod of 777 so we need to set all files to 644 and directories to 755 Your scripts and directories can now, only have a maximum of 755 permissions Set file permission(s) to allow read access only by you (e.g., chmod 600 filename.suphp). What is the exact rule and to what files does it apply to? If I have README file somewhere in a nested directory does its permission need to be updated, or does it only apply to .php files?

    Read the article

  • How can I keep websites from knowing where I live?

    - by D Connors
    This questions is related to issues and practicality, not security. I live in Brazil and, apparently, every single website I visit knows about it. Usually that's ok, but there are quite a few sites that don't make use of that information adequately. For instance: Bing keeps thinking that brazilian pages are way more relevant to me than american ones (which they're not). Google.com always redirects me to google.com.br. Microsoft automatically sends me to horribly translated support pages in portuguese (which would just be easier to read in english). These are just a few examples. Usually it's stuff I can live with (or work around), but some of them are just plain irritating. I have geolocation disabled in firefox, so I guess they're either getting this information from my IP or from windows itself (which I bought here). Is there a way to avoid this? Either tell them nothing or make them think I live somewhere else? Thanks

    Read the article

  • Why won't apache load a symlinked file from conf.d?

    - by kdt
    I have an apache configuration file which works fine when it's placed directly in /etc/httpd/conf.d/foo.conf. However, when I move the same file somewhere else (for example, move it to /tmp/foo.conf) and then create a symlink with ln -s /tmp/foo.conf /etc/httpd/conf.d then apache fails on startup with: httpd: could not open document config file /etc/httpd/conf.d/foo.conf I've tried making the file and the symlink mode 777, and tried changing them to be owned by the apache user. It seems like apache is failing to load the file purely on the basis of it being a symlink, but I'm sure I've used symlinks successfully on other machines. Is there something I'm missing? Does apache have an option for refusing to load config files if they're symlinks? The operating system is CentOS 4.4, apache version 2.0.52.

    Read the article

  • Updating AFP version on Windows Server 2003

    - by Niclas Lindqvist
    Hi, We have a couple of Macs, running Leopard and Snow Leopard in our otherwise pure Windows environment. We cannot get file names longer than 31 characters and some times the file permissions gets scrabbled (users can't delete their own folders or access files they usually access). This only occurs when connecting through AFP, We've tried with SMB but it's horribly slow when working with larger documents, so it's not an option. I somewhere read about AFP 3.0 being the answer. So my question is: Is there some way of updating my "AFP version" on Windows Server 2003, or does anyone have a different idea of what might help to resolve this situation?

    Read the article

  • Best way to find the computer a user last logged on from?

    - by Garrett
    I am hoping that somewhere in Active Directory the "last logged on from [computer]" is written/stored, or there is a log I can parse out? The purpose of wanting to know the last PC logged on from is for offering remote support over the network - our users move around pretty infrequently, but I'd like to know that whatever I'm consulting was updating that morning (when they logged in, presumably) at minimum. I'm also considering login scripts that write the user and computer names to a known location I can reference, but some of our users don't like to logout for 15 days at a time. If there is an elegant solution that uses login scripts, definitely mention it - but if it happens to work for merely unlocking the station, that would be even better!

    Read the article

  • What differences are there between an official Ubuntu AMI image and a base install from an ISO?

    - by David Winter
    When creating a new instance on AWS using an official Ubuntu 12.04 server AMI, what differences are there compared to if I was to do a standard server install on a computer of my own? For example, the default user is 'ubuntu'. An SSH public key is added to that users authorized_keys file. Sudo is passwordless for that user. PasswordAuthentication is disabled for SSH. etc etc. Configurations have been changed from their defaults, and I'd like to know if there is a list, or somewhere I could find out the modifications made.

    Read the article

  • rsyslog appears to act on old configuration

    - by Jeff Lee
    I'm using a template to dynamically generate rsyslog filenames. I've made some changes from my original format, but rsyslog still appears to be using both the new template and the old after restarting. My filename template went from this: $template RemoteDailyLog,"/var/log/remote/%hostname%/%$year%/%$month%/%$day%.log" To this: $template RemoteDailyLog,"/var/log/remote/%hostname%/%fromhost-ip%/%$year%/%$month%/%$day%.log" I stopped rsyslogd using service rsyslog stop, deleted all of my log files using rm -rf /var/log/remote/*, and then restarted rsyslogd with service rsyslog start. The problem is rsyslog seems to be building folder structures of the type "/var/log/remote/%hostname%/%$year%/%$month%/%$day%.log" (i.e., without the remote IP), which no longer appears anywhere in my configuration. Is it possible that old log or config data have been cached somewhere and are being preserved through the server restart? This is creeping me out a little.

    Read the article

  • How to set IP address of Amazon EC2 instance to its Elastic IP?

    - by TWord
    Hi, I have an Amazon EC2 instance running and I am installing a program on it that needs to know what the machine's IP address is. Can I set the Elastic IP address to the IP address within the EC2 instance? Its okay if it reroutes data packets somewhere 'outside' and then back to itself, but the software NEEDs me to specify an IP address of the machine its on. I proceeded with the software installation using the "local IP" (10.xx.xx.xx) within the software installation. I don't know if this is the reason why the application is not visible publicly (as I'm trying to determine in the question http://serverfault.com/questions/166946/allowing-web-access-to-an-amazon-ec2-windows-server-2008-instance-running-tomcat)

    Read the article

  • MacBook Pro Trackpad freeze / holds click on its own

    - by lajuette
    My MacBook Pro's Trackpad shows an annoying behavior in the last few days. I think the only relevant thing that changed is the firmware for the Trackpad. An update has been distributed via Software Update a few days ago. Sometimes when i click somewhere the Trackpad won't recognize the mouseup, i.e. hold the click on its own. Clicking again seems to stop the holding and issue a new mousedown, but then again, no mouseup I.e. when i click on a file on the desktop the file will be dragged. But i can't stop dragging! Clicking in a browser or text document will select the text/images. Did anyone notice similar behavior? Any ideas or fixes?

    Read the article

  • Mount Docker container contents in host file system

    - by dflemstr
    I want to be able to inspect the contents of a Docker container (read-only). An elegant way of doing this would be to mount the container's contents in a directory. I'm talking about mounting the contents of a container on the host, not about mounting a folder on the host inside a container. I can see that there are two storage drivers in Docker right now: aufs and btrfs. My own Docker install uses btrfs, and browsing to /var/lib/docker/btrfs/subvolumes shows me one directory per Docker container on the system. This is however an implementation detail of Docker and it feels wrong to mount --bind these directories somewhere else. Is there a proper way of doing this, or do I need to patch Docker to support these kinds of mounts?

    Read the article

  • copy/paste on MacOS gets 'Stuck'

    - by bmargulies
    SnowLeopard. I commonly use copy in a terminal window to grab stuff I have to paste. A few times in the last few days, the clipboard has gotten 'stuck' on some old text. I do the copy gesture (with keyboard or Edit menu) in the terminal window. I paste somewhere else, and get something other than what I just copied. If all else fails, I can always tell the terminal to write it's content to a file and use that. But I'm wondering if there's something I'm doing to get into this pickle and if there's any way out.

    Read the article

  • Safe to Use an Amazon EBS Volume While Snapshot in Progress?

    - by Justin Noel
    Is it safe to use an EBS volumne while a snapshot is being created? I've currently got a 100Gb EBS volume mounted. I am in the process of snapshotting it. Goodness it's slow!! It's going to end up taking more than 45 minutes to snapshot. My question : Is the EBS volume already copied and just being saved somewhere? Or, is the snapshot actively copying from my mounted volume right now? Basically, if I start using it before the snapshot completes, am I hosed? I just can't believe it takes this long to copy. There really isn't even 100GB in use. It's more like 25Gb.

    Read the article

  • I want to version control my entire slice

    - by Tom
    I'm renting a slice (i.e., a VPS) from Slicehost. I've a spent a day or two filling up /usr with my favorite packages, /etc with configs and init scripts, and so on. Now I want to: save this whole setup somewhere (e.g., to load onto another machine). see what changes I've made to which files revert changes, tag revisions, and all that other good version control stuff Saving a disk image gives me (1), but not (2) and (3). Using Subversion (svn import / svn://someotherhost) might give me all three, but I expect problems if I actually try to check a project out into / and maintain .svn directories in root-owned areas. And to load my setup onto a fresh slice, I'd need to install an svn client on it first. Is there a good way to do what I want to do?

    Read the article

  • Configure Postfix for outgoing mail

    - by Savvas Sopiadis
    I have the following scenario i must implement using Postfix (but don't know how to begin): Say we have a domain aaa.com which is hosted somewhere (this is already functioning without any problems). On this envirnoment there is already a Mail server though which we can send and receive mail (this mail server is limiting us to a very small number of emails/day). Now we have setup a VPS (virtual private server) on which we installed Postfix, which sole porpuse is to send emails on behalf of domain aaa.com. The Postfix server will be used by a program like Outlook to send the email. What configuration has to be done? (i 'm a bloody beginner in this field!) Thanks in advance

    Read the article

< Previous Page | 95 96 97 98 99 100 101 102 103 104 105 106  | Next Page >