Search Results

Search found 26640 results on 1066 pages for 'a troubled linux newbie'.

Page 268/1066 | < Previous Page | 264 265 266 267 268 269 270 271 272 273 274 275  | Next Page >

  • Subscribe to feed in Thunderbird from the command line?

    - by Coderer
    From reading around the web, it looks like Firefox's "quick view" of an RSS feed sometimes lets you "Subscribe to this feed using" Thunderbird. For whatever reason, that's not an automatically-added option with my setup (FF 3.5.something + Thunderbird 3.0.something on Linux), so I figured I could just "Choose Application...", point at the Thunderbird binary, and be on my way. Not so -- nothing appears to happen. If I run thunderbird from the command line as thunderbird "http://path/to/feed" the app launches as normal. If it's already running, absolutely nothing happens. Is this impossible? Is there some mojo I can pass Firefox to tell it that Thunderbird exists? Should I just suck it up and copy/paste the URLs manually?

    Read the article

  • remove apache tar binary package in centos

    - by user119720
    I need help in removing binary package that had been installed in my linux machine. The scenario that I am having are such as like this: I've already install latest apache that I get from its website(httpd.apache.org) through Unix binary package(tar.gz) After successfully install program,the apache server web perfectly without any issues. But then I having a thought,if there are another latest release version of apache in the future,then I need to make sure that i can upgrading current apache or reinstalling the new version apache. So my question is, how do I ensure that I have remove the old remove apache and all its dependencies so that it will not having conflict(probably) when installing the new apache. Right now the only thing that i can think of is to remove all the apache folder manually : rm -rf /apache2 Hope someone can shed some light about this.Thanks.

    Read the article

  • Can gedit on mac be used to edit files over ssh?

    - by Dave
    I use a linux machine at work and a mac at home. I can ssh from my machine at home to my work machine. But the only editor that I have access to on the command line then is vi, which I don't like. Is there a way to use gedit on my mac to edit files remotely over an ssh connection? This page says that it can be done, but I think that it assumes that you are using gedit on ubuntu. On my mac (os 10.5.8) I don't have the "bookmark" option when I click "connect to server". http://thecodecentral.com/2010/04/02/use-gedit-as-remote-file-editor-via-ftp-and-ssh-ubuntu/comment-page-1#comment-50558

    Read the article

  • Why is "chmod -R 777 /" destructive?

    - by samwise
    This is a Canonical Question about File Permission and Why 777 is "destructive". I'm not asking how to fix this problem, as there are a ton of references of that already on Server Fault (reinstall OS). Why does it do anything destructive at all? If you've ever ran this command you pretty much immediately destroy your operating system. I'm not clear why removing restrictions has any impact on existing processes. For example, if I don't have read access to something and after a quick mistype in the terminal suddenly I now have access well... why does that cause Linux to break?

    Read the article

  • Ubuntu from console/command-line/shell

    - by Xolve
    Earlies linux distros though required lot of manual work they were quite good to use from commandline. If the X-server didn't start or you just want a shell to work they all supported. Network was configured by init; sound was up and ready; new devices inserted would be configured and their configureation was placed in fstab. Also there were small scripts I found on many distros which on X used windows while on console they switched to ncurses. But now this all needs GUI with a desktop manager (KDE, GNOME) for the new paradigms :'-( require GUI (NetworkManger, hal etc.). So if on just command line you have to be root, looks like they believe only geeky admins need that, and need to edit config files or type big commands. Any way so that this is easy in Ubnubtu through shell again.

    Read the article

  • Unix Shell/SSH config to allow TCP port forwarding without showing a command prompt

    - by Raphael K
    I'm running a Debian Linux. I'd like to have a user account that is able to connect via SSH for TCP-forwarding only, without a command prompt. e.g the following would work (from a remote computer): ssh -D1234 user@myhost but no command prompt would appear. Using a shell like /bin/false or /sbin/nologin is too restrictive as it doesn't even allow the user to log in. A shell that only allows the "exit" or Ctrl+D commands would do the job. I know that something similar is possible to allow only SFTP, but I can't find the equivalent for TCP forwarding. Thanks

    Read the article

  • Wifi and eth behavior

    - by r00ster
    I have a wireless router 150M Wireless Lite N Router Model No. TL-WR740N / TL-WR740ND. Normally, when I'm connected to the local network using eth0 I can ping other machines by issuing ping name. When I'm connected through wifi I have to issue ping name.domain.com. The machine is only visible in intranet. How to achieve the same behavior with wifi? The second problem is, that I can not connect to some external sites through wifi but through eth everything is ok. I guess that is related to some port forwarding, but I'm not sure. How can I resolve this issue? EDIT: I'm using Linux Mint.

    Read the article

  • need for tcp fine-tuning on heavily used proxy server

    - by Vijay Gharge
    Hi all, I am using squid like Internet proxy server on RHEL 4 update 6 & 8 with quite heavy load i.e. 8k established connections during peak hour. Without depending much on application provider's expertise I want to achieve maximum o/p from linux. W.r.t. that I have certain questions as following: How to find out if there is scope for further tcp fine-tuning (without exhausting available resources) as the benchmark values given by vendor looks poor! Is there any parameter value that is available from OS / network stack that will show me the results. If at all there is scope, how shall I identify & configure OS tcp stack parameters i.e. using sysctl or any specific parameter Post tuning how shall I clearly measure performance enhancement / degradation ?

    Read the article

  • Port forwarding with DNAT and SNAT without touching other packets

    - by w00t
    I have a Linux gateway with iptables which does routing and port forwarding. I want the port forwarding to happen independent of the routing. To port forward, I add this to the nat table: iptables -t nat -A "$PRE" -p tcp -d $GW --dport $fromPort -j DNAT --to-destination $toHost:$toPort iptables -t nat -A "$POST" -p tcp -d $toHost --dport $toPort -j SNAT --to $SRC $PRE and POST are actually destination-specific chains that I jump to from the PREROUTING and POSTROUTING chains respectively so I can keep the iptables clean. $SRC is the IP address I'm SNATing to which is different from the gateway IP $GW. The problem with this setup is that regular routed packets that were not DNATed but happen to go to the same $toHost:$toPort combo will also be SNATed. I wish to avoid this. Any clever things I can do?

    Read the article

  • /srv/mm/Music (etc) Twonky won't scan here for media

    - by Hamid
    Is there something special about /srv/mm/ that Twonky server refuses to scan there? I previously had my system set up with all my Music, Video and Photo folders in /srv/mm shared by Samba, miniDLNA etc, with no problems. I came to install Twonky to replace miniDLNA and after two days of tearing my hair out, changing permissions and owners of the directories I ended up making a new directory at /multimedia and moved my Music, Video and Photo folders in there. Twonky then scanned them all straight away with no problems. I'm running Arch Linux (plugapps specifically) on a NAS. The solution is already implemented (moving the directory) I'm just wondering technically why Twonky might have refused to look for my media in the /srv/mm directories.

    Read the article

  • Cannot connect to FTP server from external host

    - by h3.
    I have a FTP server (vsftpd) setuped on a Linux box (Ubuntu server). When I try to connect with a computer on the same network everything works fine as expected. But as soon the IP is external it won't connect.. I first assumed the port was blocked, but then: localserver:$ sudo tail -f /var/log/vsftpd.log Wed Jan 13 14:21:17 2010 [pid 2407] CONNECT: Client "xxx.xxx.107.4" remotemachine:$ netcat svn-motion.no-ip.biz 21 220 FTP Server And it hangs there. Do any ports other than 21 need to be open?

    Read the article

  • rsync server side limit bandwidth/connection

    - by c2h2
    In a VOIP application, I have upto 3000 clients rsync audio files from there linux server in a daily, server is placed at a data center (10Mbps in/out bound), the server works as a VOIP sip server running FreeSWITCH (low ping latency should be ensured.) Therefore I would like to have server side control of rsync which controls: Limit total outbound bandwidth. Limit total number of connections. (Reject clients while at max number of connection and let it retry after a specific time frame.) OPTIONAL: list/kill individual connections. Normally I would use ssh + rsync + pem_keys with some extra options, but above requirements are not feasible by simple command lines. Can anyone point me some direction. or show some scripts/tools? I would also probably integrate them and release on github. Thanks!

    Read the article

  • How can I fix my C++ compiler as it isn't loaded by default?

    - by GNR
    Recently I had installed the Ubuntu flavour of the Linux operating system. I had opened a terminal and just wrote a sample C program to check if it is compiling. When I saved the sample file and compiled with cc a.c, errors comes that the standard library is not loaded (i.e stdio.h). When I went to help pages, it says that the C or C++ compiler doesnt gets loaded by default and we should do it ourselves. So can anyone help me out to fix this problem, i.e to load the C/C++ compiler.

    Read the article

  • Configure redis to not have everything in memory?

    - by acidzombie24
    I like redis because it lets me do operations on data structures. I wanted to see what would happen if i were to put more data into redis then i have for ram. So i wrote a loop that inserted 30k bytes repeatedly and set maxmemory 100MB. I figure it would stay at 100mb. It kept growing. Past 1gb then past 2gb. Suddenly it crashed because i ran the 32bit version. Now... i dont understand what the point of maxmemory is? I am using the windows version so maybe its ignored. Does redis have to have everything in memory? If i have a site (on linux) with a 10gb database and 512mb on the machine will redis work? I don't need it to be amazingly fast i just prefer to modify data in it then sql (although i hope it is still faster then mysql)

    Read the article

  • How do I setup Apache with FastCGI and Ruby?

    - by sdellysse
    I've been reading alot of forum posts, tutorials, etc., about setting up apache under linux with fastcgi. I'm trying to run fcgi for some of my ruby development (not rails), but I figure it should apply to any language. Please note that I already have apache, fastcgi, ruby-fcgi, and the like already installed on my web box. Conceptually, I'm having a hard time with it. Does apache spawn the specified fcgi processes as needed? Do I start them along side apache? Basically, I'm trying to get to a point where I am able to ask the proper questions, so apologies if the title is misleading.

    Read the article

  • How do you synchronise huge sparse files (VM disk images) between machines?

    - by chrisdew
    Is there a command, such as rsync, which can synchronise huge, sparse, files from one linux server to another? It is very important that the destination file remains sparse. It may be longer (but not bigger) than the drive which contains it. Only changed blocks should be sent across the wire. I have tried rsync, but got no joy. groups.google.com/group/mailing.unix.rsync/browse_thread/thread/94f39271980513d3 If I write a programme to do this, am I just reinventing the wheel? http://www.finalcog.com/synchronise-block-devices Thanks, Chris.

    Read the article

  • How to install ubuntu OS using an android mobile as a usb drive?

    - by Rochelle Canale
    I just want to ask if How can I install Ubuntu OS from android phone? I have the ubuntu.img and .md5 file. But I don't knwo how to run this. I also have Linux installer and VNC. Everytime I launch the application the terminal emulator always force close. So I think that there's an error with the application. Any idea how can I install this? I am using myphone A848 duo same as Noir A6 and micromax A68.

    Read the article

  • Deactivate GPU and use IGP

    - by squelos
    I am having some trouble with my Sony Vaio SVE1511W1e laptop. It has an ATI Radeon and the i5 has an IGP (i5 2450m). I don't often use my GPU, and the IGP would be just enough for most usage I do. Therefore, in order to improve the battery life, I wish to deactivate the GPU and use only the IGP. The problem is that my BIOS doesnt allow me to do so. But I believe it is possible to deactivate the GPU 'programatically'. I'm running Debian Wheezy on the 3.2.0.4 AMD64 kernel. The first problem I'm running into is that when I run lspci, my IGP doesnt show up. Could this be because im lacking a kernel module? (I chose a targeted installation). What are the solutions to deactivating a GPU and using an IGP on a Linux System such as debian?

    Read the article

  • Unable to login into CentOS

    - by Rendl
    I had setup a multinode cluster using CentOS with VMware yesterday. Today when I reboot the nodes I get the below error on startup. "there is a problem with the configuration server status 256 centOS" (/usr/libexec/gconf-sanity-check-2 ) I am unable to login as root or any user as the screen is frozen. The solutions online is to change the permissions for some tmp files. My problem is I am unable to access the terminal as I cannot login. Also on reboot I do not have any recovery options in CentOS. I only see command line GRUB. I am new to linux and Hadoop.Pls help asap.

    Read the article

  • how to setup a ssh acount with no terminal but port forwarding?

    - by admalledd
    I am trying to set up a new user account I can give to friends so they can SSH into my forward computer, and only allow forwarding of certain ports. I do not want my friends to have a shell, or be able to change what ports to where they are allowed to forward. example session: joe(friend) connects using PuTTY (that I have pre-set, he isn't good with computers) to example.com(my Internet facing computer) forwarding ports 8080,1990,25565 to him(with what ever end ports he wants, preferably they stay the same numbers) example ssh command to do similar (but he can still change the ports on my computer!) ssh -N [email protected] -p443 -L8080:192.168.1.2:8080 -L1990:127.0.0.1:1990 -L25565:127.0.0.1:25565 then, same story with other friend smith(same ports, same user even) except he is using linux, so cant use putty. is it possible to also leave default SSH functinality for all other users but this one? I found this when I was searching google, but alas, I did not quite understand what was being suggested, and I don't think they covered restricting port forwarding

    Read the article

  • RAID 10, how layout works ?

    - by Bastien974
    I'm trying to figure out how exactly works the RAID 10 in linux with mdadm. I want to create a RAID 10 out of 4 partitions, let's say a, b, c and d. a and b are on the array 1, c and d array 2. So what I want is to have the couple a and b, c and d in RAID 0. Then on top of that, a RAID 1. The option in the mdadm command to configure the layout is -p, --layout with option : near, far, offset see here I want to keep my data safe if the array 1 fails for example, that would mean that every chunk of data are always copied on both arrays. How do I have to set my RAID 10, near or far ?

    Read the article

  • What is the difference between yum, apt-get, rpm, ./configure && make install

    - by Saif Bechan
    I am new to Linux and am running CentOs. When I want to update or install certain software I came across three ways. Sometimes it's: yum install program rpm -i program.rpm wget program.tar.gz unpack ./configure make make install That last one is a real pain, esp when you come from windows where a program install is usually one click and then a nice guide. Now can someone please explain to me: Why are there so many different ways to do this? Which one do you recommend to use and why? Are there any other ways for installing programs?

    Read the article

  • Resized NTFS partition, now it wont mount.

    - by H4Z3Y
    I have had a 1.5TB drive used as an external for 6 months or so, then I decided to put it in my linux server for network storage. ntfs was being crazy inefficient so I wanted to change the filesystem to ext4. I used the ntfsresize command to reduce the partition to 650GB and that took abour 2 hours, then I deleted all of the entries in fstab like a guide told me too and created a new one the size of the ntfs partition, or, 650GB. after I modified fstab the ntfs partition would no longer mount and when plugging it in to windows it says "This Hard Drive needs to be formatted". any ideas on how I can recover the data off of the drive? I have 600GB of free space on a different drive so I just need some way of copying them off.

    Read the article

  • Creating user accounts in Amazon EC2

    - by Tvanover
    I am putting together a test environment using Amazon's EC2 for me and some friends to collaborate on a project. I am not a server guy but I do know my way around a bash prompt and have done some work on ubuntu before. I am using Amazon Linux AMI i386 EBS and have gotten apache and php running. Now I need to create the user accounts my friends and I will use to upload files (sftp) and work on the project (ssh). How should I go about this? Should I just use adduser and configure it like normal? Or should I use the AWS IAM groups?

    Read the article

  • What are USB device resets?

    - by ILIV
    I have this brand-new software RAID-1 built with two WD Elements external USB drives, and I can see that Linux repeats the following message quite a bit: ... [302148.036912] usb 1-3.1: reset high-speed USB device number 19 using ehci_hcd [302153.052029] usb 1-3.3: reset high-speed USB device number 20 using ehci_hcd [302186.031481] usb 1-3.3: reset high-speed USB device number 20 using ehci_hcd [302217.050210] usb 1-3.3: reset high-speed USB device number 20 using ehci_hcd [302281.043543] usb 1-3.3: reset high-speed USB device number 20 using ehci_hcd [302312.090158] usb 1-3.3: reset high-speed USB device number 20 using ehci_hcd [302351.076851] usb 1-3.1: reset high-speed USB device number 19 using ehci_hcd ... So, what does it say? Is this normal? Is this a problem that I should try to fix?

    Read the article

< Previous Page | 264 265 266 267 268 269 270 271 272 273 274 275  | Next Page >