Search Results

Search found 238 results on 10 pages for 'blankman'.

Page 1/10 | 1 2 3 4 5 6 7 8 9 10  | Next Page >

  • How to debug a .bash_profile

    - by Blankman
    I was updating my .bash_profile, and unfortunetly I made a few updates and now I am getting: env: bash: No such file or directory env: bash: No such file or directory env: bash: No such file or directory env: bash: No such file or directory env: bash: No such file or directory -bash: tar: command not found -bash: grep: command not found -bash: cat: command not found -bash: find: command not found -bash: dirname: command not found -bash: /preexec.sh.lib: No such file or directory -bash: preexec_install: command not found -bash: sed: command not found -bash: git: command not found My bash_profile actually pulls in other .sh files (sources them) so I am not exactly sure which modification may have caused this. Now if I even try and to a list of files, I get: >ls -bash: ls: command not found -bash: sed: command not found -bash: git: command not found Any tips on how to trace the source of the error, and how to be able to use the terminal for basic things like listing files etc?

    Read the article

  • Every command fails with "command not found" after changing .bash_profile?

    - by Blankman
    I was updating my .bash_profile, and unfortunetly I made a few updates and now I am getting: env: bash: No such file or directory env: bash: No such file or directory env: bash: No such file or directory env: bash: No such file or directory env: bash: No such file or directory -bash: tar: command not found -bash: grep: command not found -bash: cat: command not found -bash: find: command not found -bash: dirname: command not found -bash: /preexec.sh.lib: No such file or directory -bash: preexec_install: command not found -bash: sed: command not found -bash: git: command not found My bash_profile actually pulls in other .sh files (sources them) so I am not exactly sure which modification may have caused this. Now if I even try and to a list of files, I get: >ls -bash: ls: command not found -bash: sed: command not found -bash: git: command not found Any tips on how to trace the source of the error, and how to be able to use the terminal for basic things like listing files etc?

    Read the article

  • Using tar, the entire folder structure is including, I don't want that

    - by Blankman
    I am taring a folder and for some reason the entire directory structure that preceds the folder I am tarring is included. I am doing this in a script like: 'tar czf ' + dir + '/asdf.tgz ' + dir + 'asdf/' Where dir is like: /Downloads/archive/ In the man pages, I see I can fix this but I can't get it to work. I tried: tar czf -C dir ... But now I have some kind of a file -C in my folder (which I can't seem to delete btw!). Please help!

    Read the article

  • Where can I change the webpage showed by apache?

    - by Blankman
    I have ubuntu 10.10 installed, and when I hit my IP I see: It works! This is the default web page for this server. The web server software is running but no content has been added, yet. Where should I look to see where this is being served from? The strange thing is, I just installed nginx and setup a static page that is served on port 80 also (should be conflicting with the default apache that is serving the 'it works page'), when I restart nginx I don't get bind error sayign port 80 is already taken. I dont' use apache, and don't want it but the default install seems to have it. I ahve no idea where it is, I don't see it in /opt or etc/apache . thanks!

    Read the article

  • Should all new web projects build their backend based on xml/json result sets?

    - by Blankman
    If you were building a new Saas project, would it make sense to start with all of the backend services returning xml/json? Because these days you need to build for both the web and mobile devices, and having a backend that is build from the start to return xml and json, you are ready to go mobile (all services have the business logic, so you won't be repeating anything). Now the web would be MVC, so the controller would just be routing the request to your service backend, and converting the json or xml to html. The obviousl downside is that you have to build a backend, and then another web project that calls your backend. But this also goes to you favor as it forces you to seperate your concerns, and not leak business logic in your controller/view layer. Thoughts?

    Read the article

  • Top reasons why programmers are fired. [closed]

    - by Blankman
    Have you ever been fired before, or one of your colleagues? What were the main reasons for him being fired? The objective of this question is to get a list/description of other peoples experiences too see if we can come up with a theme of why people are let go. For those that work for consulting companies, I would guess one of the main reasons would be a developer is just too slow pumping out code. So my guesses would be: 1) attitude 2) too slow Please specify if you are saying something from experience or just guessing.

    Read the article

  • Undertanding .bashrc and .bash_profile

    - by Blankman
    On a server when I login as root I see .bashrc (ubuntu 10.10). On my Mac I have a .bash_profile, does ubuntu always have only a .bashrc file and not .bash_profile? (just confused so asking, I realize they are different o/s's but maybe there is a relationship somehow?) On my server, I want to create an alias, should I put it in .bashrc? What if I want this alias to be applied so all users can use it?

    Read the article

  • Moving a .aspx site over to nginx, need to server .aspx as static files or rewrite

    - by Blankman
    I have a simple CMS site that was written in asp.net. My site looks like: www.example.com/ www.example.com/content/index.aspx www.example.com/content/get.aspx?id=234 (loads an article) It uses a database currently, but I am going to dump all the content to file, and then I can just pull the contents of the file based on the id=234 value. I want to move this site over to my ubuntu nginx server. What options do I have? Suggestions? I want to keep the URL structure exactly how it is now, how can I do this? Would this be any easier using apache?

    Read the article

  • Effects of HTTP/TCP connection handshakes and server performance

    - by Blankman
    When running apache bench on the same server as the website like: ab -n 1000 -c 10 localhost:8080/ I am most probably not getting accurate results when compared to users hitting the server from various locations. I'm trying to understand how or rather why this will effect real world performance since a user in china will have different latency issues when compared to someone in the same state/country. Say my web server has a maximum thread limit of 100. Can someone explain in detail how end user latency can/will effect server performance. I'm assuming here that each request will be computed equally at say 10ms. What I'm not understand is how external factors can effect overal server performance, specifically internet connections (location, or even device like mobile) and http/tcp handshakes etc.

    Read the article

  • nginx start failing, says error.log doesn't exist

    - by Blankman
    I structured my sites like: /home/www/domain.com/public,private, log, backup In the log folder, I created a blank error.log and access.log. My nginx file in sites-available for the domain looks like: server { access_log /home/www/domain1.com/log/access.log; error_log /home/www/domain1.com/log/error.log; } Trying to start nginx it says: starting nginx: the config file /etc/nginx/nginx/conf syntax is ok [emrg] open() ".../access.log" failed (2: no such file or directory) Is this a permission issue?

    Read the article

  • Is it possible to rate limit based on host headers? i.e. not just on ip address

    - by Blankman
    I have a web service endpoint that I am building where people will post an xml file to, and it will really get pounded with over 1K requests per second. Now they are sending in these xml files via http post, but a good majority of them will be rate limited. The problem is, the rate limiting will be done by the web application by looking up the source_id in the xml, and if it is over x requests per minute, it will not be processed further. I was wondering if I could do rate limit checking earlier in the processing somehow and thus save the 50K file going threw the pipeline to my web servers and eating up resources. Could a load balancer make a call out to verify rate usage somehow? If this is possible, I could maybe put the source_id in a host header so even the XML file doesn't have to be parsed and loaded into memory. Is it possible to just look at host headers and not load up the entire 50K xml file into memory? I really appreciate your insights as this takes more knowledge of the entire tcp/ip stack etc.

    Read the article

  • When connecting to a unix box, how does it know you have SSH setup on your desktop?

    - by Blankman
    When you use something like putty to connect to a linux box, and you setup your SSH keys etc. When connecting, how does it tell the server that you want to connect using your SSH keys? Is SSH running as a service on a particular port or does it simply pass your private-key and then the login service sees that and tries to connect using it? Just looking for a fairly high level understanding (with maybe some detail if you want to...)

    Read the article

  • Is there a max thread per mongrel?

    - by Blankman
    I don't know much about ruby, much less how or what is involved with hosting a ruby on rails web app. BUT, I recall hearing someone saying that they have to run multiple mongrels b/c of a limit of 50 threads? Is this true (or something similiar)? Why does it have this limitation?

    Read the article

  • how to add solr to run on startup?

    - by Blankman
    I asked this question before, but I think I got answers that were cron job type answers. I want the solr service to startup if I reboot my computer, and I also want it to run by itself instead of me opening up cmd.exe and running it manually, and then having to close my terminal connection (not log off as that will close it). What options do I have? I am on windows server 2008, IIS7 running a asp.net application that uses solr.

    Read the article

  • new windows 08 server, can't download using browser

    - by Blankman
    by default it seems you can't download anything using the browser on the server when terminal server'ing into the box. I need to download the java sdk to install solr, and also need to install winrar since I can't even unzip a file I uploaded from my mapped drive. What setting do I have to change?

    Read the article

  • Puppet nodes cant' find master, ec2 public versus internal ip addresses and hosts files

    - by Blankman
    If I setup my hosts files such that they reference all other ec2 nodes using the internal ip addresses, will this work or do I have to use the external ip addresses? Do I need to specify anything in my security group to get internal ip addresses to work? e.g. /etc/hosts ip-10-11-12-13.internal some_node_name If I do this, can I reference some_node_name anywhere in my scripts where I would have used the ip address previously? On my puppet agent servers, I have a reference to my puppet master like: public-ip-here puppet When I reboot my puppet agent's, syslog shows they couldn't find the master with the message: getaddinfo : name or service not known I did get it to work by updating /etc/default/puppet and I added to the options: --server=public-ip-here From what I read, puppet will by default try using 'puppet', and I set this in my hosts file so why wouldn't it be picking this up?

    Read the article

1 2 3 4 5 6 7 8 9 10  | Next Page >