Search Results

Search found 5658 results on 227 pages for 'eric fail'.

Page 4/227 | < Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >

  • Fail to download repository information

    - by BRvitorio
    Why lately I've been getting this error when I try to update my computer? Ubuntu 12.04 Kernel 3.5.2 Processor i386 W:Failed to fetch http://ppa.launchpad.net/kernel-ppa/ppa/ubuntu/dists/precise/main/source/Sources 404 Not Found, W:Failed to fetch http://ppa.launchpad.net/kernel-ppa/ppa/ubuntu/dists/precise/main/binary-i386/Packages 404 Not Found, E:Some index files failed to download. They have been ignored, or old ones used instead.

    Read the article

  • Why do packages fail to install from command line

    - by calebB
    Every once in a while I can't install packages in Ubuntu 11.04. I'll try to run the software center and it and it will pop up then disappear. If I call it from the command line it says it seg-faults. And then trying to install anything from the command line I get this: sudo apt-get install pdftk Reading package lists... Done And thats all I get for anything I try. Usually restarting fixes the problem but its very annoying.

    Read the article

  • Fail to access Network options

    - by Konstantinos Marinis
    I am trying to use OpenDNS for my newly installed Ubuntu 12.10... However I cannot insert custom DNS addresses... I am accessing Network, then at my wireless connection, no matter how many times I press the "options" tab at the low right corner (I am not using english Ubuntu, so the button might have a different name), nothing happens. Any ideas why or how should I configure my OpenDNS connection?

    Read the article

  • Why does Ubuntu 12.04 dual boot fail?

    - by Tranas
    Fresh install of XP followed by a fresh install of Ubuntu 12.04 results in the following error: error: unknown filesystem. grub rescue and the machine will not boot. Prior to the 12.04 install, XP worked fine. During the 12.04 install, all partitions and free space was visible, and the install seemed to complete without issues until the error message. Although I can fix the MBR via recovery console in XP and allow the machine to boot to windows, why is GRUB/Ubuntu trashing the boot sequence?

    Read the article

  • Is Assert.Fail() considered bad practice?

    - by Mendelt
    I use Assert.Fail a lot when doing TDD. I'm usually working on one test at a time but when I get ideas for things I want to implement later I quickly write an empty test where the name of the test method indicates what I want to implement as sort of a todo-list. To make sure I don't forget I put an Assert.Fail() in the body. When trying out xUnit.Net I found they hadn't implemented Assert.Fail. Of course you can always Assert.IsTrue(false) but this doesn't communicate my intention as well. I got the impression Assert.Fail wasn't implemented on purpose. Is this considered bad practice? If so why? @Martin Meredith That's not exactly what I do. I do write a test first and then implement code to make it work. Usually I think of several tests at once. Or I think about a test to write when I'm working on something else. That's when I write an empty failing test to remember. By the time I get to writing the test I neatly work test-first. @Jimmeh That looks like a good idea. Ignored tests don't fail but they still show up in a separate list. Have to try that out. @Matt Howells Great Idea. NotImplementedException communicates intention better than assert.Fail() in this case @Mitch Wheat That's what I was looking for. It seems it was left out to prevent it being abused in another way I abuse it.

    Read the article

  • Software load balancing fail-over vs hardware

    - by SmartLemon
    Please correct me, but my understanding is that with software load balancing a service must be run on each server while there is one DS that notifies the other servers that a server has gone down and that they should consume that servers load. With hardware load balancing what happens in a fail-over? Could someone explain? Is there advantages with using hardware load balancing when it comes to fail-over, or is there advantages with software? Or do they both have their pros and cons?

    Read the article

  • How to stop a random ramp in FCGI Processes Killing the server

    - by Andy Main
    So got the below earlier to day... Around that time the logs show a ramp in processes(600) and associated memory (1.2g), cpu usage load average (80) untill the server gave out. Server had to be hard reset by host as there was no ssh or plesk panel access. Fast CGI is configured as below and is setup for one high use site. As I understand it FcgidMaxProcesses 20 should protect against what happen but has not. I've read many forums with differing answers and references to many different fcgi directives, but have found nothing conclusive. Any one got some definitive answers on how to stop this sort of server process ramping and subsequent server failure? If you need more info let me know. Cheers Andy  /var/log/apache2/error_log [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17651 graceful kill fail, sending SIGKILL [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17650 graceful kill fail, sending SIGKILL [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17649 graceful kill fail, sending SIGKILL [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17644 graceful kill fail, sending SIGKILL [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17643 graceful kill fail, sending SIGKILL [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17638 graceful kill fail, sending SIGKILL [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17633 graceful kill fail, sending SIGKILL [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17627 graceful kill fail, sending SIGKILL [Thu May 17 07:40:47 2012] [warn] mod_fcgid: process 17622 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17674 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17673 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17672 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17667 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17666 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17665 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17664 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17659 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17658 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17657 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17656 graceful kill fail, sending SIGKILL [Thu May 17 07:40:51 2012] [warn] mod_fcgid: process 17651 graceful kill fail, sending SIGKILL https://docs.google.com/a/thesugarrefinery.com/open?id=0B_XbpWChge0VRmFLWEZfR2VBb2M https://docs.google.com/a/thesugarrefinery.com/open?id=0B_XbpWChge0VWTcwZEhoV2Fqejg https://docs.google.com/a/thesugarrefinery.com/open?id=0B_XbpWChge0VUUtVWWFINHZjZ0U https://docs.google.com/a/thesugarrefinery.com/open?id=0B_XbpWChge0VZEVMclh6ZUdaOUE <IfModule mod_fcgid.c> <IfModule !mod_fastcgi.c> AddHandler fcgid-script fcg fcgi fpl </IfModule> FcgidIPCDir /var/lib/apache2/fcgid/sock FcgidProcessTableFile /var/lib/apache2/fcgid/shm FcgidIdleTimeout 40 FcgidProcessLifeTime 30 FcgidMaxProcesses 20 FcgidMaxProcessesPerClass 20 FcgidMinProcessesPerClass 0 FcgidConnectTimeout 30 FcgidIOTimeout 120 FcgidInitialEnv RAILS_ENV production FcgidIdleScanInterval 10 FcgidMaxRequestLen 1073741824 </IfModule>

    Read the article

  • Win 7: apps crash, then explorer crashes, then services fail, then boom

    - by snorfys
    Periodically, every 2-3 days one of my systems will go haywire: every app will crash search will fail via the start menu and then explorer will fail. Restarting explorer via taskmanager will cause it to fail again, then it'll BSOD and restart. The eventlog for when this happens goes something like this every time: ERROR: Session "ReadyBoot" stopped due to the following error: 0xC0000188 (supposedly not a problem) WARNING: The maximum file size for session "ReadyBoot" has been reached... (forget where I found out, but also 'not a problem') ERROR: Session "Circular Kernel Context Logger" stopped due to the following error: 0xC0000188 (again, supposedly not a problem) WARNING: The maximum file size for session "Circular Kernel Context Logger" has been reached... ERROR: Faulting application name: Explorer.EXE, version: 6.1.7600.16450, time stamp:... ERROR: Faulting application name: explorer.exe, version: 6.1.7600.16450, time stamp:... ERROR: Faulting application name: svchost.exe_iphlpsvc, version: 6.1.7600.16385, time stamp:... ERROR: The Service Name service terminated unexpectedly. It has done this 1 time(s) That last one happens a number of times but with a different service name. Then finally we have: ERROR: The Service Control Manager tried to take a corrective action (Restart the service) after the unexpected termination of the Server service, but this action failed with the following error: An instance of the service is already running. After that, I have my BSOD and logs complaining that windows started up without shutting down. It's a new machine: Intel i3 530 4gb RAM (Ran memtest for 4 hrs, no problems) 320GB WD/250GB Seagate HDDs (Happened on fresh installs on 2 separate HDDs) Win7 Pro/Ultimate x64 (wife's copy of pro, my copy of ult, no change) Fresh install + driver and windows update (happened without updates as well) I'm at a bit of a loss as to what I can look at next. Especially since it'll work like a charm for 2-3 days and then it's hooped for a night (I'm on it now in fact - no problems).

    Read the article

  • "Rien n'est sécurisé sur le site Web de l'Hadopi" déclare Eric Walter, lors d'un débat sur les cyberguerriers

    "Rien n'est sécurisé sur le site Web de l'Hadopi" déclare Eric Walter, le secrétaire général était l'invité d'un débat sur les cyberguerriers Ce 8 février 2011 avait lieu dans un endroit huppé de la capitale un débat organisé par le Cercle, un réseau de 500 professionnels de la sécurité de l'information. Autour de la table, étaient réunis Olivier Laurelli, aka Bluetouff, blogueur spécialisé dans les problématiques liées à la sécurité et aux libertés individuelles ; Pierre Zanger, psychiatre et psychanalyste ; et Eric Walter, secrétaire général de l'Hadopi. Ce dernier, habituellement très contesté dans ce genre d'exercices, fut accueilli par ces mots de la part du "Monsieur Loyal" de la soirée :"Je n'ai pas ...

    Read the article

  • Quelle firme représente le plus grand danger pour Google : Facebook ou Microsoft ? Eric Schmidt penche pour le second choix

    Quelle firme représente le plus grand danger pour Google : Facebook ou Microsoft ? Eric Schmidt penche pour le second choix Depuis quelques mois, les citations et rapports avançant que Facebook est le plus grand rival de Google ne cessent de se succéder. Mais est-ce vraiment le cas ? Pas au goût d'Eric Schmidt en tous cas, le CEO sortant de la firme de Mountain View. Selon lui, l'entreprise faisant le plus de concurrence à Google est Microsoft. Il précise que Redmond possède plus de cash, d'ingénieurs et de clients, ce qui pousse Google à "sentir la concurrence de Microsoft chaque jour". Alors que, parallèlement à cela, Facebook a clairement expliqué qu'il ne s'attaquerait pas ...

    Read the article

  • "Le secteur technologique dominé par une Bande des Quatre" pour Eric Schmidt, qui exclut Microsoft un peu dépassé

    Le secteur technologique dominé par une Bande des Quatre pour Eric Schmidt qui exclut Microsoft un peu dépassé, mais capable de faire un grand coup avec Surface Eric Schmidt, le président du conseil d'administration de Google estime que la bataille du mobile se joue actuellement entre le géant de la recherche et Apple. S'exprimant lors d'une interview avec le blog technologique AllThingsDigital, l'ancien PDG de Google a déclaré que la bataille que livre sa firme avec Apple joue actuellement le rôle central dans le secteur de l'IT. « Nous n'avons pas vu de combat de plateformes à cette échelle », souligne Schmidt. Le secteur du mobile est actuellement largement dominé par App...

    Read the article

  • Android : Pourquoi Apple n'attaque-t-il pas directement Google ? Eric Schmidt étonné que la firme vise plutôt les constructeurs

    Android : Pourquoi Apple n'attaque-t-il pas directement Google ? Eric Schmidt étonné que la firme vise plutôt les constructeurs Dans un entretien accordé au magazine Wall Street Journal concernant les relations tendues entre Google et Apple, Eric Schmidt, Président du conseil d'administration de Google, exprime son étonnement au fait qu'Apple n'attaque pas directement Google en justice pour violation de brevets. La firme à la pomme croquée depuis plusieurs années est en conflit avec les constructeurs de terminaux Android, qu'elle accuse de violer dans leurs dispositifs ses propriétés intellectuelles. Dans sa dernière plainte contre Samsung, Apple s'attaque directement à l...

    Read the article

  • Why do many software projects fail today?

    - by TomTom
    As long as there are software projects, the world is wondering why they fail so often. I would like to know if there is a list or something equivalent which shows how many software projects fail today. Would be nice if there would be a comparison over the last 20 - 30 years. You can also add your top reason why a software project fails. Mine is "Requirements are poor or not even existing." which includes also "No (real) customer / user involved". EDIT: It is nearly impossible to clearly define the term "fail". Let's say that fail means: The project was more than 10% over budget and time. In my opinion the 10% + / - is a good range for an offer / tender. EDIT: Until now (Feb 11) it seems that most posters agree that a fail of the project is basically a failure of the project management (whatever fail means). But IMHO it comes out, that most developers are not happy with this situation. Perhaps because not the manager get penalized when a project was not successful, but the lazy, incompetent developer teams? When I read the posts I can also hear-out that there is a big "gap" between the developer side and the managment side. The expectations (perhaps also the requirements) seem to be so different, that a project cannot be successful in the end (over time / budget; users are not happy; not all first-prio features implemented; too many bugs because developers were forced to implement in too short timeframes ...) I',m asking myself: How can we improve it? Or do we have the possibility to improve it? Everybody seems to be unsatisfied with the way it goes now. Can we close the gap between these two worlds? Should we (the developers) go on strike and fight for "high quality reqiurements" and "realistic / iteration based time shedules"? EDIT: Ralph Westphal and Stefan Lieser have founded a new "community" called: Clean-Code-Developer. The aim of the group is to bring more professionalism into software engineering. Independently if a developer has a degree or tons of years of experience you can be part of this movement. Clean Code Developers live principles like SOLID every day. A professional developer is the biggest reviewer of his own work. And he has an internal value system which helps him to improve and become better. Check it out on: Clean Code Developer EDIT: Our company is doing at the moment a thing called "Application Development and Maintenance Benchmarking". This is a service offered by IBM to get a feedback from someone external on your software engineering process quality etc. When we get the results, I will tell you more about it.

    Read the article

  • Reasons to fail a build

    - by Brian Laframboise
    As a build engineer, I'm constantly looking for new and interesting ways to improve our build process - and that includes looking for new and interesting ways to fail our builds! I have yet to find a canonical list of reasons to fail a build ... so I figure it's time to get one created. With that in mind: What build-time checks - both obvious and creative - have you seen fail builds?

    Read the article

  • Fail rate of EBS snapshots and AMIs?

    - by user784637
    According to Amazon the fail rate of EBS volumes is: As an example, volumes that operate with 20 GB or less of modified data since their most recent Amazon EBS snapshot can expect an annual failure rate (AFR) of between 0.1% – 0.5%, where failure refers to a complete loss of the volume http://aws.amazon.com/ebs/ However, I was curious to know the fail rate of EBS snapshots and private AMIs that a system admin would take. Since the EBS snapshots and AMIs are stored in Amazon s3, is it safe to assume that the likelihood you cannot rollback to a previous snapshot or AMI the same likelihood that a file gets lost in s3? Amazon S3’s standard storage is: ... Designed to provide 99.999999999% durability and 99.99% availability of objects over a given year. http://aws.amazon.com/s3/

    Read the article

  • How does RAM fail?

    - by ethanlee16
    I have an issue with a Dell Inspiron 15 (1545) laptop that refuses to open any applications (save select Microsoft programs, e.g. Security Essentials, Ctrl Panel, Windows Explorer (not Internet), regedit, Event Viewer, etc.). I've run Microsoft Memory Diagnostics Tool and it found a 'hardware problem was detected.' Does this indicate that the RAM has failed? I notice when I open programs like Word, Excel, Internet Explorer, etc., it always give me an error from WerFault.exe saying The instruction at xxxxxxx referenced memory at xxxxxxxxx. The memory could not be written. and sometimes something about illegal instructions. If it is a hardware problem, does this mean that replacing the RAM is my only option? Again, I would also like to know if RAM can fail (like hard drives) and if malware can cause RAM to fail also.

    Read the article

  • Existing connexion on Apache and mod_proxy_balancer don't fail over second JBoss node

    - by Jean-Rémy Revy
    I have a Jboss farm, load balanced by Apache HTTP + mod_proxy_balancer and mod_proxy_ajp, with the following configuration : <VirtualHost *:80> ServerName web-gui-acceptance.myorg.com ServerAlias web-gui-acceptance ProxyRequests Off ProxyPass /web-gui balancer://jbosscluster/web-gui stickysession=JSESSIONID nofailover=On ProxyPassReverse /web-gui http://srvlnx01.myorg.com:8080/web-gui ProxyPassReverse /web-gui http://srvlnx02.myorg.com:8080/web-gui <Proxy *> AuthType Kerberos [...] </Proxy> <Proxy balancer://jbosscluster> BalancerMember ajp://srvlnx01.myorg.com:8009 route=SRVLNX01_node1 BalancerMember ajp://srvlnx01.myorg.com:8009 route=SRVLNX02_node1 ProxySet lbmethod=byrequests </Proxy> </VirtualHost> When the first JBoss node fail (the hosting VM is down), my existing connexions don't fail over the second node ... the fist route is keeped (in table / .shm ?) and that provide me 503 errors. Can someone tell me what I missed ?

    Read the article

  • tomcat6 on ubuntu fails when user set to root

    - by J G
    I'm well aware that running tomcat6 is really bad from a security point of view - and opens the box it is running on to all kinds of security risks and attack vectors. That said: When I change the entry in the /etc/init.d/tomcat6 to TOMCAT6_USER=root and then run sudo /etc/init.d/tomcat6 start I get [fail] and nothing is written to the logs under /var/log/tomcat6 and no entry for tomcat6 is created under /var/run How do I diagnose what is going wrong?

    Read the article

  • Some URLs fail to load on Windows web portal

    - by jpolache
    I’m working in a large data center and have been assigned to troubleshoot and issue with a windows (IIS) web server that acts as a portal for a customer of the data center. This portal server is on a DMZ at the local data center. I don’t have access to the portal desktop and am relying on an off-site administrator to work with me to do testing and report the condition of the portal. He tells me there are no software firewalls or other filtering configured. While most of the remote web pages work fine, several of the URSs the portal is suppose to serve up fail to load. I had wireshark installed on the portal system and had a capture taken of one of the failures. I used IE to access one of the remote web servers at issue. I could see the TCP SYN-ACK coming back from the remote server, but after several HTTP GETs fail to get a response the portal server sends a reset. The webmaster of the remote web server assures me that no sites are being blocked. I had a capture taken outside the local firewall, so there should be no issue there. Another tech set up a laptop and used the IP address of the portal (we took the portal off-line for the test). The laptop loads the URL as expected. I tried having Firefox loaded to make sure that the HTTP GET was not mal-formed. Same failure as with IE. So, it seems it is not the remote web server or the network, because there was no problem with the laptop. At this point, I’m not sure what other questions to ask or tests to do.

    Read the article

  • SSL certificate fail apache

    - by Murkin
    Hello, I have setup a certificate on Apache server. When I access my site's https pages, I see the certificate flashing (in FireFox's url tab) and disappearing immediately. The browser stays in the same windows (https) but now certificate info is not displayed and the connection is not encrypted. Can anyone recommend on how I can even start checking what is wrong ? (Certificate fail ? Bad config ?) Thanks

    Read the article

  • Plink/Putty fail on return

    - by jasemccarty
    Any one have any idea why a remote command to a NetApp filer would fail when using Plink/Putty (to execute a command) across 2 networks? Everything is open both ways (multiple firewalls), but the filer seems to be responding from port 22 to some random port. So my Windows box on the other side doesn't know to answer because the target port isn't 22. Any thoughts?

    Read the article

  • Will Windows fail activation on a new hard drive after previous hard drive failed

    - by ServerBloke
    I have a failing hard drive which won't boot, that has Windows 7 Home Premium installed. I have a replacement hard drive on the way. My question is will I run into problems trying to install Windows 7 using the same cdkey and DVD on the new hard drive? I assume activation will find that the cdkey has already been activated and fail, especially if a hardware ID is checked which will probably be different because even though the other hardware is the same, the hard drive will be different.

    Read the article

< Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >