Search Results

Search found 44763 results on 1791 pages for 'first responder'.

Page 548/1791 | < Previous Page | 544 545 546 547 548 549 550 551 552 553 554 555  | Next Page >

  • Why 11.10 says plymouth errors and "Checking battery state..."?

    - by Aleksi Kinnunen
    I've just installed new Ubuntu 11.10. The installation was successful. But when I start the Ubuntu, the splash screen does not work, and I get error (always some of these) "mountall: Plymouth command failed", "mountall: Connect to Plymouth was lost" or "Checking battery state..." I'm sorry, if the error messages are incorrect because I translated two first of them from Finnish (my system is in Finnish) to English, so they maybe couldn't be just right. I have already asked from our Finnish community, but they couldn't answer. Could anybody here?

    Read the article

  • Problem after upgrading to 14.04: Unable to mount - Not authorized

    - by Elisa Velasco Lorenzo
    This is the first time I ever post on this forum. I have used Ubuntu for 7 years, and every time I've had a problem, I have browsed the web and I have been able to solve it thanks to the help given on forums to other users experiencing the same issues (I do not have technical knowledge on Ubuntu). This time, I've spent hours reading forums and typing commands and nothing helps. I am exhausted. Upgrading to 14.04 has proved itself a mistake. Many things won't work anymore. But the one which worries me right now is external media not automounting anymore. I know other people have had the same issue and I have carefully read their threads, but nothing helps!!

    Read the article

  • How can I have my VPN connect automatically when the wireless connects?

    - by ams
    I have a working VPN connection using NetworkManager, OpenConnect, and the network-manager-openconnect-gnome package, but I have to start it manually every time I connect to a network, and I have to enter my password manually each time. How can I get it to connect automatically, and remember my password (securely)? I have checked the 'Connect Automatically' box on the Configure VPN page, but this seems to have no effect. I've also got the 'Start connecting automatically' box checked in the pop-up box, and that does avoid the need to press the connect button in that window, but seems to have no part in kicking off the whole process in the first place. There is no option to remember the password in the window, but maybe there's one somewhere else?

    Read the article

  • Laptop not syncing with Ubuntu One

    - by Lolwhites
    I have a desktop running Maverick, and a laptop running Lucid. Both are supposed to be linked to my Ubuntu One account. The desktop syncs fine, but when I started the laptop for the first time in a couple of months, it wouldn't sync any more. The Ubuntu One Preferences window either reports "Synchronisation complete" even though no recent files have been downloaded nor have any test files created in the relevant synced folder been uploaded, or it says "Synchronisation in progress", which does not appear to be happening as it stays like this for ages and the lights on my router suggest no traffic is going through. Have repeatedly tried disconnecting and reconnecting, and removing the device from the account then reattaching it, all to no avail.

    Read the article

  • How to connect to the internet reliably

    - by user12689
    I have a DSL internet connection. For a long time the internet functioned fine in Ubuntu using the default Network Manager. Recently however, my ISP was bought by another ISP, and the service name had to be changed in the DSL settings. In Windows, it connects without any issue whatsoever. But in Ubuntu, the first time I connect I get a "Apache: Not Found" page, the next time, I get to the previous ISP's login page, it continues like this for some time until I get it working. I have to try connecting it again and again. When I asked the customer service, they said that because of the takeover, they have switched servers. But they said, it should work properly in Windows, and they were clueless about Ubuntu. Is there anything I can do to solve this issue?

    Read the article

  • Cloud Computing Forces Better Design Practices

    - by Herve Roggero
    Is cloud computing simply different than on premise development, or is cloud computing actually forcing you to create better applications than you normally would? In other words, is cloud computing merely imposing different design principles, or forcing better design principles?  A little while back I got into a discussion with a developer in which I was arguing that cloud computing, and specifically Windows Azure in his case, was forcing developers to adopt better design principles. His opinion was that cloud computing was not yielding better systems; just different systems. In this blog, I will argue that cloud computing does force developers to use better design practices, and hence better applications. So the first thing to define, of course, is the word “better”, in the context of application development. Looking at a few definitions online, better means “superior quality”. As it relates to this discussion then, I stipulate that cloud computing can yield higher quality applications in terms of scalability, everything else being equal. Before going further I need to also outline the difference between performance and scalability. Performance and scalability are two related concepts, but they don’t mean the same thing. Scalability is the measure of system performance given various loads. So when developers design for performance, they usually give higher priority to a given load and tend to optimize for the given load. When developers design for scalability, the actual performance at a given load is not as important; the ability to ensure reasonable performance regardless of the load becomes the objective. This can lead to very different design choices. For example, if your objective is to obtains the fastest response time possible for a service you are building, you may choose the implement a TCP connection that never closes until the client chooses to close the connection (in other words, a tightly coupled service from a connectivity standpoint), and on which a connection session is established for faster processing on the next request (like SQL Server or other database systems for example). If you objective is to scale, you may implement a service that answers to requests without keeping session state, so that server resources are released as quickly as possible, like a REST service for example. This alternate design would likely have a slower response time than the TCP service for any given load, but would continue to function at very large loads because of its inherently loosely coupled design. An example of a REST service is the NO-SQL implementation in the Microsoft cloud called Azure Tables. Now, back to cloud computing… Cloud computing is designed to help you scale your applications, specifically when you use Platform as a Service (PaaS) offerings. However it’s not automatic. You can design a tightly-coupled TCP service as discussed above, and as you can imagine, it probably won’t scale even if you place the service in the cloud because it isn’t using a connection pattern that will allow it to scale [note: I am not implying that all TCP systems do not scale; I am just illustrating the scalability concepts with an imaginary TCP service that isn’t designed to scale for the purpose of this discussion]. The other service, using REST, will have a better chance to scale because, by design, it minimizes resource consumption for individual requests and doesn’t tie a client connection to a specific endpoint (which means you can easily deploy this service to hundreds of machines without much trouble, as long as your pockets are deep enough). The TCP and REST services discussed above are both valid designs; the TCP service is faster and the REST service scales better. So is it fair to say that one service is fundamentally better than the other? No; not unless you need to scale. And if you don’t need to scale, then you don’t need the cloud in the first place. However, it is interesting to note that if you do need to scale, then a loosely coupled system becomes a better design because it can almost always scale better than a tightly-coupled system. And because most applications grow overtime, with an increasing user base, new functional requirements, increased data and so forth, most applications eventually do need to scale. So in my humble opinion, I conclude that a loosely coupled system is not just different than a tightly coupled system; it is a better design, because it will stand the test of time. And in my book, if a system stands the test of time better than another, it is of superior quality. Because cloud computing demands loosely coupled systems so that its underlying service architecture can be leveraged, developers ultimately have no choice but to design loosely coupled systems for the cloud. And because loosely coupled systems are better… … the cloud forces better design practices. My 2 cents.

    Read the article

  • What are good JS libraries for game dev? (HTML5)

    - by acidzombie24
    If i decide to write a simple game both text and graphical (2d) what libs may i use? Assuming we are using a HTML5 compatible browser. The main things i can think of Rendering text on screen Animating sprites (using images/css) Input (capturing the arrow keys and getting relative mouse positions) Perhaps some preloading resource or dynamically loading resources and choosing order Sound (but i am unsure how important this will be to me at first). Perhaps with mixing and chaining sounds or looping forever until stop. Networking (low priority) to connect a user to another or to continuously GET data without multiple request (i know this exist but i dont know how easy it is to setup or use. But this isnt important to me. Its for the question).

    Read the article

  • File access forbidden in htpasswd

    - by Nerd-Herd
    I have been using the htpasswd generated in this question and it seemed to have been working well until recently. Since yesterday, I am not able to access the newest file created in the folder ChatLogs(named 10_07_2012.txt). The server returns a 403 Forbidden error saying: Forbidden You don't have permission to access /ChatLogs/2012/07/10_07_2012.txt on this server. I am still able to access older files(until 09 July, 2012). At first I thought it might be because of file permissions, but they are the same as on other 9 files in the folder. What could be the problem? Please Help.

    Read the article

  • I want to record a screencast of a processing sketch

    - by nathanvda
    I have a created a music visualisation using Processing. I now want to convert that to a video, and the least obtrusive way I could think of is to record a screencast. I figured exporting Processing to video including audio, from within Processing itself, on ubuntu seemed an unsolved issue. Very hard and also could cause timing sync issues (since the music keeps running while images are captured). So move on to the screencast method. Dead-easy, I figured. But I was wrong. First hurdle was to find a way to record the sound from the audio (and not the mic). I did find a tutorial for that here. In short: use gtk-recordmydesktop and pulse audio. But, apparently, what happens: Processing does not use ALSA. When the sound is playing, it does not appear in the Pulse Audio mixer. How can I record the audio now?

    Read the article

  • making "xwacom set" changes permanent

    - by Philippe
    On my Thinkpad X220 Tablet the touch-finger works flawlessly but the pen is terribly miscalibrated. The calibration tool in System settings - Wacom tablet does not work. Instead, whenever I wan't to use the pen I first need to sudo xsetwacom set 'Wacom ISDv4 E6 Pen stylus' Area 0 0 27760 15690 These changes do not remain permanent. That is, after every reboot they are gone. How can I make the change permanent - I'm not looking for a startup scrip, I'd like to set the right area once for all. Any idea?

    Read the article

  • LinuxCon North America 2014

    - by Chris Kawalek
    As the first day of LinuxCon North America 2014 draws to a close, we want to thank all the folks that stopped by our booth today! If you're at the show, please stop by our booth #204 and have a chat with our experts. And you won't want to miss these Linux and virtualization related sessions coming up tomorrow and Friday: Thursday Aug 21 - Static Analysis in the Linux Kernel Using Smatch - Dan Carpenter, Oracle Friday Aug 22 - The Proper Care of Feeding of MySQL Database for Linux Admins Who Also Have DBA duties - Morgan Tocker, Oracle Friday Aug 22 - Why Use Xen for Large Scale Enterprise Deployments, Konrad Rzeszutek Wilk , Oracle We look forward to meeting with you and we hope you enjoy the rest of LinuxCon North America 2014! 

    Read the article

  • How to disable integrated webcam and still be able to use an external one

    - by alikasundara
    I am looking for a way to disable the webcam that is integrated into my laptop. The webcam is using uvcvideo module but I do not want to blacklist it since it is also being used by an external webcam I have. Is there any way of disabling the device itself without touching the modules list? This is how the webcams are listed by lsusb. The first one is an integrated one (It is identified by some apps as BisonCam NB Pro), the second one is the external Logitech C525: Bus 002 Device 004: ID 5986:0361 Acer, Inc Bus 003 Device 002: ID 046d:0826 Logitech, Inc. I have already checked BIOS - there is no way of disabling the webcam from there. Besides I would love to learn how to disable the device by ID anyway. Thanks!

    Read the article

  • Google Analytics goal funnel does not recognize virtual page views

    - by Webber Smith
    I have a setup wizard with 3 steps. Since I'm using AJAX each step uses a virtual pageview with an appropriate URL for each step (see below). The pageviews are being recorded in the Content section of Google Analytics but the Goal Funnel still shows zero for each step. I've tried advise on other forums such as... Make sure Goal URL is set to Exact match Make sure no steps or the Goal URL are a parent directory of any other steps. For example, don't track /wizard/ as a Goal/step and track /wizard/step2/. Not sure why this would be a problem since it is an exact match, but it shouldn't hurt so I tried it... Require (or don't require - tried both) the first step in the funnel ...but none of these seem to work. Thoughts? Goal Settings Exact match : "/wizard/setup-complete/" Funnel Step 1 : "/wizard/step1/" Step 2 : "/wizard/step2/" Step 3 : "/wizard/step3/"

    Read the article

  • Broken package ubuntuone-control-panel-qt

    - by baale7
    An interrupted installation of updates resulted in ubuntuone-control-panel-qt breaking and now I cannot install or update anything. There seems to be an error in python2.7 - ImportError: No module named _sysconfigdata_nd dpkg: error processing archive ubuntuone-control-panel-qt_13.09-0ubuntu1_all.deb (--install): When I try to reinstall the package via .deb a dpkg --audit yields: The following packages are in a mess due to serious problems during installation. They must be reinstalled for them (and any packages that depend on them) to function properly: ?python-pexpect Python module for automating interactive applications ?ubuntuone-control-panel-qt Ubuntu One Control Panel - Qt frontend ?hplip-data HP Linux Printing and Imaging - data files ?python-libxml2 Python bindings for the GNOME XML library ?apport automatically generate crash reports for debugging The following packages are only half configured, probably due to problems configuring them the first time. The configuration should be retried using dpkg --configure or the configure menu option in dselect: ?python-ubuntuone-client Ubuntu One client Python libraries ?python-ubuntuone-control-panel Ubuntu One Control Panel - Python Libraries --configure or --configure -a doesn't work Any help?

    Read the article

  • View space lighting in deferred shading

    - by kochol
    I implemented a simple deferred shading renderer. I use 3 G-Buffer for storing position (R32F), normal (G16R16F) and albedo (ARGB8). I use sphere map algorithm to store normals in world space. Currently I use inverse of view * projection matrix to calculate the position of each pixel from stored depth value. First I want to avoid per pixel matrix multiplication for calculating the position. Is there another way to store and calculate position in G-Buffer without the need of matrix multiplication Store the normal in view space Every lighting in my engine is in world space and I want do the lighting in view space to speed up my lighting pass. I want an optimized lighting pass for my deferred engine.

    Read the article

  • Why are 3D game scenarios so big? and.. are them shareable? [closed]

    - by Néstor Sánchez A.
    First, I apologize if this is not the forum for these type of questions. I was wondering... 1- Why 3D games (such as GTA IV) uses so much gigabytes of space? And my half-answar was... because of their huge 3D scenarios. But wait! are not these scenarios made with vector graphics and textures? then.. why are so big??? 2- If they are so big... would really these companies make a whole new scenario for every new game they develop? that would be very inefficient. Doesn't exist "shareable" scenarios that can be used in different games and expanded to no rebuild a whole one (like "New York" or "Liberty City")???. I mean.. the scenarios is almost the same in terms of vertices, just with improved textures.

    Read the article

  • I cant acess my HDD

    - by user286283
    First of all, I'm a total Linux noob. I was running the latest version of Ubuntu but eventually I wasn't able to boot no more, the Ubuntu logo appeared and then it was only a blinking white bar indefinitely, without charging. Now, my problem is the following. I want to do a fresh install and I am trying to recover my latest files from the HDD using the Ubuntu live CD. I click on the HDD and obtain the following I honestly don't really know whats going on. Looking at disks I have this... Can anyone help me out? Cheers.

    Read the article

  • Wordpress : Automatically transfer media files to Amazon S3

    - by Ron Ranieri
    I've been using VPS to host 7 Wordpress websites, most of them require big storage but very little RAM and traffic. So I'm thinking of moving the static files(uploads folder) content to Amazon S3 and I'm looking for the most viable solution to this. I want every website to have their own bucket and newly uploaded media files automatically uploaded to Amazon S3 without using plugin. I'm ok with cron job, for example the files were uploaded first to my server, then transferred to S3 and deleted from my server every 24 hour. Or is there any way for me to change the default upload directory to my S3 bucket without sacrificing any Wordpress functionality(resize/title etc)? What do you think the most efficient way to do this? Currently I'm looking at this plus cron job but I would like to know better option if it exist.

    Read the article

  • pulseaudio: no microphone configuration

    - by mitsch
    Updated Ubuntu to precise from oneiric on a Dell Inspirion Mini 10 (with an Intel HDA-Soundcard). I can't remember having any issues with the microphone. I didn't need it - I tried ekiga the first time in precise. I couldn't hear any sound in the echo test of ekiga, so I switched to the "System Preferences" and looked for the microphone to boost it. Surprisingly, the microphone input was greyed out - I couldn't mute or unmute, i couldn't even move the volume-slider. On the other hand: I could change the microphone-setting on console with alsamixer, so don't worry about that… :) I just wanted to ask, how to get pulseaudio back to the known, comfortable behaviour. Some newbies won't know the trick to use alsamixer… My soundcard (output of lspci): 00:1b.0 Audio device: Intel Corporation N10/ICH 7 Family High Definition Audio Controller (rev 02) Greets!

    Read the article

  • About cdn architecture to route way

    - by Tony Lee
    Our web system, use the third-party cdn service. Assume that the user set the local dns with the googledns or opendns to visit our web sites, so cdn service will select the closest cdn proxy node. all right, but in fact the user's actual access position might outside there, cdn service may chose the one furthest away from the user node, so static resource access slower.. At present, my idea is if user local set dns server with googledns, and then first one we get the actual ip address of the user, tracerote to test a best routing lines, set up a cookie in user browser, and then set 302 header for response to jump to the which best cdn node. Whether the user's browser side traceroute tool can provide the best route decision-making ? Because we find that, once the user to set local dns server with the foreign network segment, for example : set dns with 8.8.8.8, so cdn routing will choose the foreign service node.

    Read the article

  • Reinstall Ubuntu on Custom Partitions

    - by Forerunner117
    I am attempting to reinstall ubuntu 13.04 without losing my installed software and /home docs. I have read countless threads on this same topic, but nothing seems to apply to my situation. When I originally installed, I had created a separate partition for /home, but I am now unsure of which partition that was. Based on the picture below, where should I be installing the new copy? Also, will I run into problems since I am now running 13.10 and want to put 13.04 back on it? Should I grab 12.04 or 13.10 for this reinstall? Picture: http://i.stack.imgur.com/FL2SY.png (Note: I am performing this reinstall due to a complete muck up of my unity/compiz settings and configuration, resulting in no desktop. I've done my best to resolve this problem first before resorting to this.)

    Read the article

  • Hosting a website from a dynamic IP

    - by nick
    I recently upgraded my internet to the point that it is much faster and more reliable than my current webhost. I would like to move my current domain to be hosted at home, but my IP address is dynamic. As far as I know, I only get a new IP when I restart my modem and or router (which is almost never) or when cable one (my ISP) pushes out a firmware update (rarely). There are a few ways I can see doing this 1) convince my ISP to give me a static IP 2) assign my router my current IP to force a static IP (which might work?) 3) set my dns record to my current IP address and update it on the rare occasions that it changes. Obviously I'm hoping that the first one works, but I don't want to pay a lot of extra money (if that's what it takes) to get a static IP address. Has anyone had any luck with something like that?

    Read the article

  • Displaying a Sorted, Paged, and Filtered Grid of Data in ASP.NET MVC

    Over the past couple of months I've authored five articles on displaying a grid of data in an ASP.NET MVC application. The first article in the series focused on simply displaying data. This was followed by articles showing how to sort, page, and filter a grid of data. We then examined how to both sort and page a single grid of data. This article looks at how to add the final piece to the puzzle: we'll see how to combine sorting, paging and filtering when displaying data in a single grid. Like with its predecessors, this article offers step-by-step instructions and includes a complete, working demo available for download at the end of the article. Read on to learn more! Read More >

    Read the article

  • The Gates Books&ndash;Finished

    - by MarkPearl
    Today I can finally say that I finished both of my Bill Gates books that had been lying on the shelf for several years. The books were… The Road Ahead (1995) Business @ the Speed of Thought (1999) I enjoyed “The Road Ahead”, purely because it was fun to read about someone looking into the future at technology, while I could read it looking at the past. In fact I was quite impressed with how much he got right and it was also nice to remember “The good old days”. Business @ the Speed of Thought was a harder read for me. The book still had some good insights, but was tough going at times (possibly because it was several hundred more pages than the first). All that being said, I can now finally place them back on the shelf knowing that they have been read.

    Read the article

  • Skype on Ubuntu 12.04 x64 - anyone has it installed?

    - by Zevaka
    I am wondering if there's anyone here who managed to use skype @ 12.04. It used to work (ever since i first tried Ubuntu 10.04) flawlessly, but after upgrade to 12.04 just doesn't start. If I start skype from terminal, it just says "Segmentation fault (core dumped). I tried to remove skype and installed it from different locations: Skype website (did not install at all, "package error") From official Ubuntu repository (doesn't start at all, segmentation fault) From repository "deb http://archive.canonical.com/ $(lsb_release -sc) partner" (doesn't start at all, segmentation fault) Interesting thing: even if I remove / purge skype from the system, it still shows up in Unity menu (of course doesn't start). Any help here? Thanks!

    Read the article

< Previous Page | 544 545 546 547 548 549 550 551 552 553 554 555  | Next Page >