Search Results

Search found 14570 results on 583 pages for 'midi virtual keyboard'.

Page 312/583 | < Previous Page | 308 309 310 311 312 313 314 315 316 317 318 319  | Next Page >

  • How do I record audio through M-Audio Keystudio?

    - by interstar
    Hi, I'm trying to get my M-Audio Keystudio (which has an audio input as well as the keyboard) to record audio to Audacity. I'm in Ubuntu 10.10. When I look at the Sound Preferences I can select "M-Audio RunTime DFU Analog Stereo" as my input device. However, when I try to record in Audacity, Audacity remains frozen. The program seems to be running and recording, but the recording cursor won't advance. If I reset the audio input to the internal sound card, recording works normally. Any ideas what to look for?

    Read the article

  • Ubuntu 14.04 Bluetooth Magic Mouse doesn't pair (No agent available)

    - by Rafael Xavier
    Mouse gets discovered. Although, it doesn't pair. /var/log/syslog: Apr 23 10:05:15 xavier bluetoothd[9873]: No agent available for request type 0 Apr 23 10:05:15 xavier bluetoothd[9873]: btd_event_request_pin: Operation not permitted Apr 23 10:05:15 xavier bluetoothd[9873]: Connection refused (111) It's worth saying that: Keyboard has paired and it's working just fine though; Mouse used to work just fine in Ubuntu 12.04, and 13, and it works when I reboot on Mac; This is the hci device. $ hcitool dev Devices: hci0 E0:F8:47:3A:3F:47 How to get it working?

    Read the article

  • How can I disable Hibernate completely?

    - by Lekensteyn
    I have seen the answer on How to disable hibernating?, but I have no such file. Possibly because that suggestion was written for Ubuntu, not Kubuntu (KDE, not Gnome). I do not have a swap on my encrypted SSD, my system freezes (cannot even change Caps Lock) if I accidentally press the "Hibernate" button at "Energy management". My keyboard has a Hibernate button (Fn + F4) next to the volume control buttons and every time I press the wrong key, the system will freeze after. So, what is the correct way to disable it? If there is no solution, a work-around is welcome too.

    Read the article

  • (Xubuntu) no touchpad or mouse on Aspire One D270

    - by Sman789
    I installed Xubuntu 12.10 on an Aspire One D270 using a Live USB. Initially, the live USB wouldn't boot, but by running sudo apt-get dist-upgrade on it (on a different machine), I upgraded the kernel and it now works perfectly. However, after installing Xubuntu from the stick onto the D270, neither the touchpad nor an external mouse works. If I use the keyboard to navigate to the mouse settings, the 'device' setting is greyed out. I don't see why the mouse should work perfectly when running from the live USB but not once installed to the machine...? Thanks for any help.

    Read the article

  • Build 2012, some thoughts..

    - by Dennis Vroegop
    I think you probably read my rant about the logistics at Build 2012, as posted here, so I am not going into that anymore. Instead, let’s look at the content. (BTW If you did read that post and want some more info then read Nia Angelina’s post about Build. I have nothing to add to that.) As usual, there were good speakers and some speakers who could benefit from some speaker training. I find it hard to understand why Microsoft allows certain people on stage, people who speak English with such strong accents it’s hard for people, especially from abroad, to understand. Some basic training might be useful for some of them. However, it is nice to see that most speakers are project managers, program managers or even devs on the teams that build the stuff they talk about: there was a lot of knowledge on stage! And that means when you ask questions you get very relevant information. I realize I am not the average audience member here, I am regular speaker myself so I tend to look for other things when I am in a room than most audience members so my opinion might differ from others. All in all the knowledge of the speakers was above average but the presentation skills were most of the times below what I would describe as adequate. But let us look at the contents. Since the official name of the conference is Build Windows 2012 it is not surprising most of the talks were focused on building Windows 8 apps. Next to that, there was a lot of focus on Azure and of course Windows Phone 8 that launched the day before Build started. Most sessions dealt with C# and JavaScript although I did see a tendency to use C++ more. Touch. Well, that was the focus on a lot of sessions, that goes without saying. Microsoft is really betting on Touch these days and being a Touch oriented developer I can only applaud this. The term NUI is getting a bit outdated but the principles behind it certainly aren’t. The sessions did cover quite a lot on how to make your applications easy to use and easy to understand. However, not all is touch nowadays; still the majority of people use keyboard and mouse to interact with their machines (or, as I do, use keyboard, mouse AND touch at the same time). Microsoft understands this and has spend some serious thoughts on this as well. It was all about making your apps run everywhere on all sorts of devices and in all sorts of scenarios. I have seen a couple of sessions focusing on the portable class library and on sharing code between Windows 8 and Windows Phone 8. You get the feeling Microsoft is enabling us devs to write software that will be ubiquitous. They want your stuff to be all over the place and they do anything they can to help. To achieve that goal they provide us with brilliant SDK’s, great tooling, a very, very good backend in the form of Windows Azure (I was particularly impressed by the Mobility part of Azure) and some fantastic hardware. And speaking of hardware: the partners such as Acer, Lenovo and Dell are making hardware that puts Apple to a shame nowadays. To illustrate: in Bellevue (very close to Redmond where Microsoft HQ is) they have the Microsoft Store located very close to the Apple Store, so it’s easy to compare devices. And I have to say: the Microsoft offerings are much, much more appealing that what the Cupertino guys have to offer. That was very visible by the number of people visiting the stores: even on the day that Apple launched the iPad Mini there were more people in the Microsoft store than in the Apple store. So, the future looks like it’s going to be fun. Great hardware (did I mention the Nokia Lumia 920? No? It’s brilliant), great software (Windows 8 is in a league of its own), the best dev tools (Visual Studio 2012 is still the champion here) and a fantastic backend (Azure.. need I say more?). It’s up to us devs to fill up the stores with applications that matches this. To summarize: it is great to be a Windows developer. PS. Did I mention Surface RT? Man….. People were drooling all over it wherever I went. It is fantastic :-) Technorati Tags: Build,Windows 8,Windows Phone,Lumia,Surface,Microsoft

    Read the article

  • Visual Studio 2010 Short Cut Links!

    - by Dave Noderer
    This week Scott Cate came to South Florida and gave a great talk on his Visual Studio shortcuts and how he uses them. You can find a collection of short video’s he has done at: http://scottcate.com/tricks/ Also you might want to check out Sara Ford’s blog: http://blogs.msdn.com/saraford/default.aspx, she started doing a tip a day but has many more now. Scott covers many of these in the videos. And.. as with past releases, the languages team has provided PDF’s with a lot of keyboard shortcuts, this time for VB, C#, F# and C++. You can find downloads for all of these at the top of the FlaDotNet.com page and are included below: VB: http://www.fladotnet.com/downloads/VS2010VB.pdf C#: http://www.fladotnet.com/downloads/VS2010CSharp.pdf F#: http://www.fladotnet.com/downloads/VS2010FSharp.pdf C++: http://www.fladotnet.com/downloads/VS2010CPP.pdf Happy Keyboarding!!

    Read the article

  • Static Application Switcher is empty after show dekstop

    - by Jepzen
    Im using the Static Application Switcher to configure my ALT + TAB to cycle trough all applications not grouped. So far so good. Then I press Super + D to "hide all normal windows" (show desktop). This setting is from Keyboard - Shortcuts - Navigation. This also works BUT After hide all normal windows has been activated and I press ALT + TAB nothing comes up. I suspect they have been hidden so well that the Application Switcher does not longer sees them. Anyone know if this can be fixed trough settings somewhere?

    Read the article

  • Push or Pull Input Data In the Game Logic?

    - by Qua
    In the process of preparing my game for networking I'm adding a layer of seperation between the physical input (mouse/keyboard) and the actual game "engine"/logic. All input that has any relation to the game logic is wrapped inside action objects such as BuildBuildingAction. I was thinking of having an action processing layer that would determine what to do with the input. This layer could then be set up to either just pass the actions locally to the game engine or send it via sockets to the network server depending on whether the game was single- or multiplayer. In network games it would make sense that the player's actions should be sent to the server, but should the game logic be pulling (polling?) the data through some sort of interface or should the action processing layer be adding the actions to an input queue in the game logic code?

    Read the article

  • How to Run files and open folders as an administrater 11.10

    - by Mattlinux1
    Solved!!! Enable open as administrator in nautilus! To get started, press Ctrl – Alt – T on your keyboard to open Terminal. When it opens, run the commands below: enter code here sudo apt-get install nautilus-gksu After installing that application run copy and paste the line below and press enter. sudo cp /usr/lib/nautilus/extensions-2.0/libnautilus-gksu.so /usr/lib/nautilus/extensions-3.0/ Finally, log out and back in then go and test it, by clicking your right mouse click on the file you want to run as Admin and your see a popup menu with the words: open as administrater. Visit: http://www.liberiangeek.net/2011/12/add-open-as-administrator-to-nautilus-context-menu-in-ubuntu-11-10-oneiric-ocelot/ for help and from here. Enjoy!

    Read the article

  • Why did ctrl-f stop working in LibreOffice?

    - by amandabee
    I've basically settled into Unity, but one constant that is kind of giving me agita is that in LibreOffice Calc (and Writer) ctrl-f does nothing. I can go to edit > find ... and the edit menu even suggests ctrl-f as a keyboard shortcut, but actually typing it doesn't do anything. ctrl-alt-f works fine. The only thing that doesn't work is ctrl-f. I had a really hard time getting settled with Unity (mostly trying to get my old Gnome-do and Compiz settings back) so I spent a lot of time monkeying with the Compiz Settings Manager. But I can't figure out where I would have unset/ overrode this. Plus, it isn't like ctrl-f does something unexpected. It just does nothing.

    Read the article

  • Geek Deal: Refurbished Kindle Fire for $139; Today Only

    - by Jason Fitzpatrick
    If you’re looking to pick up a Kindle Fire on the cheap, Amazon is offering them–refurbished with a 1-year warranty–for $139. $139 is an even better price than we see on our local Craiglist (where Kindle Fires usually go for $180 or so) and it comes with a 1-year warranty. We’ve purchased several Kindle Keyboard units through Amazon’s refurbished warehouse deals over the last two years and, frankly, we can’t tell them apart from the brand new ones–if you’re looking to pick up a Kindle Fire this is a great deal. Kindle Fire for $139 How to Own Your Own Website (Even If You Can’t Build One) Pt 1 What’s the Difference Between Sleep and Hibernate in Windows? Screenshot Tour: XBMC 11 Eden Rocks Improved iOS Support, AirPlay, and Even a Custom XBMC OS

    Read the article

  • .NET development on a Retina MacBook Pro with Windows 8

    - by Jeff
    I remember sitting in Building 5 at Microsoft with some of my coworkers, when one of them came in with a shiny new 11” MacBook Air. It was nearly two years ago, and we found it pretty odd that the OEM’s building Windows machines sucked at industrial design in a way that defied logic. While Dell and HP were in a race to the bottom building commodity crap, Apple was staying out of the low-end market completely, and focusing on better design. In the process, they managed to build machines people actually wanted, and maintain an insanely high margin in the process. I stopped buying the commodity crap and custom builds in 2006, when Apple went Intel. As a .NET guy, I was still in it for Microsoft’s stack of development tools, which I found awesome, but had back to back crappy laptops from HP and Dell. After that original 15” MacBook Pro, I also had a Mac Pro tower (that I sold after three years for $1,500!), a 27” iMac, and my favorite, a 17” MacBook Pro (the unibody style) with an SSD added from OWC. The 17” was a little much to carry around because it was heavy, but it sure was nice getting as much as eight hours of battery life, and the screen was amazing. When the rumors started about a 15” model with a “retina” screen inspired by the Air, I made up my mind I wanted one, and ordered it the day it came out. I sold my 17”, after three years, for $750 to a friend who is really enjoying it. I got the base model with the upgrade to 16 gigs of RAM. It feels solid for being so thin, and if you’ve used the third generation iPad or the newer iPhone, you’ll be just as thrilled with the screen resolution. I’m typically getting just over six hours of battery life while running a VM, but Parallels 8 allegedly makes some power improvements, so we’ll see what happens. (It was just released today.) The nice thing about VM’s are that you can run more than one at a time. Primarily I run the Windows 8 VM with four cores (the laptop is quad-core, but has 8 logical cores due to hyperthreading or whatever Intel calls it) and 8 gigs of RAM. I also have a Windows Server 2008 R2 VM I spin up when I need to test stuff in a “real” server environment, and I give it two cores and 4 gigs of RAM. The Windows 8 VM spins up in about 8 seconds. Visual Studio 2012 takes a few more seconds, but count part of that as the “ReSharper tax” as it does its startup magic. The real beauty, the thing I looked most forward to, is that beautifully crisp C# text. Consolas has never looked as good as it does at 10pt. as it does on this display. You know how it looks great at 80pt. when conference speakers demo stuff on a projector? Think that sharpness, only tiny. It’s just gorgeous. Beyond that, everything is just so responsive and fast. Builds of large projects happen in seconds, hundreds of unit tests run in seconds… you just don’t spend a lot of time waiting for stuff. It’s kind of painful to go back to my 27” iMac (which would be better if I put an SSD in it before its third birthday). Are there negatives? A few minor issues, yes. As is the case with OS X, not everything scales right. You’ll see some weirdness at times with splash screens and icons and such. Chrome’s text rendering (in Windows) is apparently not aware of how to deal with higher DPI’s, so text is fuzzy (the OS X version is super sharp, however). You’ll also have to do some fiddling with keyboard settings to use the Windows 8 keyboard shortcuts. Overall, it’s as close to a no-compromise development experience as I’ve ever had. I’m not even going to bother with Boot Camp because the VM route already exceeds my expectations. You definitely get what you pay for. If this one also lasts three years and I can turn around and sell it, it’s worth it for something I use every day.

    Read the article

  • I need acpi=off to boot, but this seems to disable wifi

    - by TheExp3rt
    To boot my computer, I have to use acpi=off in kernel options. Doing this apparently causes my wifi adapter to be switched off and prevents my mouse and USB keyboard to not work. If I change acpi=off to nomodeset, it fixes these problems. But then it sets my computer at a lower resolution. I am currently running Kubuntu 13.10, although none of the updates, starting from 12.04, changed the need for acpi=off. I assume I need acpi=off in because I have an oldish computer with an intel motherboard How can I boot without acpi=off? Output of uname -a: 3.11.0-13-generic #20-Ubuntu SMP Wed Oct 23 07:38:26 UTC x86_64 x86_64 x86_64 GNU/Linux

    Read the article

  • Become a Vi Master by Learning These 30+ Key Bindings

    - by Chris Hoffman
    Vi is a powerful text editor included on most Linux systems. Many people swear by vi and find it faster than any other editor once they’ve learned its key bindings. You can even use vi key bindings in Bash. We’ve already covered getting started with vi for beginners. If you haven’t used vi in a while, you might want to give that post a look to get a refresher on the basics. How To Customize Your Wallpaper with Google Image Searches, RSS Feeds, and More 47 Keyboard Shortcuts That Work in All Web Browsers How To Hide Passwords in an Encrypted Drive Even the FBI Can’t Get Into

    Read the article

  • OpenGL programming vs Blender Software, which is better for custom video creation?

    - by iammilind
    I am learning OpenGL API bit by bit and also develop my own C++ framework library for effectively using them. Recently came across Blender software which is used for graphics creation and is in turn written in OpenGL itself. For my part time hobby of graphics learning, I want to just create small-small movie or video segments; e.g. related to construction engineering, epic stories and so on. There may be very minimal to nil mouse-keyboard interaction for those videos, unlike video games which are highly interactive. I was wondering if learning OpenGL from scratch is worth for it or should I invest my time in learning Blender software? There are quite a few good movie examples are created using Blender and are shown in its website. Other such opensource cross platform alternatives are also welcome, which can serve my aforementioned purpose.

    Read the article

  • Is it possible to get the old workspaceswitcher back in the top-panel?

    - by Arne
    I just updated my Ubuntu from 10.04 to 12.04 yesterday and so far I'm very satisfied with it. There is only one thing which is still missing to get back just the same comfort as I had with the previous version. I work on 4 workspaces and would like to have the ability to switch from one to another by just one click. With Gnome there was the possibility to add the workspace switcher to the top-panel next to the system tray and I could easily click on just the workspace I wanted to switch to. Of course there I could switch by using the keyboard shortcuts, but I would like to use the mouse in addition. I don't like using the switcher on the left side because it's much slower. Is it possible to get the old workplace switcher back in the panel at the top?

    Read the article

  • use ubuntu notebook as second monitor for fedora desktop

    - by YumTum
    I was wondering if there is a possibility to use my netbook with ubuntu 12.04 as a second monitor for my fedora desktop, because its keyboard is broken but everything else is running just fine. I already read about x2x, but I hope there is a more direct way to do this (I mean not over the WLAN but e.g. via LAN etc... ). Would it be possible to directly use the monitor without even running the laptop? For example through an VGA adapter? Thanks for your replies!

    Read the article

  • Managed Service Architectures Part I

    - by barryoreilly
    Instead of thinking about service oriented architecture, a concept that is continually defined, redefined, abused and mistreated, perhaps it is time to drop the acronym and consider what we actually need to get the job done.   ‘Pure’ SOA involves the modeling of an organisation’s processes, the so called ‘Top Down’ approach, followed by the implementation of these processes as services.     Another approach, more commonly seen in the wild, is the bottom up approach. This usually involves services that simply start popping up in the organization, and SOA in this case is often just an attempt to rein in these services. Such projects, although described as SOA projects for a variety of reasons, have clearly little relation to process driven architecture. Much has been written about these two approaches, with many deciding that a hybrid of both methods is needed to succeed with SOA.   These hybrid methods are a sensible compromise, but one gets the feeling that there is too much focus on ‘Succeeding with SOA’. Organisations who focus too much on bottom up development, or who waste too much time and money on top down approaches that don’t produce results, are often recommended to attempt an ‘agile’(Erl) or ‘middle-out’ (Microsoft) approach in order to succeed with SOA.  The problem with recommending this approach is that, in most cases, succeeding with SOA isn’t the aim of the project. If a project is started with the simple aim of ‘Succeeding with SOA’ then the reasons for the projects existence probably need to be questioned.   There are a number of things we can be sure of: ·         An organisation will have a number of disparate IT systems ·         Some of these systems will have redundant data and functionality ·         Integration will give considerable ROI ·         Integration will already be under way. ·         Services will already exist in the organisation ·         These services will be inconsistent in their implementation and in their governance   So there are three goals here: 1.       Alignment between the business and IT 2.     Integration of disparate systems 3.     Management of services.   2 and 3 are going to happen,  in fact they must happen if any degree of return is expected from the IT department. Ignoring 1 is considered a typical mistake in SOA implementations, as it ignores the business implications. However, the business implication of this approach is the money saved in more efficient IT processes. 2 and 3 are ongoing, and they will continue happening, even if a large project to produce a SOA metamodel is started. The result will then be an unstructured cackle of services, and a metamodel that is already going out of date. So we get stuck in and rebuild our services so that they match the metamodel, with the far reaching consequences that this will have on all our LOB systems are current. Lets imagine that this actually works ( how often do we rip and replace working software because it doesn't fit a certain pattern? Never -that's the point of integration), we will now be working with a metamodel that is out of date, and most likely incomplete if the organisation is large.      Accepting that an object can have more than one model over time, with perhaps more than one model being  at any given time will help us realise the limitations of the top down model. It is entirely normal , and perhaps necessary, for an organisation to be able to view an entity from different perspectives.   So, instead of trying to constantly force these goals in a straight line, why not let them happen in parallel, and manage the changes in each layer.     If  company A has chosen to model their business processes and create a business architecture, there will be a reason behind this. Often the aim is to make the business more flexible and able to cope with change, through alignment between the business and the IT department.   If company B’s IT department recognizes the problem of wild services springing up everywhere, and decides to do something about it, by designing a platform and processes for the introduction of services, is this not a valid approach?   With the hybrid approach, it is recommended that company A begin deploying services as quickly as possible. Based on models that are clearly incomplete, and which will therefore change rapidly and often in the near future. Natural business evolution will also mean that the models can be guaranteed to change in the not so near future. To ‘Succeed with SOA’ Company B needs to go back to the drawing board and start modeling processes and objects. So, in effect, we are telling business analysts to start developing code based on a model they are unsure of, and telling programmers to ignore the obvious and growing problems in their IT department and start drawing lines and boxes.     Could the problem be that there are two different problem domains? And the whole concept of SOA as it being described by clever salespeople today creates an example of oft dreaded ‘tight coupling’ between these two domains?   Could it be that we have taken two large problem areas, and bundled the solution together in order to create a magic bullet? And then convinced ourselves that the bullet actually exists?   Company A wants to have a closer relationship between the business and its IT department, in order to become a more flexible organization. Company B wants to decrease the maintenance costs of its IT infrastructure. If both companies focus on succeeding with SOA, then they aren’t focusing on their actual goals.   If Company A starts building services from incomplete models, without a gameplan, they will end up in the same situation as company B, with wild services. If company B focuses on modeling, they could easily end up with the same problems as company A.   Now we have two companies, who a short while ago had one problem each, that now have two problems each. This has happened because of a focus on ‘Succeeding with SOA’, rather than solving the problem at hand.   This is not to suggest that the two problem domains are unrelated, a strategy that encompasses both will obviously be good for the organization. But only if the organization realizes this and can develop such a strategy. This strategy cannot be bought in a box.       Anyone who has worked with SOA for a while will be used to analyzing the solutions to a problem and judging the solution’s level of coupling. If we have two applications that each perform separate functions, but need to communicate with each other, we create a integration layer between them, perhaps with a service, but we do all we can to reduce the dependency between the two systems. Using the same approach, we can separate the modeling (business architecture) and the service hosting (technical architecture).     The business architecture describes the processes and business objects in the business domain.   The technical architecture describes the hosting and management and implementation of services.   The glue that binds these together, the integration layer in our analogy, is the service contract, where the operations map the processes to their technical implementation, and the messages map business concepts to software objects in the implementation.   If we reduce the coupling between these layers, we should be able to allow developers to develop services, and business analysts to develop models, without the changes rippling through from one side to the other.   This would allow company A to carry on modeling, and company B to develop a service platform, each achieving their intended goal, without necessarily creating the problems seen in pure top down or bottom up approaches. Company B could then at a later date map their service infrastructure to a unified model, and company A could carry on modeling, insulating deployed services from changes in the ongoing modeling.   How do we do this?  The concept of service virtualization has been around for a while, and is instantly realizable in Microsoft’s Managed Services Engine. Here we can create a layer of virtual services, which represent the business analyst’s view, presenting uniform contracts to the outside world. These services can then transform and route messages to the actual service implementations. I like to think of the virtual services with their beautifully modeled interfaces as ‘SOA services’, and the implementations as simple integration ‘adapter’ services providing an interface to a technical implementation. The Managed Services Engine also provides policy based control over services, regardless of where they are deployed, simplifying handling of security, logging, exception handling etc.   This solves a big problem. The pressure to deliver services quickly is always there in projects. It is very important to quickly show value when implementing service architectures. There is also pressure to deliver quality, and you can’t easily do both at the same time. This approach allows quick delivery with quality increasing over time, allowing modeling and service development to occur in parallel and independent of each other. The link between business modeling and service implementation is not one that is obvious to many organizations, and requires a certain maturity to realize and drive forward. It is also completely possible that a company can benefit from one without the other, even if this approach is frowned upon today, there are many companies doing so and seeing ROI.   Of course there are disadvantages to this. The biggest one being the transformations necessary between the virtual interfaces and the service implementations. Bad choices in developing the services in the service implementation could mean that it is impossible to map the modeled processes to the implementation with redevelopment of the service. In many cases the architect will not have a choice here anyway, as proprietary systems are often delivered with predeveloped services. The alternative is to wait until the model is finished and then build the service according the model. However, if that approach worked we wouldn’t be having this discussion! And even when it does work, natural business evolution will mean that the two concepts (model and implementation) will immediately start to drift away from each other, so coupling them tightly together so that they are forever bound to the model that only applies at the time of the modeling work will not really achieve a great deal. Architecture is all about trade offs, and here a choice has to be made. The choice is between something will initially be of low quality but will work, or something that may well be impossible to achieve in most situations.         In conclusion, top-down is a natural approach for business analysts, and bottom-up  is a natural approach for developers. Instead of trying to force something on both that neither want, and which has not shown itself to be successful,  why not let them get on with their jobs, and let an enterprise architect coordinate the processes?

    Read the article

  • Good text editor for Ruby on Rails programming?

    - by Andrew
    I'm trying to find a text editor that I can use for doing Ruby on Rails development. I have been using TextMate on my Mac and would love to find something that even comes close to that experience. My Ubuntu laptop is a little old, and doesn't have a lot of memory, so I need something lightweight. I don't need/want a bloated IDE because the performance on my slow laptop would be terrible. It would be nice if this text editor had: Syntax highlighting A project/file browser view to be able to open files in my project Keyboard shortcuts (don't need them as much)

    Read the article

  • Eight Geektacular Christmas Projects for Your Day Off

    - by Jason Fitzpatrick
    It’s Christmas Eve and if you’re lucky you’ve got some time off ahead of you. Let’s put that time to good use with some holiday-centered geeking out. Come on in for LEGO ornaments, Darth Vader snow flakes, and Christmas light hacks galore. Latest Features How-To Geek ETC How to Use the Avira Rescue CD to Clean Your Infected PC The Complete List of iPad Tips, Tricks, and Tutorials Is Your Desktop Printer More Expensive Than Printing Services? 20 OS X Keyboard Shortcuts You Might Not Know HTG Explains: Which Linux File System Should You Choose? HTG Explains: Why Does Photo Paper Improve Print Quality? An Alternate Star Wars Christmas Special [Video] Sunset in a Tropical Paradise Wallpaper Natural Wood Grain Icons for Your Desktop and App Launcher Docks My Blackberry Is Not Working! The Apple Too?! [Funny Video] Hidden Tracks Your Stolen Mac; Free Until End of January Why the Other Checkout Line Always Moves Faster

    Read the article

  • Terminal will not accept password / terminale non accetta password

    - by elvizz67
    Translation, by Google Translate: I just installed lxde Lubuntu errrato and a terminal command I locked the package management and updates, after erasing software from sources I typed the wrong command sudo apt-get upgrade pero'il terminal asks me the password and the keyboard is not me accept any command. What should I do to unlock? Original text: Ho appena installato lubuntu lxde e per una errrato comando il terminale mi ha bloccato la gestione pacchetti e aggiornamenti, dopo avere cancellato dalle fonti software il comando sbagliato ho digitato sudo apt-get upgrade pero'il terminale mi chiede la password e la tastiera non mi accetta nessun comando. cosa devo fare per sbloccare ?

    Read the article

  • Configure mounting timeout at boot

    - by René Pieters
    On remotely rebooting a 12.04 machine I found it hanging at the "unable to mount soandso: Skip, Manual Abort? " (That's pretty much how I remember the message) The machine was basically stopped there until I hooked up a keyboard and pressed "s". I can see the rationale for the question, but I'd really like to know where to configure it or turn it off altogether. A mandatory question like this makes sense in a desktop environment but for servers I'd like more flexibility. So where do I fiddle and tweak this?

    Read the article

  • Alt Tab with xdotool and xkeybind

    - by Gary
    I'm trying to map alt tab behavior to a mouse button using xdotool in my .xbindkeysrc. What I want to do is map something like "xdotools keydown alt key Tab" release b:7 but by doing this the alt key gets stuck, and it prevents my other mappings from working until I press the alt key on the keyboard. Is there a way to do this without having to press the alt key to get things unstuck? Without keydown it just switches back and forth between two windows, with keydown it behaves like I want (displaying the alt-tab switcher) but the alt key is stuck. Is there a way to unstick the key after using the alt tab switcher using xdotools?

    Read the article

  • Ubuntu 10.10 freezes

    - by albert green
    Hi, I have the following configuration: CPU: amd athlon 64bit Motherboard: Phoenix with awardBIOS v6.0 APIC mode = disabled Video card: GeForce 5500 Sound card: soundblaster live 24bit RAM: 1.5GB Partitions: ext3 32GB where I installed Ubuntu. Swap 500MB. 180GB Fat32 unmounted. contains data only. I installed Ubuntu 10.10 and used the "Additional driver" program to get the nVidia proprietary drivers. The system works for sometime and randomly freezes. The mouse and keyboard are not responding and I am not able to do a SSH. The only solution is pressing the power button. I tried Kubuntu 10.10 and 32/64-bits versions as well. There is nothing in the logs to suggest what the reason might be. Any help will be appreciated. Thanks in advance.

    Read the article

  • Minimum brightness level is brighter than the levels just above it

    - by greggory.hz
    I'm using a macbook pro 8,1. When I use the brightness keys on the keyboard, decreasing the brightness works until it reaches the lowest brightness level. At that point (when on the lowest level), the brightness is something like 70 or 80% instead of 0-10% (or whatever it usually is). All the other levels are fine. How can I adjust/calibrate these brightness controls? PS: I'm currently using 12.04 but I've had the same issue no later than 11.10 (and I'm fairly confident it was present in 11.04 as well). UPDATE: If I go to System Settings and select "Brightness and Lock" the exact same thing happens, but there is more fine grained control of the brightness. It goes to completely black when the dial is about 10% from the left edge. When it passes that point, anywhere inside that left 10% appears to be about 70 - 80% brightness (there's no change regardless of where the dial is in this section).

    Read the article

< Previous Page | 308 309 310 311 312 313 314 315 316 317 318 319  | Next Page >