Search Results

Search found 212 results on 9 pages for 'gestures'.

Page 5/9 | < Previous Page | 1 2 3 4 5 6 7 8 9  | Next Page >

  • Week in Geek: Google Drive Desktop Client Allows Backdoor Access to Google Accounts

    - by Asian Angel
    Our last edition of WIG for October is filled with news link coverage on topics such as Microsoft may not issue a second Windows 7 service pack, Windows Media Center is free for Windows 8 Pro users for limited time, CyanogenMod logged swipe gestures used to unlock Android devices, and more. What Is the Purpose of the “Do Not Cover This Hole” Hole on Hard Drives? How To Log Into The Desktop, Add a Start Menu, and Disable Hot Corners in Windows 8 HTG Explains: Why You Shouldn’t Use a Task Killer On Android

    Read the article

  • What is the difference between development and R&D?

    - by MainMa
    I was asked by a colleague to explain clearly the difference between ordinary development and research and development (R&D) and was unable to do it. After reading Wikipedia, I still don't have the precise answer. According to Wikipedia (slightly modified): There are two primary models: In one model, the primary function is to develop new products; in the other model, the primary function is to discover and create new knowledge about scientific and technological topics for the purpose of uncovering and enabling development of valuable new products, processes, and services. The first model is confusing. Does it mean that development (not R&D) consists exclusively in adding new features to a product, solving bugs and doing maintenance? What if something which was previously developed as a new feature becomes a separate product? The second model is less confusing, but still, how to qualify whether something is new knowledge or existent knowledge which is just rediscovered? Later, Wikipedia adds that ordinary development is different from R&D because of its: nearly immediate profit or immediate improvement. It's still not clear enough. How to qualify "nearly immediate profit"? What if a task has an immediate profit but requires heavy research? Or if it is basic but has uncertain profit, like the enforcement of a common style over the codebase? For example, does it belong to development or R&D to: Develop an engine which abstracts the access to the database, simplifying and shortening enormously the code of other applications (existent or ones which will be written in future) which should access to the database? Establish a new service-oriented architecture for the entire organization of company resources, in order to move from a bunch of separate and autonomous applications to a set of well-organized, interconnected web services, like what is used by Amazon? Design a new communication protocol to allow faster replication of data between two data centers of the company? Conceive a new type of software testing while working on a specific product, knowing that this type of testing will improve/simplify the testing process? Prove that Functional programming is more appropriate than OOP for a specific application, based on evidence, logic and previous experience? Enhance the existent application by adding gestures on tactile screens, after doing studies and testing that shows that those gestures improve the productivity of the users by a ratio of at least 1.4 for a precise set of tasks? Find a way to strongly enhance the Power usage effectiveness (PUE) of a data center? Create a Domain-Specific Language (DSL)? In short, how could I determine whether I'm doing R&D while working on something?

    Read the article

  • Acer touchpad multitouch not working

    - by Lassi
    I tried to get my Acer laptop's touchpad working properly. I installed driver package (psmouse-alps-dkms_0.10_all) which made the side scrolling working. However, the touchpad supports multitouch (works on Windows), but Synaptiks claims that only 1 finger is detected by the touchpad. Thus 2 finger scrolling, pinch & pan zoom and those are not working. Any ideas how I could get these multi-finger gestures working?

    Read the article

  • What resources are available for iPhone development using AS3?

    - by woodscreative
    I've just realeased my first game developed for the iPhone using AS3 and the iPhone Packager http://itunes.apple.com/us/app/snapshot-paintball/id407362440?mt=8&uo=4 I want to take the game to the next level but I am not using the native iPhone SDK so I need some other ideas, I am fresh to iPhone development and it's hard to find good resources, any AS3 developers out there willing to share some links? Highscore frameworks and best practices, connecting to Facebook, ui classes/gestures. Thanks.

    Read the article

  • Can I invoke a graphical program to run with a hidden window?

    - by Willi Ballenthin
    I have a program that interprets the input from my mouse and provides multitouch gestures. I wrote it using a game engine, which always creates a window when it runs. I'd like to run this program without seeing the window, so this is the point of the questions: Can invoke a graphical program but have the window hidden? By hidden I mean anything that keeps me from encountering the window on my desktop. So I would accept something like, "open on a different, previously non-existant screen."

    Read the article

  • iPad. UIBarButtonItem has an undocumented view of type UIToolbarTextButton. Huh?

    - by dugla
    I have an iPad app where I have a view controller that is the UIGestureRecognizerDelegate for a number of UIGestureRecognizers. I have implemented the following method of UIGestureRecognizerDelegate: - (BOOL)gestureRecognizer:(UIGestureRecognizer *)gestureRecognizer shouldReceiveTouch:(UITouch *)touch { // Double tapping anywhere on the screen hides/shows the toolbar if ([gestureRecognizer isKindOfClass:[UITapGestureRecognizer class]] == YES) { if (touch.tapCount == 2) { self.toolbar.hidden = self.toolbar.isHidden ? NO : YES; } // if (touch.tapCount == 2) } // if ([gestureRecognizer isKindOfClass:[UITapGestureRecognizer class]] == YES) // All gestures are ignored unless they happen on the fullscreen EAGLView if ([touch.view isKindOfClass:[EAGLView class]] == NO) { return NO; } // if ([touch.view isKindOfClass:[EAGLView class]] == NO) return YES; } My setup is a fullscreen EAGLView with a UIToolbar atop the EAGLView. There is a UIBarButtonItem on the toolbar. The idea here is that double-tapping anywhere toggles the appearance of the toolbar. All other gestures must occur on the EAGLView. My problem is that taps directly on the UIBarButtonItem show touch.view to be the UIView subclass UIToolbarTextButton which is undocumented and can't be introspected. Huh? Can someone suggest a work around, preferably that uses introspective goodness of some form? Thanks, Doug Thanks, Doug

    Read the article

  • Overlaying 2D paths on UIImage without scaling artifacts

    - by tat0
    I need to draw a path along the shape of an image in a way that it is always matching its position on the image independent of the image scale. Think of this like the hybrid view of Google Maps where streets names and roads are superimposed on top of the aerial pictures. Furthermore, this path will be drawn by the user's finger movements and I need to be able to retrieve the path keypoints on the image pixel coordinates. The user zooms-in in order to more precisely set the paths location. I manage to somehow make it work using this approach: -Create a custom UIView called CanvasView that handles touches interaction and delivers scaling, rotation, translation values to either the UIImageView or PathsView (see bellow) depending on a flag: deliverToImageOrPaths. -Create a UIImageView holding the base image. This is set as a children of CanvasView -Create a custom UIView called PathsView that keeps track of the 2D paths geometry and draws itself with a custom drawRect. This is set as children of the UIImageView. So hierarchy: CanvasView - UIImageView -PathsView In this way when deliverToImageOrPaths is YES, finger gestures transforms both the UIImageView and its child PathsView. When deliverToImageOrPaths is NO the gestures affect only the PathsView altering its geometry. So far so good. QUESTION: The problem I have is that when scaling the base UIImageView (via its .transform property) the PathsView is scaled with aliasing artifacts. drawRect is still being called on the PathsView but I guess it's performing the drawing using the original buffer size and then interpolating. How can I solve this issue? Are there better ways to implement these features? PS: I tried changing the PathsView layer class to CATiledLayer with levelsOfDetailBias 4 and levelsOfDetail 4. It solves the aliasing problem to some extent but it's unacceptable slow to render.

    Read the article

  • How can I create a finger scrollable Textbox in WM 6.5?

    - by Papajohn
    Hi everybody. I just noticed something weird in WM 6.5 emulators. Unlike 6.1 where finger panning kind of worked, the only way to scroll a Textbox appears to be through scrollbars. This behaviour is in contrast to what they have done for comboboxes: they are now gesture-friendly without the programmer's intervention. I.e. the user can select a choice from a standard drop down menu by panning and scrolling. Previously, you had to use the embedded scrollbar. The combobox's case implies that MS took some measures to provide standard gesture support for classic finger gestures, yet I cannot see something similar for textboxes. This makes me ask the following: Is there anything that can be done to make textboxes finger scrollabe easily? Note that I refer to managed .NET CF development. It is my understanding that in native development I could use the new Gestures API to achieve the scrolling effect. Yet, I am not sure if there is an easier and more straightforward method that I have missed.

    Read the article

  • How do you get the viewport scale after pinch/zoom on an iPhone web app?

    - by Loktar
    Does anyone know how to get the size in pixels or scale value of the viewport after a user has pinched or double tapped to zoom in/out on a page in JavaScript? I've tried using window.innerWidth but I've had mixed results. Sometimes it seems to accurately give the number of pixels the viewport is showing, however, if I zoom way in on a page and then do a large pinch to zoom back out, window.innerWidth will be around 600-700 even though it is only showing ~200px of the page. The page is only 400px wide and it didn't show the checkered "you've gone too far" background you see when you zoom out beyond the page size. If I do small pinches to zoom in and out, window.innerWidth appears to work just fine. Unfortunately I can't rely on a user only making small pinch gestures :) I've also tried to use the scale property on the gesture event object, but I've found that unreliable because you don't always know the initial scale when you reload the page or use back/forward buttons to navigate to it even when using the meta tag to specify it. Ultimately, I'm trying to make an app that is aware of when a user is trying to zoom out beyond the maximum zoom level so if there is another way to do this I'm interested in hearing about it :) Here's the code I'm using to get the innerWidth: document.body.addEventListener('gestureend', function (evt) { console.log(window.innerWidth); // inaccurate when doing large pinch gestures }, false); Thanks!

    Read the article

  • Is there a way to touch-enable scrolling in a WPF ScrollViewer?

    - by Brian Sullivan
    I'm trying to create a form in a WPF application that will allow the user to use iPhone-like gestures to scroll through the available fields. So, I've put all my form controls inside a StackPanel inside a ScrollViewer, and the scrollbar shows up as expected when there are too many elements to be shown on the screen. However, when I try to test this on my touch-enabled device, a panning gesture (placing a finger down on the surface and dragging it upward) does not move the viewable area down as I would expect. When I simply put a number of elements inside a ListView, the touch gestures work just fine. Is there any way to enable the same kind of behavior in a ScrollViewer? My window is structured like this: <Window x:Class="TestTouchScrolling.MainWindow" xmlns="http://schemas.microsoft.com/winfx/2006/xaml/presentation" xmlns:x="http://schemas.microsoft.com/winfx/2006/xaml" Title="MainWindow" Height="350" Width="525" Loaded="Window_Loaded"> <Grid> <ScrollViewer Name="viewer" VerticalScrollBarVisibility="Auto"> <StackPanel Name="panel"> <StackPanel Orientation="Horizontal"> <Label>Label 1:</Label> <TextBox Name="TextBox1"></TextBox> </StackPanel> <StackPanel Orientation="Horizontal"> <Label>Label 2:</Label> <TextBox Name="TextBox2"></TextBox> </StackPanel> <StackPanel Orientation="Horizontal"> <Label>Label 3:</Label> <TextBox Name="TextBox3"></TextBox> </StackPanel> <!-- Lots more like these --> </StackPanel> </ScrollViewer> </Grid>

    Read the article

  • How to access the Quick Link / Admin Menu in Windows 8 using touch?

    - by it depends
    In Windows 8 and Windows RT you can right-click in the bottom-left corner to access a menu of commonly used desktop links. You can access the same menu with the keyboard using Windows+X. Is there a way to access this menu using touch? I have a Surface RT and have tried a number of gestures in the corner (e.g., press and hold, swipe down) on both the Desktop and the Start Screen without any luck.

    Read the article

  • Does Apple Magic Mouse fully work on Windows 7 x86/x64?

    - by Sorin Sbarnea
    I would like to know if Apple Magic Mouse works on Windows 7 (x86/x64) on non-Apple computers. Here are some checklists: x64 compatibility left click right click middle button? vertical scroll horizontal scroll bind additional gestures to keystrokes? are any usage problems? In case it works please advise on how to get the drivers.

    Read the article

  • Kinect Turns DaVinci Physics Application Super Cool

    - by Gopinath
    Guys at RazorFish who are well known for their Microsoft Surface impressive stuff has ported their Da Vinci Application over to Kinect device. The end result is a super cool gesture based application. Check out the embedded video demonstration below If you wondering what is Da Vince Application is all about, here are few lines from RazorFish DaVinci is a Microsoft Surface application that blurs the lines between the physical and virtual world by combining object recognition, real-world physics simulation and gestural interface design. Related:Kinect + Windows 7 = Control PC With Hand Gestures This article titled,Kinect Turns DaVinci Physics Application Super Cool, was originally published at Tech Dreams. Grab our rss feed or fan us on Facebook to get updates from us.

    Read the article

  • Jaw Dropping Kinect Integration With Combat Solider Game

    - by Gopinath
    The innovation in natural user interface for interacting with computers and other devices is riding on the brilliance of Microsoft’s XBox Kinect. The amazing technology behind Kinect lets users to plays games without touching game controllers. It enables users to control and interact with XBox 360 games using gestures(body movements) and spoken commands. Earlier we have seen Kinect in controlling Windows 7 PCs, simulating Da Vince application. At Microsoft’s E3 keynote, game publisher Ubisoft demoed Kinect integration with the future version of a Soldier game. The usage of Kinect to change weapons and play the game is jaw dropping. It is tough to explain the experience in words, check out the embedded demo video This article titled,Jaw Dropping Kinect Integration With Combat Solider Game, was originally published at Tech Dreams. Grab our rss feed or fan us on Facebook to get updates from us.

    Read the article

  • Disabling Navigation Flicks in WPF

    - by Brian Genisio's House Of Bilz
    I am currently working on a multi-touch application using WPF.  One thing that has been irritating me with this development is an automatic navigation forward/back command that is bound to forward and backwards flicks.  Many of my touch-based interactions were being thwarted by gestures picked up by WPF as navigation.  I just wanted to disable this behavior. My programmatic back/forward calls are not affected by this change, which is nice.  Here is how I did it:  In my main window, I added the following command bindings:<NavigationWindow.CommandBindings> <CommandBinding Command="NavigationCommands.BrowseBack" Executed="DoNothing" /> <CommandBinding Command="NavigationCommands.BrowseForward" Executed="DoNothing" /> </NavigationWindow.CommandBindings> Then, the DoNothing method in the code-behind does nothing:private void DoNothing(object sender, ExecutedRoutedEventArgs e) { } There may be a better way to do this, but I haven’t found one.

    Read the article

  • Flutter Enabled Hand Gesture Control of Your Music and Movies

    - by Jason Fitzpatrick
    You’ve got a dozen windows open, your media player is buried underneath somewhere, and the phone rings. Do you reach for the speaker power button, dig through the windows, or simply wave your hand to pause the music? With Flutter, a simple gesture is all it takes. Flutter is a free application for Windows and Mac OS X that turns your webcam into a gesture-recognition tool and your hand into a remote for your media applications. By lifting your your hand palm out, making a fist, or using other simple gestures, you can pause or otherwise interact with your media players. Hit up the link below to grab a free copy of Flutter. FlutterApp [via MakeUseOf] How Hackers Can Disguise Malicious Programs With Fake File Extensions Can Dust Actually Damage My Computer? What To Do If You Get a Virus on Your Computer

    Read the article

  • Sluggish/unresponsive trackpad on pre-unibody MacBook Pro

    - by Art
    I am running Ubuntu Natty Narwhal on pre-unibody MacBook Pro (2007 I believe). There seems to be a problem with trackpad - it barely works, in order to move cursor you have to move your finger a lot, and it terminates the 'gestures' abruptly - say you are moving the cursor with your finger and out of the blue it just stops, although the finger is still in contact with the surface of the trackpad. Those issues seem to dissapear as soon as I boot Mac OS X, so I suspect it is something Ubuntu-specific. Also, if I try to move the cursor with not just fingertip but increase the contact area, it seems to work just fine, although it is hardly convenient.

    Read the article

  • Chrome + Touchscreen + Unity (14.04)

    - by Luis Fernando P. Gonzalez
    I've got a Dell 14R with touchscreen with Ubuntu 14.04. When I start Google Chrome by clicking the launcher with the TouchPad or with the Super+ combination, touch support in Chrome doesn't work. When I touch some element, it only gets focus, but doesn't activate the element (button, field, etc). And touch gestures don't work. In the other hand, if I start Chrome by touching the launcher button, touch scroll works in Chrome, even flicking to left does back to the previous page, etc. It doesn't makes sense to me, but is always like this, even across reboots. Anyone can confirm this? Any idea on how to fix?

    Read the article

  • Mouse wheel on Dell Vostro V13

    - by Scaine
    I have a Dell Vostro V13, which Canonical claim is "Certified" for Ubuntu (http://www.ubuntu.com/certification/make/Dell/laptops). So it surprises me that the synaptics trackpad isn't recognised as such and defaults to psmouse emulation. This means no multitouch, no gestures, and more importantly, no mousewheel scrolling. Because it's not loading the synaptic driver, running gpointing-device-settings doesn't help. I've found this bug : https://bugs.launchpad.net/ubuntu/+source/xserver-xorg-input-synaptics/+bug/380126 which suggests a fix has been committed, but I can't see a way of getting this loaded on my Maverick build. I've tried backports and proposed repos in synaptic, but no joy. Any ideas?

    Read the article

  • in ubuntu 12.04 how may i know if my devices have it's driver installed??

    - by Aldo
    i have a dell N4110 laptop, and i want to know if the driver is installed and working well, something like the device manager in windows , or another way to know if a device is driverless or if the device might have a better driver, like my mousepad, in windows it have multi-touch gestures , that scrolls or zoom with two fingers (like an ipod) but in ubuntu it just works the right part as a scroll bar, so maybe it is installed one driver, but i need other one that uses well my devices. and the grphics card, i have not idea if it is well installed or isn't. i have a Intel 3000hd graphics card. thank you for your time. have a nice day people! =D

    Read the article

  • Pinch gesture in QLPreviewController?

    - by Arne
    I am using the QLPreviewController in my project, but cannot use pinch gestures to zoom in. What do I need to set up first? This is how I create the view controller: - (void)tableView:(UITableView *)tableView didSelectRowAtIndexPath:(NSIndexPath *)indexPath { QLPreviewController *previewer = [[QLPreviewController alloc] init]; [previewer setDataSource:self]; [previewer setCurrentPreviewItemIndex:indexPath.row]; [[self navigationController] pushViewController:previewer animated:YES]; }

    Read the article

  • Android: How to detect end of scroll

    - by Mix
    Hi! I need to implement HorizontalScrollView which is scrolled to predefined positions (similar to Home behaviour). It works with slow gestures, but does not work with flings. I hooked to View.onScrollChanged() and it is called when scrolling happened, but I can't determine when scrolling ends. In theory there should be a way to say that fling is over. Is there such API?

    Read the article

  • iphone - double tap fail safe way

    - by mike
    I am trying to detect double taps on a view, but when the double tap comes, the first tap triggers an action on TouchesBegan, so, before detecting a double tap a single tap is always detected first. How can I make this in a way that just detects the double tap? I cannot use OS 3.x gestures, because I have to make it compatible with old OS versions. thanks

    Read the article

  • How to display image in Swing? in textarea?

    - by Nitesh Panchal
    Hello, I am creating a chat application using JApplet. I have a TextArea where all chat messages go. Everything in working fine and smooth just as you would expect a basic chat application to do. Now i want to add support for gestures. I wanted to know, how can we show an icon in textarea? it only takes string in append() method. Thanks in advance :)

    Read the article

< Previous Page | 1 2 3 4 5 6 7 8 9  | Next Page >