Search Results

Search found 10018 results on 401 pages for 'sample rate'.

Page 99/401 | < Previous Page | 95 96 97 98 99 100 101 102 103 104 105 106  | Next Page >

  • Windows Phone 7 dev: C# or silverlight for a simple app?

    - by OneWorld
    I'm about to hire a programmer to develop Windows Phone 7 apps. The current app that shall be developed is quite simple. The app will download content from a Web-API. There are two lists to select content. There is a single item content page. Users can rate the content and upload photos. The screens will be produced by our artist. I am pretty sure that most of the available programmers haven't touched WP7 development yet. Now the questions are: What technology is suitable for this kind of app? What technology requires less research, learning and production time? Do you already have experience of limitations of silverlight compared to C#? (I am also thinking of future projects) My guess is that silverlight programmers are more experienced in UI programming than C# coders. Do you feel the same way?

    Read the article

  • OTN Article: The Enterprise Side of JavaFX (part 1 of 2)

    - by terrencebarr
    OTN just published part 1 of a series by Adam Bien on “The Enterprise Side of JavaFX”. In this article, learn how to use LightView to convert REST services into a bindable set of properties, using JavaFX, Glassfish, LightFish, and Maven. Sample code included. Part 2 will discuss the integration of a JavaServer Faces 2 UI with WebView. Cheers, – Terrence Filed under: Mobile & Embedded Tagged: glassfish, JavaFX

    Read the article

  • Should you charge clients hours spent on the wrong track?

    - by Lea Verou
    I took up a small CSS challenge to solve for a client and I'm going to be paid on a hourly rate. I eventually solved it, it took 4 hours but I spent roughly 30% of the time in the wrong track, trying a CSS3 solution that only worked in recent browsers and finally discovering that no fallback is possible via JS (like I originally thought). Should I charge the client that 30%? More details: I didn't provide an estimate, I liked the challenge per se, so I started working on it before giving an estimate (but I have worked with him before, so I know he's not one of those people that have unrealistic expectations). At the very worst I will have spent 4 unpaid hours on an intriguing CSS challenge. And I will give the fairest possible estimate for both of us, since I will have already done the work. :)

    Read the article

  • Examples of datecentric LOB-application in Silverlight with creative design?

    - by Alexander Galkin
    I am a database developer and I have rather limited experience with interface design. I am currently working on a project in my free-time a freetime which is mostly data centric and I would like to develop an eye-catchy interface for it using Silverlight. What I am looking for are examples of nice and interesting LOB applications in Siverlight without the use of paid frameworks. So far, I could only find something like Telerik sample application, but it uses a lot of Telerik controls I can't afford to buy.

    Read the article

  • Help us with our git workflow

    - by Brandon Cordell
    We have a web application that gets deployed to multiple regions around our state. An instance of the application for each region. We maintain a staging and production (master) branch in our repository, but we were wondering what is the best way of maintaining each instances codebase. It's similar at the core, but we have to give each region the ability to make specific requests that may not make it into the core of the application. Right now we have branches for each region, like region_one_staging, and region_one_production. At the rate we're growing we'll have hundreds of branches here in the next few years. Is there a better way to do this?

    Read the article

  • I'm blogging again, and about time too

    - by fatherjack
    No, seriously, this one is about time. I recently had an issue in a work database where a query was giving random results, sometimes the query would return a row and other times it wouldn't. There was quite a bit of work distilling the query down to find the reason for this and I'll try to explain by demonstrating what was happening by using some sample data in a table with rather a contrived use case. Let's assume we have a table that is designed to have a start and end date for something, maybe...(read more)

    Read the article

  • Palm take-over watch

    This is just in: HTC is said to be considering taking over Palm. That would be an interesting development, since HTC is one of the few parties we can trust not messing up webOS but actually using it as it’s supposed to work.Palm has an absolutely first-rate product in webOS, especially its user interface. As far as I’m concerned the Palm Pre is the only phone that’s (almost?) on a par with the iPhone when it comes to UI, although the system is completely different (and has supported multitasking...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Fork dead SVN based project on GitHub

    - by Quinn Bailey
    I previously asked this at stack overflow but it was closed, I believe because 'programmers' is a more appropriate venue for this question. I have done some work on the SVN Importer project (Apache license), which appears to be effectively dead (no published changes in 5 years). I have a login to their svn server but do not have commit rights. At any rate, I'd like to convert this project to Git and push my own changes to GitHub. The GitHub site suggests the svn2git tool for converting svn projects to Git, so I was planning to convert the SVN repository to Git, add my changes, and then push this Git repository to GitHub. I'm wondering, what are the legal requirements and common conventions of this process? Is it acceptable to clone the entire history of the project and move it to GitHub? Also, even though this is essentially a dead project, once I've translated the repository to Git should I put all of my commits onto a non master branch or is it acceptable to use master in this case?

    Read the article

  • slow speed transfering files to a local network device

    - by F. Ariel Jung
    I'm having troubles copying files from my pc(ubuntu 13.04 64 bits) to a Media Center(WDTV Live: with an SATA2 HDD attached) through the local network. the speed is to slow, it's up to 2.0 MB/s image: http://s11.postimg.org/g1nuai92b/Captura_de_pantalla_de_2013_06_24_07_46_55.png (I don't use image tag because it is a big image) my PC has a wifi network card Tp Link TL-WN951N and it is connected to a TP Link TL-WDR3500 router via wifi, and the WDTV Live is connected via LAN here is a iwconfig: eth0 no wireless extensions. lo no wireless extensions. wlan0 IEEE 802.11bgn ESSID:"TL-WDR3500" Mode:Managed Frequency:2.462 GHz Access Point: A0:F3:C1:6C:3B:F1 Bit Rate=39 Mb/s Tx-Power=20 dBm Retry long limit:7 RTS thr:off Fragment thr:off Power Management:off Link Quality=62/70 Signal level=-48 dBm Rx invalid nwid:0 Rx invalid crypt:0 Rx invalid frag:0 Tx excessive retries:33521 Invalid misc:2245290 Missed beacon:0 if more data is required I'll post it as needed. I can't figure this out, needed help. please

    Read the article

  • Building an MVC application using QuickBooks

    - by dataintegration
    RSSBus ADO.NET Providers can be used from many tools and IDEs. In this article we show how to connect to QuickBooks from an MVC3 project using the RSSBus ADO.NET Provider for QuickBooks. Although this example uses the QuickBooks Data Provider, the same process can be used with any of our ADO.NET Providers. Creating the Model Step 1: Download and install the QuickBooks Data Provider from RSSBus. Step 2: Create a new MVC3 project in Visual Studio. Add a data model to the Models folder using the ADO.NET Entity Data Model wizard. Step 3: Create a new RSSBus QuickBooks Data Source by clicking "New Connection", specify the connection string options, and click Next. Step 4: Select all the tables and views you need, and click Finish to create the data model. Step 5: Right click on the entity diagram and select 'Add Code Generation Item'. Choose the 'ADO.NET DbContext Generator'. Creating the Controller and the Views Step 6: Add a new controller to the Controllers folder. Give it a meaningful name, such as ReceivePaymentsController. Also, make sure the template selected is 'Controller with empty read/write actions'. Before adding new methods to the Controller, create views for your model. We will add the List, Create, and Delete views. Step 7: Right click on the Views folder and go to Add -> View. Here create a new view for each: List, Create, and Delete templates. Make sure to also associate your Model with the new views. Step 10: Now that the views are ready, go back and edit the RecievePayment controller. Update your code to handle the Index, Create, and Delete methods. Sample Project We are including a sample project that shows how to use the QuickBooks Data Provider in an MVC3 application. You may download the C# project here or download the VB.NET project here. You will also need to install the QuickBooks ADO.NET Data Provider to run the demo. You can download a free trial here. To use this demo, you will also need to modify the connection string in the 'web.config'.

    Read the article

  • How to remove tearing in videos with Nvidia driver?

    - by Daimon
    I'm using ubuntu 11.10 as my HTPC OS on ION board. Tearing was always a big problem but this guide (http://www.omgubuntu.co.uk/2010/01/how-to-fix-video-tearing-in-videos-nvidia-ubuntu/) used to solve it in previous versions. I mean it used to because it doesn't solve it anymore. I tried making same changes (although compiz config is little different and same options are in different panels) but with no success. I also tried to manually set resolution and refresh rate of HDTV in nvidia settings - no changes at all. I know that open source drivers would solve a problem but it's not an option because I need VDPAU acceleration. I've also read that completely disabling of Composite solves the problem - can't do that because compizconfig-settings-manager crashes on attempt. More info about setup: Ubuntu 11.10 64bit Nvidia ION board newest official version of proprietary driver SMplayer with vdpau settings Full HD HDTV

    Read the article

  • Frame timing for GLFW versus GLUT

    - by linello
    I need a library which ensures me that the timing between frames are more constant as possible during an experiment of visual psychophics. This is usually done synchronizing the refresh rate of the screen with the main loop. For example if my monitor runs at 60Hz I would like to specify that frequency to my framework. For example if my gameloop is the following void gameloop() { // do some computation printDeltaT(); Flip buffers } I would like to have printed a constant time interval. Is it possible with GLFW?

    Read the article

  • How do I fix the Google Webmaster Tools warning: "URL not followed?"

    - by user3611500
    A few days after submitting my sitemap to Google, I received this warning: When we tested a sample of URLs from your Sitemap, we found that some URLs redirect to other locations. We recommend that your Sitemap contain URLs that point to the final destination (the redirect target) instead of redirecting to another URL. The example URL Google gave me is http://iketqua.net/?_escaped_fragment_=CIDTKT/mien-trung/xo-so-kon-tum I checked all possible things that I could think of, but still can't figure out what the warning is about! My sitemap: http://iketqua.net/sitemap.xml

    Read the article

  • Music server + Banshee

    - by slosd
    I'm currently trying to figure out some kind of set-up for my music collection. The music collection is located on my "main PC" and I'm using Banshee to play/manage the music. Since I'm often at a different location where I use my laptop I have a FTP server running on my main PC, use curlftpfs to mount the music in /home/username/Music and sync the Banshee DB/configuration via Dropbox. The problem at the moment is that Banshee skips songs very often when the connection is a little slower. I've been looking at some servers like Ampache but they are not really an option since they don't let me use Banshee (which means I have no ability to change metadata or rate songs). Syncing the whole collection to my laptop isn't an option either because of its size. It would be great if you had some ideas how to make my current set-up more reliable or maybe know an alternative.

    Read the article

  • Interpolation gives the appearance of collisions

    - by Akroy
    I'm implementing a simple 2D platformer with a constant speed update of the game logic, but with the rendering done as fast as the machine can handle. I interpolate positions between actual game updates by just using the position and velocity of objects at the last update. This makes things look really smooth in general, but when something hits a wall/floor, it appears to go through the wall for a moment before being positioned correctly. This is because the interpolator is not taking walls into account, so it guesses the position into walls until the actual game update fixes it. Are there any particularly elegant solutions for this? Simply increasing the update rate seems like a band-aid solution, and I'm trying to avoid increasing the system reqs. I could also check for collisions in the actual interpolator, but that seems like heavy overhead, and then I'm no longer dividing the drawing and the game updating.

    Read the article

  • CPU Usage in Very Large Coherence Clusters

    - by jpurdy
    When sizing Coherence installations, one of the complicating factors is that these installations (by their very nature) tend to be application-specific, with some being large, memory-intensive caches, with others acting as I/O-intensive transaction-processing platforms, and still others performing CPU-intensive calculations across the data grid. Regardless of the primary resource requirements, Coherence sizing calculations are inherently empirical, in that there are so many permutations that a simple spreadsheet approach to sizing is rarely optimal (though it can provide a good starting estimate). So we typically recommend measuring actual resource usage (primarily CPU cycles, network bandwidth and memory) at a given load, and then extrapolating from those measurements. Of course there may be multiple types of load, and these may have varying degrees of correlation -- for example, an increased request rate may drive up the number of objects "pinned" in memory at any point, but the increase may be less than linear if those objects are naturally shared by concurrent requests. But for most reasonably-designed applications, a linear resource model will be reasonably accurate for most levels of scale. However, at extreme scale, sizing becomes a bit more complicated as certain cluster management operations -- while very infrequent -- become increasingly critical. This is because certain operations do not naturally tend to scale out. In a small cluster, sizing is primarily driven by the request rate, required cache size, or other application-driven metrics. In larger clusters (e.g. those with hundreds of cluster members), certain infrastructure tasks become intensive, in particular those related to members joining and leaving the cluster, such as introducing new cluster members to the rest of the cluster, or publishing the location of partitions during rebalancing. These tasks have a strong tendency to require all updates to be routed via a single member for the sake of cluster stability and data integrity. Fortunately that member is dynamically assigned in Coherence, so it is not a single point of failure, but it may still become a single point of bottleneck (until the cluster finishes its reconfiguration, at which point this member will have a similar load to the rest of the members). The most common cause of scaling issues in large clusters is disabling multicast (by configuring well-known addresses, aka WKA). This obviously impacts network usage, but it also has a large impact on CPU usage, primarily since the senior member must directly communicate certain messages with every other cluster member, and this communication requires significant CPU time. In particular, the need to notify the rest of the cluster about membership changes and corresponding partition reassignments adds stress to the senior member. Given that portions of the network stack may tend to be single-threaded (both in Coherence and the underlying OS), this may be even more problematic on servers with poor single-threaded performance. As a result of this, some extremely large clusters may be configured with a smaller number of partitions than ideal. This results in the size of each partition being increased. When a cache server fails, the other servers will use their fractional backups to recover the state of that server (and take over responsibility for their backed-up portion of that state). The finest granularity of this recovery is a single partition, and the single service thread can not accept new requests during this recovery. Ordinarily, recovery is practically instantaneous (it is roughly equivalent to the time required to iterate over a set of backup backing map entries and move them to the primary backing map in the same JVM). But certain factors can increase this duration drastically (to several seconds): large partitions, sufficiently slow single-threaded CPU performance, many or expensive indexes to rebuild, etc. The solution of course is to mitigate each of those factors but in many cases this may be challenging. Larger clusters also lead to the temptation to place more load on the available hardware resources, spreading CPU resources thin. As an example, while we've long been aware of how garbage collection can cause significant pauses, it usually isn't viewed as a major consumer of CPU (in terms of overall system throughput). Typically, the use of a concurrent collector allows greater responsiveness by minimizing pause times, at the cost of reducing system throughput. However, at a recent engagement, we were forced to turn off the concurrent collector and use a traditional parallel "stop the world" collector to reduce CPU usage to an acceptable level. In summary, there are some less obvious factors that may result in excessive CPU consumption in a larger cluster, so it is even more critical to test at full scale, even though allocating sufficient hardware may often be much more difficult for these large clusters.

    Read the article

  • A good way to build a game loop in OpenGL

    - by Jeff
    I'm currently beginning to learn OpenGL at school, and I've started making a simple game the other day (on my own, not for school). I'm using freeglut, and am building it in C, so for my game loop I had really just been using a function I made passed to glutIdleFunc to update all the drawing and physics in one pass. This was fine for simple animations that I didn't care too much about the frame rate, but since the game is mostly physics based, I really want to (need to) tie down how fast it's updating. So my first attempt was to have my function I pass to glutIdleFunc (myIdle()) to keep track of how much time has passed since the previous call to it, and update the physics (and currently graphics) every so many milliseconds. I used timeGetTime() to do this (by using <windows.h>). And this got me to thinking, is using the idle function really a good way of going about the game loop? My question is, what is a better way to implement the game loop in OpenGL? Should I avoid using the idle function?

    Read the article

  • From Transactions To Engagement

    - by David Dorf
    I've mentioned in the past that Oracle has invested quite a bit in acquiring social companies to build out its Social Relationship Management suite.  The concept is to shift away from transactions and towards engagement.  Social media represents a great opportunity to engage with customers, learn what they want, and personalize the shopping experience for them. I look at SRM as the bridge between traditional CRM and CX.  If you're looking for ideas, check out Five Social Retailing Suggestions and Social Analytics and the Customer.  There are lots of ways to leverage social media to enhance the customer experience and thus drive more sales. My friends over at 8th Bridge have just released their Social IQ report in which they rate retailers on their social capabilities.  They also produced a nice infographic so you can consume the data quickly, but I'd still encourage you to download the full report. Retailers interested in upping their SRM abilities should definitely stop by the Oracle booth at NRF in January.

    Read the article

  • Infrared usb adapter not hot pluggable?

    - by A. Goossens
    For transferring exercise files from my Polar heart rate monitor i use the software Protrainer 5 with an external infrared usb adapter. Using Wine to run Protrainer 5 and some adjustments mentioned here (using irda-utils) i managed to make it work. Well, sort of: every time i want to transfer exercise files and the adapter is not plugged in i have to reboot my laptop with the usb adapter plugged in to make it work. Since i have a laptop it unfortunately is no option to just always leave the adapter in. It seems to me that the usb adapter isn't hot pluggable (which seems odd for a usb device) or there must be another problem. Polar WebLink (other software to transfer exercise files) experiences the same issue. Can anyone help me to fix this problem? Infrared usb adapter: Polar IRDA USB adapter OS: Ubuntu 12.10 (64-bit) Hardware: Lenovo Thinkpad T500

    Read the article

  • NEW: Oracle Certification Exam Preparation Seminars

    - by Harold Green
    Hi Everyone, I am really excited about a new offering that we are announcing this week - Oracle Certification Exam Preparation Seminars. These are something that will make a big difference for many of you in your efforts to become certified and move your career forward. They are also something that have previously only been available (but very popular) to the limited number of customers who have attended our annual conferences in San Francisco (Oracle OpenWorld and JavaOne). These are the first in a series of offerings that we are releasing over the next few months. So for those of you either preparing or considering Oracle certification - keep watching here on the blog, Facebook, Twitter and the Oracle Certification website for additional announcements related to our most popular certification areas. Details of the new Exam Preparation Seminars are found below: NEW: ORACLE CERTIFICATION EXAM PREPARATION SEMINARS Becoming Oracle certified is a great way to build your career, gain additional credibility and improve your earning power. We know that the decision to become certified is not trivial. Our surveys indicate that people consider their time investment a critical factor in their decision to become certified. Your time is important. In order to help candidates maximize the efficiency of their study time we are releasing a new series of video-based seminars called Exam Preparation Seminars. These seminars are patterned after the extremely popular Exam Cram sessions that until now have only been available at our annual customer conferences (Oracle Open World and JavaOne). Beginning today they are now available to anyone, anywhere as a part of this Exam Prep Seminar series. Features: Fast-paced objective by objective review of the exam topics - led by top Oracle University instructors 24/7 access through Oracle University's training on demand platform. Ability to re-watch all or part of the the seminar. All the conveniences of video-based training: start, stop, fast-forward, skip, rewind, review. Tips that will help you better understand what you need to know to pass the exam. The Exam Preparation Seminars are meant to help anyone with a working knowledge of the technology get that extra boost to help them finalize their preparation, and will help anyone who wants a better understanding of the the depth and breadth of the exam topics and objectives. Benefits: Save time by understanding what you should study. Makes you efficient because you will understand the breadth and depth of each of the exam topics. Helps you create a better, more efficient study plan. Improves your confidence in your skills and ability to pass the certification exam. Exam Preparation Seminars are available individually, or in convenient Value Packages (which include the Exam Preparation Seminar, and an exam voucher which includes one free-retake if you need it). Currently we are releasing two seminars - one for DBA SQL and one for DBA Administration I. Additional offerings are in process. Find out more: General WEB: Oracle Certification Exam Preparation Seminars VIDEO: Exam Preparation Seminars Promo (1:27) Oracle Database Administration I (11g, 10g) VIDEO: Instructor Introduction (1:08) VIDEO: Sample Video (2:16) Oracle Database SQL VIDEO: Instructor Introduction (1:08) VIDEO: Sample Video (2:16)

    Read the article

  • Advertising Campaign Bidding Network

    - by David
    I hope this is the correct stackexchange to post on. Seems to most relevant. I'm looking for more ways to monetize my site. I'm already using the usual Google Adsense, tradedoublers, ad networks, etc. What I want to know, is if anyone knows of a service/network out there (similar to tradedoubler), which allows you to accept (or bid) on campaigns (primarily CPM based). For example: I log in, I get a list of advertising campaigns that are available and the prices they are willing to pay or I can put a bid in with imp count and cpm rate. I'm looking for something ideally europe or UK based, but im open to others. Thanks!

    Read the article

  • Wireless speed drops over time and doesn't get up if not reconnected

    - by Vili Lehto
    I am using Ubuntu 12.10 64bit and I have a Asus PCE-N15 wireless card(in PCI-E slot). The problem is that first when I connect to my WiFi network the speed is just fine, and actually my link speed never drops, it shows a solid 150mb/s+ link speed and good signal. But the download and upload speeds drop dramatically after couple minutes of use and I have to reconnect to fix this. Network is type IEEE 802.11bgn and uses WPA2/AES encryption. I don't have similar problems on the windows side, and the network is working fine on every other device. iwconfig shows this: wlan0 IEEE 802.11bgn ESSID:"WLAN-AP" Mode:Managed Frequency:2.412 GHz Access Point: 00:1E:AB:05:EF:31 Bit Rate=300 Mb/s Tx-Power=20 dBm Retry long limit:7 RTS thr=2347 B Fragment thr:off Encryption key:off Power Management:off Link Quality=61/70 Signal level=-49 dBm Rx invalid nwid:0 Rx invalid crypt:0 Rx invalid frag:0 Tx excessive retries:0 Invalid misc:2 Missed beacon:0 So the question is: Is there a way to fix this? Thanks in advance.

    Read the article

  • A Look at Exceptions in .NET Applications

    Memory dumps are a wonderful way of finding out what caused an exception in a managed .NET application, particularly if it is happening in a production application. It is when tracking exceptions in applications where you can't use Visual Studio that the techniques of using cdb and sos.dll come into their own. They may nor be skills that you need to use regularly, but at some point, they will be invaluable. Edward supplies sample memory dumps and gives you a simple introduction.

    Read the article

  • Track those visitors who come through a particular link

    - by busybee235
    I want to track visitors who come to my site through a particular link. For example, those visitors coming from http://www.domain.com/abc123, I can get their pageviews, time on site, bounce rate, referrer pages per visit etc. After that I can store that info into by database on daily basis. Can anyone suggest any service or api or any software for the same? I have used Google Analytics utm tags that work straight well for my requirement but I don't know how many links I can track with it. I have around 80-100 links to track a day and the number of links will be increasing. I couldn't find any documentation regarding limit of campaigns in GA. If there's no such limit, I can start this project. Thanks

    Read the article

  • Oracle Fusion HCM Gains Traction and Customer Recognition

    - by Scott Ewart
    Oracle Fusion HCM Gains Traction and Customer Recognition at the HRO Summit Europe in Barcelona Audience voted Oracle Fusion Human Capital Management as best in Most Reliable, Most Innovative and Best in Class. During the annual European HRO Summit in Barcelona, HRO buyers, service providers, third party advisors and other attendees were visibly impressed with the Fusion HCM product stack. Following the “present-off” among four technology vendors, Oracle was voted first in the following categories: Which technology could best suit the needs for your company Which technology do you think came across as the most reliable Which technology offers the most innovation Based on what you heard today, which technology presentation would you rate as best in class Oracle was voted second in the two other remaining categories Click here for the full article ==> http://bit.ly/sxC3tX

    Read the article

< Previous Page | 95 96 97 98 99 100 101 102 103 104 105 106  | Next Page >