Search Results

Search found 7204 results on 289 pages for 'almost dead'.

Page 72/289 | < Previous Page | 68 69 70 71 72 73 74 75 76 77 78 79  | Next Page >

  • Using glReadBuffer/glReadPixels returns black image instead of the actual image only on Intel cards

    - by cloudraven
    I have this piece of code glReadBuffer( GL_FRONT ); glReadPixels( 0, 0, width, height, GL_RGB, GL_UNSIGNED_BYTE, buffer ); Which works just perfectly in all the Nvidia and AMD GPUs I have tried, but it fails in almost every single Intel built-in video that I have tried. It actually works in a very old 945GME, but fails in all the others. Instead of getting a screenshot I am actually getting a black screen. If it helps, I am working with the Doom3 Engine, and that code is derived from the built-in screen capture code. By the way, even with the original game I cannot do screen capture on those intel devices anyway. My guess is that they are not implementing the standard correctly or something. Is there a workaround for this?

    Read the article

  • What if you've been asked to develop a site and the client later introduces Ts&Cs that you'll breach whilst doing your job?

    - by Matt Lacey
    Disclaimer : this is all made up. Honest. And it represents no clients or employers living or dead, blah blah blah, etc. [Allegedly] As part of a website I've built, I've now been provided the Terms and Conditions of site usage to display on the site. These terms--which must be agreed to to access the site--include my (or any visitor to the sites) compliance with a number of clauses. Many of these clauses refer to general computer use and are not tied specifically to use of the site. Some of these clauses refer to things I have had to previously do as a legitimate part of my job and would expect to have to do again. When I've raised similar issues previously my line manager has said just to ignore it but that doesn't seem to be the professional thing to do. So, what do I do? Abiding by the terms would mean that I could no longer work on the project and would cause issues with my employer and the owner of the business the site is being created for. Ignoring them could lead to possible future issues with the business owner and is not something I'm necessarily happy with (the deliberate breaking of a legal contract). Neither option is one I'd choose and could have major consequences. Any thoughts?

    Read the article

  • Skynet Big Data Demo Using Hexbug Spider Robot, Raspberry Pi, and Java SE Embedded (Part 4)

    - by hinkmond
    Here's the first sign of life of a Hexbug Spider Robot converted to become a Skynet Big Data model T-1. Yes, this is T-1 the precursor to the Cyberdyne Systems T-101 (and you know where that will lead to...) It is demonstrating a heartbeat using a simple Java SE Embedded program to drive it. See: Skynet Model T-1 Heartbeat It's alive!!! Well, almost alive. At least there's a pulse. We'll program more to its actions next, and then finally connect it to Skynet Big Data to do more advanced stuff, like hunt for Sara Connor. Java SE Embedded programming makes it simple to create the first model in the long line of T-XXX robots to take on the world. Raspberry Pi makes connecting it all together on one simple device, easy. Next post, I'll show how the wires are connected to drive the T-1 robot. Hinkmond

    Read the article

  • How can I stop a bot attack on my site?

    - by tnorthcutt
    I have a site (built with wordpress) that is currently under a bot attack (as best I can tell). A file is being requested over and over, and the referrer is (almost every time) turkyoutube.org/player/player.swf. The file being requested is deep within my theme files, and is always followed by "?v=" and a long string (i.e. r.php?v=Wby02FlVyms&title=izlesen.tk_Wby02FlVyms&toke). I've tried setting an .htaccess rule for that referrer, which seems to work, except that now my 404 page is being loaded over and over, which is still using lots of bandwidth. Is there a way to create an .htaccess rule that requires no bandwidth usage on my part? I also tried creating a robots.txt file, but the attack seems to be ignoring that. #This is the relevant part of the .htaccess file: RewriteCond %{HTTP_REFERER} turkyoutube\.org [NC] RewriteRule .* - [F]

    Read the article

  • Does my approach for building a real time monitoring system make sense? [closed]

    - by sameer
    I am developing an application that will display a dashboard that will display data from different SQL databases. This needs to happen in almost real time, our refresh time is about 5 minutes. My approach so far is: Develop a Windows service to accumulate the data from various SQL Server instances. Persist those details into a SQL DB, from which the dashboard will display them on the web page. Trigger fetching of data from the Windows service will every x minutes. The details of the SQL Server instances will be stored in the SQL DB which the Windows service will be referring. Does my approach make sense?

    Read the article

  • Why subdomains of Blogspot/WordPress like sites are treated as different domains or sites?

    - by Thedijje
    As I know, maps.google.com or mail.google.com all comes under the same domain and its all are subdomain. Entire web treats these subdomain as the part of main domain and they have same Alexa rank, PageRank and all. But in another hand, take a look on blogspot.com/wordpress.com/webs.com; these are different sites but blogs or websites under those domains are treated as different sites. Its new URL, all have different PageRank and Alexa rank as well. Tts about millions of subdomains under those few domain, have almost similar IP address, hosting and CMS, still why they are called different domains?

    Read the article

  • Coherence Data Guarantees for Data Reads - Basic Terminology

    - by jpurdy
    When integrating Coherence into applications, each application has its own set of requirements with respect to data integrity guarantees. Developers often describe these requirements using expressions like "avoiding dirty reads" or "making sure that updates are transactional", but we often find that even in a small group of people, there may be a wide range of opinions as to what these terms mean. This may simply be due to a lack of familiarity, but given that Coherence sits at an intersection of several (mostly) unrelated fields, it may be a matter of conflicting vocabularies (e.g. "consistency" is similar but different in transaction processing versus multi-threaded programming). Since almost all data read consistency issues are related to the concept of concurrency, it is helpful to start with a definition of that, or rather what it means for two operations to be concurrent. Rather than implying that they occur "at the same time", concurrency is a slightly weaker statement -- it simply means that it can't be proven that one event precedes (or follows) the other. As an example, in a Coherence application, if two client members mutate two different cache entries sitting on two different cache servers at roughly the same time, it is likely that one update will precede the other by a significant amount of time (say 0.1ms). However, since there is no guarantee that all four members have their clocks perfectly synchronized, and there is no way to precisely measure the time it takes to send a given message between any two members (that have differing clocks), we consider these to be concurrent operations since we can not (easily) prove otherwise. So this leads to a question that we hear quite frequently: "Are the contents of the near cache always synchronized with the underlying distributed cache?". It's easy to see that if an update on a cache server results in a message being sent to each near cache, and then that near cache being updated that there is a window where the contents are different. However, this is irrelevant, since even if the application reads directly from the distributed cache, another thread update the cache before the read is returned to the application. Even if no other member modifies a cache entry prior to the local near cache entry being updated (and subsequently read), the purpose of reading a cache entry is to do something with the result, usually either displaying for consumption by a human, or by updating the entry based on the current state of the entry. In the former case, it's clear that if the data is updated faster than a human can perceive, then there is no problem (and in many cases this can be relaxed even further). For the latter case, the application must assume that the value might potentially be updated before it has a chance to update it. This almost aways the case with read-only caches, and the solution is the traditional optimistic transaction pattern, which requires the application to explicitly state what assumptions it made about the old value of the cache entry. If the application doesn't want to bother stating those assumptions, it is free to lock the cache entry prior to reading it, ensuring that no other threads will mutate the entry, a pessimistic approach. The optimistic approach relies on what is sometimes called a "fuzzy read". In other words, the application assumes that the read should be correct, but it also acknowledges that it might not be. (I use the qualifier "sometimes" because in some writings, "fuzzy read" indicates the situation where the application actually sees an original value and then later sees an updated value within the same transaction -- however, both definitions are roughly equivalent from an application design perspective). If the read is not correct it is called a "stale read". Going back to the definition of concurrency, it may seem difficult to precisely define a stale read, but the practical way of detecting a stale read is that is will cause the encompassing transaction to roll back if it tries to update that value. The pessimistic approach relies on a "coherent read", a guarantee that the value returned is not only the same as the primary copy of that value, but also that it will remain that way. In most cases this can be used interchangeably with "repeatable read" (though that term has additional implications when used in the context of a database system). In none of cases above is it possible for the application to perform a "dirty read". A dirty read occurs when the application reads a piece of data that was never committed. In practice the only way this can occur is with multi-phase updates such as transactions, where a value may be temporarily update but then withdrawn when a transaction is rolled back. If another thread sees that value prior to the rollback, it is a dirty read. If an application uses optimistic transactions, dirty reads will merely result in a lack of forward progress (this is actually one of the main risks of dirty reads -- they can be chained and potentially cause cascading rollbacks). The concepts of dirty reads, fuzzy reads, stale reads and coherent reads are able to describe the vast majority of requirements that we see in the field. However, the important thing is to define the terms used to define requirements. A quick web search for each of the terms in this article will show multiple meanings, so I've selected what are generally the most common variations, but it never hurts to state each definition explicitly if they are critical to the success of a project (many applications have sufficiently loose requirements that precise terminology can be avoided).

    Read the article

  • Is "PHP and MySQL Web Development, 4th ed." an outdated book to learn from?

    - by ForeverLearnNeverMaster
    I want to get into web stuff, and thought of going with PHP+MySQL. I have experience in C#, C++, desktop graphics, but no web stuff besides HTML, CSS which I started learning 2 months ago. On SO PHP-book question, the highest rated book is "PHP Objects, Patterns and Practice". However, I'm not sure if it's a book to start PHP with... or is it? Most of the books mentioned in that SO question seem to be for those who already know PHP. I considered "PHP & MySQL Web Development, 4th Ed.", but someone mentioned that "it uses the mysql_* functions which are almost deprecated." Help?

    Read the article

  • Ubuntu 11.10 Unity - all Windows decorations disappear after I maximize any window

    - by korda
    When I maximize some of the windows all decorations disappear (by all, I mean on all windows)... Is that common issue on Unity or I'm just unlucky to have some prone to that bug configuration? Anyone have idea how to fix this? EDIT: It won't fix after unmaximize. It seems like maximazing window simply crashes window decorator. Decoration isn't displayed for all existing windows and any new ones. Only way I found to fix this is to run compiz --replace (but this ruins current windows placement - all windows end up on same desktop). It happens almost every time I maximize window.

    Read the article

  • How do I make my volume indicator operate in decibels instead of percentage?

    - by ethana2
    When I want to adjust the volume of anything I'm doing, I find that using the volume controls built into Ubuntu is little but confusion. When the volume is around 100%, dropping it several increments has almost no effect on apparent volume, but when it's around 0%, the effect of one click of my mouse wheel is probably a good 3 decibels. I have observed this behavior on tens of different UC's, since I convert about one Ubuntu user a month (NE team contact). This has proven so frustrating to me that I tend to use the volume knob on my guitar amp ( mono audio :| ) instead of the volume indicator. What can I do to make my volume indicator behave properly until this is fixed? I want each volume increment to be one half or one third decibel. Is there a different piece of software I should use for system volume configuration perhaps?

    Read the article

  • google earth 7, 32bit in 12.10 runs without error but there is no image (globe)

    - by Dennis
    Everything seemed to install fine. I can start google earth and all layers are available, I can even zoom in and look at 3-D buildings. But there is absolutely no image data displayed at all. If you look at the whole globe the outlines are there on an invisible globe. As you zoom in the base looks dark grey almost black but there is no image. I have tried. Tools Options Graphics Safe Mode Tools Options Texture colors all combinations Tools Options Cache (tried several changes to the numbers) lspci shows Display controller: Intel Corporation Mobile 915GM/GMS/910GML Express Graphics Controller (rev 03) Running on a Dell Inspiron 6000 laptop (1.5Gb memory)

    Read the article

  • wireless not connecting in 12.04 update

    - by Julie
    I just updated from 10.04 to 12.04. I am not very knowledgeable at all with Ubuntu since it is still relatively new to me. I went through the whole update today and yesterday and now that it is finished it will not acknowledge that there are any wireless connections around, nor does it give me an option to pick one. it just tells me what ones I've used in the past. is there any way that someone can help me? I have read through and attempted almost every fix that everyone has found and nothing is working and I'm super confused.

    Read the article

  • Why do different browsers return different search results at Google and how can I prevent it?

    - by Sei
    I am running some websites and am constantly checking keywords' rankings on google.com. And it is really important for us to see the organic search result without logging in or setting a specific location. Since this morning my colleague and I have checked the same ranking on both IE and firefox, the result surprisingly is very different (it almost feel like IE was logged in because the ranking is much higher, while in reality it is not). I have changed computer and the same problem occurred. It did not happen before. Can anyone tell me why is it?

    Read the article

  • Verify uniqueness of new content

    - by rogerkk
    I'm working on a review site, where there is a minor issue with almost duplicate reviews across items. Just a few words are changed. It would be very nice to be able to uncover these duplicates before they are approved by a moderator, and I'm hoping someone could chime in on the best strategy to get there. The site is running Ruby on Rails on a Postgres database and using Thinking Sphinx for search (all on Heroku), and so far the best option I see is to be pulling all the reviews out of the db and using a module like amatch to compare the strings. Not very efficient, so in this case I guess I'll have to limit the number/age of reviews to scan for dupes. Anyone got a better idea?

    Read the article

  • How to fix Bluetooth Sound in Ubuntu 11.10

    - by Cerin
    Bluetooth sound worked perfectly for me in Ubuntu 10.04. I recently upgraded to 11.10, and although sound in general works fine, Bluetooth sound almost never works. My headset is paired, and Sound Settings shows it's selected as the output device, but no sound is being output. I've tested the headset with other devices, and confirmed it still works, so the issue is definitely with 11.10. I'm not seeing any explicit errors and I can't find any similar issues with 11.10 through Google. I tried restarting the bluetooth service, but that had no effect. What can I do to fix this?

    Read the article

  • Gzip compress offline?

    - by shoosh
    I've configured my site to serve compressed content by putting this line in .htaccess AddOutputFilterByType DEFLATE text/html text/plain text/xml text/javascript text/css application/javascript application/json This works perfectly for almost all files except a few large JSON files that are above 200Kb. For some reason they are not being compressed. I see that they don't using the net tab in firebug and the Network section in chrome. So as a workaround I thought I could compress these files offline and have Apache read them compressed. What tool should I use to compress them? is the linux gzip the one? any special flags or something I should use? What should I put in .htaccess so that the server would know to serve these files with content-encoding gzip ?

    Read the article

  • What are the system requirements for each flavor of Ubuntu Desktop?

    - by Braiam
    I'm thinking about installing Ubuntu Desktop, but I don't know what flavor is the better for my system. What are the minimum and recommended hardware requirements? What kind of CPU? How much memory? Should I have Hardware Acceleration? What flavor should I use? This is an attempt of a canonical answer. My answers have the "official minimal requirements", the recommended are a mix of official sources and opinion based (along with the answer it's told the source). You can comment or edit if you feel that the information is obsolete or incomplete. Is a good rule of thumb that any system capable to run Windows Vista, 7, 8 x86 OS X will almost always be a lot faster with any Ubuntu flavor even if they are lower-spec than described below.

    Read the article

  • How do I run Conkeror in Emacs?

    - by Anake
    I have been using emacs for about 3 months, and have been amazed by how much improved my interactions with my computer are. I now use eshell rather than a separate terminal, and the last thing I still need to leave my emacs environment for is for my web browser. I have been reading up on Conkeror, and it seems almost ideal (emacs keybindings, no need to use the mouse etc.) but it runs in a separate window. If it could be run within emacs, it would mean that I would never really have to have interaction with the actual OS, which would be good for someone who is forced to rotate between linux, mac and windows (i.e. me). I understand that it couldn't be run from emacs in -nw mode, but would it be possible (or likely) that Conkeror could be setup to run within an emacs window? Note: - I've tried w3m and didn't really like it

    Read the article

  • Old Lock Retrofitted for Wireless and Key-free Entry

    - by Jason Fitzpatrick
    What do you do if the old key your landlord gave you is poor fit for your apartment’s lock? If you’re the geeky sort, you build a wireless unlocking module to do the work for you. Instructables user Rybitski writes: The key to my apartment never worked quite right because it is a copy of a copy of a copy. I am fairly certain that the dead bolt is original to the building and the property manager seems to have lost the original key years ago. As a result unlocking the door was always a pain. Changing the lock wasn’t an option, but eliminating the need to use a key was. To that end, he built the device seen in the video above. An Arduino Uno drives a servo which in turn opens the deadbolt. The whole thing is controlled by a simple wireless key fob. Hit up the link below for the full build guide including code. Key Fob Deadbolt [via Hack A Day] How To Delete, Move, or Rename Locked Files in Windows HTG Explains: Why Screen Savers Are No Longer Necessary 6 Ways Windows 8 Is More Secure Than Windows 7

    Read the article

  • Is SSD TRIM support still automatic in 12.10?

    - by dam
    Folks I had automatic TRIM working on my laptop running Ubuntu Precise. As in the TRIM guide I added discard to mount options in /etc/fstab, and hdparm --read-sector read 0s immediately after a rm && sync. Using the very same hardware, laptop and SSD, TRIM seems no longer to be automatic after upgrading to Quantal. I recognise the test in the guide I mentioned above may not necessarily work. SSD erase blocks and all that. But Quantal is at least different. After deleting the file and syncing, its data are still on disk and unerased even after waiting several minutes. fstrim will then 0 the dead file's blocks. Once. Repeat the same test five minutes later, and fstrim does nothing. I figure this is probably really a kernel issue, but that box is too black for my spelunking torch. I'm prepared to believe that kernel 3.5 knows what I want better than I do, and all is well despite appearances, but it looks for all the world like TRIM isn't quite all there any more. Anybody have the scoop on TRIM in Quantal/kernel 3.5?

    Read the article

  • Mapping Object Relationships - QuickStart with NHibernate (Part 3)

    - by BobPalmer
    For this third tutorial, we'll be introducing users new to NHibernat to basic object relationships, starting with a simple many-to-one relationship.  I decided that it would make sense to at least get the readers through some basic relationship mapping (including varieties of parent/child and many to many relationships) before diverging into UI, since most folks are looking for enough to bootstrap themsevles into using NHibernate, and this almost always means some kind of relation between their objects. You can find a link to the article at: http://docs.google.com/Doc?docid=0AUP-rKyyUMKhZGczejdxeHZfMjJmM3c3M3Bnbg&hl=en As always, comments, corrections, and suggestions are appreciated! -Bob

    Read the article

  • Review before or after code commit, which is better?

    - by fifth
    Traditionally we performed code review before commit, I had an argument with my colleague today, who preferred code review after commit. First, here's some background, we got some experienced developers and we also got new hires with almost zero programming practice. we'd like to perform fast and short iterations to release our product. we all team members locate at same site. The advantages of code review before commit I've learned, mentor new hires try to prevent errors, failures, bad designs in early developing cycle learn from others knowledge backup if someone quits But I also got some bad experience, like low efficiency, some changes may be reviewed over days hard to balance speed and quality, especially for newbies some guy felt distrust As to post-review, I just knew little about this, but the most thing I worried about is the risk of losing control, people never review. Any opinions?

    Read the article

  • Red Gate join the SSIS custom component club

    I recently noticed that Red Gate have launched themselves into the SSIS component market by releasing a new Data Cleanser component, albeit in beta for now. It seems to be quite a simple component, bringing together several features that you can find elsewhere, but with a suitable level  polish that you’d expect from them. String operations include find and replace with regular expressions, case formatting and trim, all of which are available today in one form or another, but will the RedGate factor appeal to people? Benefits include ease of use, all operations in one place, versus installing a custom component which many organisations do not like. I’m also interested to see where they take this and SSIS products in general, as it almost seems too simple for RedGate, a company I normally associate with more advanced problem solving. Perhaps they are just dipping a toe in the water with a simple component for now?

    Read the article

  • Programming curricula

    - by davidk01
    There are a lot of schools that teach Java and C++ but whenever I see the syllabus for one of these classes it's almost always some cut and dry OO stuff with possibly some boring end of class project. With all the little gadgets and emulators for those gadgets why aren't more schools re-purposing those classes so that the students work their way up to building android or meego applications? That way students get to experience first hand what it takes to engineer/build a piece of software instead of doing finger exercises with syntax. Practically every self-taught programmer that I know started programming because they wanted to make their gadgets do things for them. They didn't learn a programming language with an abstract conception of using it on some far distant project so I don't understand why schools don't emulate this style of teaching.

    Read the article

  • Looking for the better way to combine deep architecture refactoring with feature based development

    - by voroninp
    Problem statement: Given: TFS as Source Control Heavy desktop client application with tons of legacy code with bad or almost absent architecture design. Clients constantly requiring new features with sound quality, fast delivery and constantly complaining on user unfriendly UI. Problem: Application undoubtedly requires deep refactoring. This process inevitably makes application unstable and dedicated stabilization phase is needed. We've tried: Refactoring in master with periodical merges from master (MB) to feature branch (FB). (my mistake) Result: Many unstable branches. What we are advised: Create additional branch for refactoring (RB) periodically synchronizing it with MB via merge from MB to RB. After RB is stabilized we substitute master with RB and create new branch for further refactoring. This is the plan. But here I expect the real hell of merging MB to RB after merging any FB to MB. The main advantage: Stable master most of the time. Are there any better alternatives to the procees?

    Read the article

< Previous Page | 68 69 70 71 72 73 74 75 76 77 78 79  | Next Page >