Search Results

Search found 15208 results on 609 pages for 'programming history'.

Page 372/609 | < Previous Page | 368 369 370 371 372 373 374 375 376 377 378 379  | Next Page >

  • "A good programmer can be as 10+ times more productive than a mediocre one"

    - by m3th0dman
    I had read an interview with a great programmer (it is not in English) and in it he said that "a great programmer can be as 100 times as good as a mediocre one" giving reason for why good programmers are very well paid and why programming companies give many facilities for their employees. The idea was that there is a very large demand for good programmers, because of the above reason and that's why companies pay very much to bring them. Do you agree with this statement? Do you know any objective facts that could support it? Edit: The question has nothing to do with experience; if you talk about one great programmer with 1 year experience then s/he should be 10 times more productive than a mediocre programmer with 1 year experience. I agree that from certain experience years onwards, things start to dissipate but that's not the purpose of the question.

    Read the article

  • Is there an industry standard for systems registered user permissions in terms of database model?

    - by EASI
    I developed many applications with registered user access for my enterprise clients. In many years I have changed my way of doing it, specially because I used many programming languages and database types along time. Some of them not very simple as view, create and/or edit permissions for each module in the application, or light as access or can't access certain module. But now that I am developing a very extensive application with many modules and many kinds of users to access them, I was wondering if there is an standard model for doing it, because I already see that's the simple or the light way won't be enough.

    Read the article

  • C redevient le langage le plus utilisé devant Java et C++, d'après le classement des langages de pro

    Le C redevient le langage de programmation le plus utilisé Devant Java et le C++, d'après TIOBE Software TIOBE Software publie chaque mois son classement (le TIOBE Programming Community index) des langages de programmation. D'après cet index, pour la première fois depuis 4 ans, Java perd sa place de langage le plus populaire au profit du C qui retrouve donc le top du classement. Le C "est assez constant au fil des années, il varie entre 15% et 20% de parts de marché depuis presque 10 ans. Donc, la raison principale de cette place de numéro 1 n'est pas une progression du C, mais plutôt la baisse de son concurrent Java", explique l'analyse qui accompagne ce ...

    Read the article

  • Is it possible to programmatically prevent a game from pausing when its window loses focus?

    - by user836045
    I'm playing Skyrim in windowed mode and I am trying to create a bot for this game for personal use. I would like to have the bot play the game in the background, while I do other things, the only problem is that the game window pauses when it loses focus. Is there a way to make the Skyrim process think that it still has the focus, so it continues to run while I do something else on another window? I'm not a windows programming expert but would this be possible if I could somehow intercept the message that says unfocused or minimized to the process, and thus let the process think its still focused? I think Skyrim uses directx, so is it possible to come up with a solution from that end?

    Read the article

  • June 14th Webcast: Using Personalization in Oracle eAM

    - by Oracle_EBS
    ADVISOR WEBCAST: Using Personalization in Oracle eAMPRODUCT FAMILY: Enterprise Asset Management June 14, 2012 at 11 am ET, 9 am MT, 8 am PT Personalization is the ability within an E-Business Suite instance to make changes to the look and behavior of OA Framework-based pages without programming (and, therefore, upgradeable!).TOPICS WILL INCLUDE: Oracle eAM customers have leveraged Personalization to address many of the following use cases: making a non-mandatory field mandatory on a Work Order adding a new field to the header region entering the specifications for a new field disabling a field to prevent user entry adding Asset Description to Create Work Request Page viewing Asset Hierarchy in Create Work Request Page setting Auto Request Material Flag to yes by default A short, live demonstration (only if applicable) and question and answer period will be included. Oracle Advisor Webcasts are dedicated to building your awareness around our products and services. This session does not replace offerings from Oracle Global Support Services. Current Schedule can be found on Note 740966.1 Post Presentation Recordings can be found on Note 740964.1

    Read the article

  • How often do CPUs make calculation errors?

    - by veryfoolish
    In Dijkstra's Notes on Structured Programming he talks a lot about the provability of computer programs as abstract entities. As a corollary, he remarks how testing isn't enough. E.g., he points out the fact that it would be impossible to test a multiplication function f(x,y) = x*y for any large values of x and y across the entire ranges of x and y. My question concerns his misc. remarks on "lousy hardware". I know the essay was written in the 1970s when computer hardware was less reliable, but computers still aren't perfect, so they must make calculation mistakes sometimes. Does anybody know how often this happens or if there are any statistics on this?

    Read the article

  • How large should I make root, home, and swap partitions?

    - by starcorn
    Hello, I have a laptop with win7 installed. I have now made a 60gb partition which I want to install ubuntu into. The question I have, before I do the installation, is how large each of the root, swap, and home partition should be? I have read some place that root could be as small as 8GB, but isn't that too small? Since I guess beside ubuntu all the softwares installed will reside there as well? And I think I'm going to set my swap to be 2GB large. My main concern is how large should the root partition should be. I'm mainly going to use ubuntu for programming and browse the web.

    Read the article

  • APress Deal of the Day 15/Jan/2011 - Windows Mobile Game Development: Building games for the Windows Phone and other mobile devices

    - by TATWORTH
    Todays Apress $10 deal of the the day is: Windows Mobile Game Development: Building games for the Windows Phone and other mobile devices This book will provide you with a comprehensive guide to developing games for both the Windows Mobile platform and the Windows Phone using the industry standard programming languages C# and VB.NET. $49.99 | Published Apr 2010 | The book dates from the VS2005/VS2008 era but should still be useful to VS2010 developers. Given that such games development can now be done within Visual Studio. Whilst these devices use the Compact Framework instead of the standard Dot Net Framework, it is still a familiar envirnment. This book should be of benefit not just to games developers but to anyone with an interest in development for mobile phones.

    Read the article

  • Getting Started with Component Architecture: DI?

    - by ashes999
    I just moved away from MVC towards something more component-architecture-like. I have no concept of messages yet (it's rough prototype code), objects just get internal properties and values of other classes for now. That issue aside, it seems like this is turning into an aspect-oriented-programming challenge. I've noticed that all entities with, for example, a position component will have similar properties (get/set X/Y/Z, rotation, velocity). Is it a common practice, and/or good idea, to push these behind an interface and use dependency injection to inject a generic class (eg. PositionComponent) which already has all the boiler-plate code? (I'm sure the answer will affect the model I use for message/passing)

    Read the article

  • 3 SQL Join Concepts to Help You Choose the Right Join

    What do SQL joins and the "teach a man to fish" Chinese proverb have in common? SQL joins, like regular expressions, are one of those commonplace programming tasks in which true success is entirely dependent upon your ability to conceptualize the outcome. Fail to do so and you'll likely wind up spending a few hours in a frustrating round of trial and error. Like regular expressions, the proliferation of online examples has actually contributed to the frustration, providing the equivalent of a day's worth of fish rather than the proverbial fishing pool. The Future of SQL Server MonitoringMonitor wherever, whenever with Red Gate's SQL Monitor. See it live in action now.

    Read the article

  • How to debug Android App Eclipse?

    - by user2534694
    Ok. So while this isnt a programming question. I wanted to know how do people debug apps? How do you view log cat, and where these exceptions are thrown etc? And do I need to run the app on the emulator to see all the stuff, or is there a way to view this after running the app on my phone(while not being connected to the computer) Links to plugins and tips would be really helpful, as im gonna start work on my next game, and while the first one works fine, had a lot of problems while debugging.

    Read the article

  • What is a good university for computer science and game development?

    - by DukeYore
    I am starting my computer science degree at a local community college in programming using C++. However, I will be transferring to a 4-year university. Does anyone have any insight on university programs? I know Cal State Fullerton has a degree with a minor in Game Development. however, is that as important as getting a degree from a really great school? If I could shoot for something like Cal Poly would that be better? Or even Stanford or SF State being so close to so many gaming companies up there in the Bay area?

    Read the article

  • Is age a factor when looking for internships? [closed]

    - by user786362
    Possible Duplicate: Is it ever too old to learn how to become a programmer? I'm 30 years old going back to school for a 2nd degree in Computer Science. I will be transferring to my local state university this fall and would like to know if my age will be a factor when applying for internships. I have already read a few threads about age and careers: Is it too late to start your career as a programmer at the age of 30? Does it matter that you started developing at 26? While it is reassuring to know that people are getting entry-level programming jobs at 30+, what about internships? Should I even bother with bigger companies like Google, Microsoft, or Apple? I know we have laws against age-discrimination but lets not pretend we live in a perfect world where everyone follows the rules.

    Read the article

  • Book suggestions for GUI related topics [closed]

    - by asrijaal
    In the past, I've developed more on the server side, so no big GUI/graphic programming topics have crossed my way so far. I'm switching a little bit to the mobile plattforms (iOS, android) and would like to gather more info on the visual/graphic side. Using the SDK controls/widgets isn't that challenging and I would love to get deeper into the view/presentation layer. Animations, drawing itself, probably OpenGL. Any books that would take me a little deeper in understanding how drawing actually works would be helpful.

    Read the article

  • Is not being paid for training normal?

    - by user23838
    I'm a recent college graduate, and I recently had a interview with a company for an entry-level programming job. The company told me that they require two months of unpaid training for all entry level programmers. The reason given was that since they are providing free training, there wouldn't be any compensation. Is this normal? Update For others junior developers looking at this: Don't go for these type of scams. This was my first interview. I interview with 10 other companies around the area and got about 9 job offers from them. I worked for a fortune 50 company for 9 months with good pay and recently found a better opportunity for even better pay and better work. I guess moral is to be patient and have confidence in yourself.

    Read the article

  • Add Matlab to main menu

    - by Tim
    I was trying to add the installed matlab to the menu of Applications under Ubuntu 10.10. I clicked System-Preference-Main Menu - Programming - New Item, where I input the Matlab file .../MatlabR2010b/bin/matlab as the command, and selected the type to be "Application". Then I finished. But when i click the item in the menu of Applications, the Matlab icon shows up a few seconds and then nothing else happens. If I select the type to be "Application in Terminal" in the last step of adding Matlab to the menu of Applications, then when I click the item in the menu of Applications, there will be firstly a terminal window and then the Matlab command window. So I was wondering how to solve the problem of Matlab not starting when the type has been selected to be "Application"? Also is there a way to eliminate the terminal appearing when the type has been selected to be "Application in Terminal"? Thanks!

    Read the article

  • Randomization of biomes

    - by user24527
    You know how in Minecraft, the World is ever-expanding and all the biomes are randomized? My generalized question is: In relation to a space simulation that is also ever-expanding as the player moves about the world, how would one go about programming this randomization in Java? My real question is: Could I get a simplified example broken down into these example classes: astroids (This would include how many astroids there are, their positioning in space, their size, how often the larger astroids occur, how close they are to each other, the limitations of how many of the large asteroids can be in one field, how often astroid fields are generated, etc.) star-types (size, color, type, how often they occur, where hey occur, etc.) inhabitable-planets (size, positioning, how often they're generated, where they are generated, etc.) This would be very helpful currently since I wish to make a simplified version of such a program.

    Read the article

  • Most Useful New Technology?

    - by Craig Ferguson
    I'm looking to take a sort of sabbatical, and I'd love to use it to learn a new technology. My question is this: What's the most useful "new" technology for a software engineer to use? Node.js, iOS programming, Android, something else? I'd prefer to stay away from anything too new or experimental, since those are, in my experience, rarely actually used in professional production environments (for better or worse). Does anyone happen to have stats on how many jobs there are for each new technology or have anecdotes about how fun each one is? I've been using python/Django, so that's out, and it's similar to Ruby so i don't think learning Ruby would be that useful to expanding my skills. Anyone have any other ideas?

    Read the article

  • Are VM-based languages becoming viable for Graphics since the move to GPU computing?

    - by skiwi
    Perhaps the title is not the most clear, so let me elaborate it more: I am talking about VM-based languages, by that I mean languages that run on the JVM (java) and for example C#. Also I am talking about 3D graphics, just to be clear. Lately the trend has been that most computing is being done on the GPU and not on the CPU, and since times the issue with programming games on a VM-based language is that garbage collecting may happen randomly. So let's take a look which is responsible for what: Showing the graphics: GPU Uploading graphics to the GPU: CPU? Needs to be done every frame? Calculating physics constraints: GPU Doing the real game logic (Determining when to move objects (independent of physics calculations), processing AI): CPU Is my list actually correct? And if it is, is for example Java becoming more viable? Or is uploading the graphics (vertices) still the most expensive operation? Would like to get more insight into this.

    Read the article

  • Can I rightfully claim this as my own project if I recieved help online?

    - by Brad Guy
    Basically I'm new to network programming in Python, so I went on a tutorial online to find out about it. Using what was taught in the tutorial (creating a socket, connecting to ports, etc), I modified the code so that I made a program where two computers can send messages to one another. If I were to apply for a job and show this to my interviewers, would the code for it technically be mine? It is fair to say that I didn't modify the code by that much; However, what if for example I modified it into something like a tic-tac-toe game, where two users play each other from different PCs, would the code then be mine? I just don't want to look like a plagiarizer hence why I ask.

    Read the article

  • iOS NSError with global handler

    - by Sebastian Dressler
    I am in the beginning of programming an iOS app. Having read the Apple guides on how to deal with errors, I got the following most important points: Exceptions are for programmers Use NSError for the user Now, NSError is usually passed as out-argument which can then be used inside and has to be checked by the caller. However, I'm asking myself whether it is a good idea to use a global error handler, say a singleton which wraps around NSError and could be used to trigger errors and error handling from within the called function. Is there anything against that method or would it be a bad practice?

    Read the article

  • Most important research article for software development [closed]

    - by Fabian Fagerholm
    Researchers all over the world collectively publish thousands of articles on software development topics every year, hoping to benefit practical software development in the long run. (Of course, some of them only publish to increase their publication count, but hopefully most still aim to advance the field.) But what is really useful for practical software development? Of all the research articles you have read, what do you consider being the most important one for the software development field? What is it about that article that makes it stand out as especially important in your view? Note: I deliberately chose the term "software development", but you can freely interpret it as "programming" or "software engineering", or anything else that fits into the "software development" category.

    Read the article

  • Career in Artificial Intelligence [closed]

    - by Rohit S
    AI has many branches and seems like it has a bigger scope. I have seen a tutorial of Neural Networks and I'm a little confused whether Neural Networks is another branch of AI or it is a technique which is being used in branches of AI. I am mainly interested in creating software like Neural Networks that can be trained for doing a task. I like to make things automated with programming languages. So can I start with Neural Networks? And also a very important matter: what will be the scope of a job in future and in which companies?

    Read the article

  • Correct process for creating builds reliant on 3rd party packages

    - by Patrick
    I work on a Symfony 2 codebase. We use a number of third-party packages (most are in the Symfony Standard Edition). We use composer for dependencies. We current have all of our third-party code committed in our repository (after changing .gitignore files) to ensure stability. According to Proper Programming Practices™, we are not supposed to have any third-party packages in our repo. We are supposed to pull them down and include them at build time. How are we to do proper QA and debugging when at any given time our dependencies could push an update that breaks functionality?

    Read the article

  • Using packages (gems, eggs, etc.) to create decoupled architectures

    - by Juan Carlos Coto
    The main issue Seeing the good support most modern programming platforms have for package management (think gem, npm, pip, etc), does it make sense to design an application or system be composed of internally developed packages, so as to promote and create a loosely coupled architecture? Example An example of this would be to create packages for database access, as well as for authentication and other components of the system. These, of course, use external packages as well. Then, your system imports and uses these packages - instead of including their code within its own code base. Considerations To me, it seems that this would promote code decoupling and help maintainability, almost in a Web-based-vs.-desktop-application kind of way (updates are applied almost automatically, single code base for single functionality, etc.). Does this seem like a rational and sane design concept? Is this actually used as a standard way of structuring applications today? Thanks very much!

    Read the article

< Previous Page | 368 369 370 371 372 373 374 375 376 377 378 379  | Next Page >