Search Results

Search found 16554 results on 663 pages for 'programmers identity'.

Page 496/663 | < Previous Page | 492 493 494 495 496 497 498 499 500 501 502 503  | Next Page >

  • What are some tasks you might ask a potential ruby developer hire to complete?

    - by opensourcechris
    I would like to start a career path as a developer. As my first Language I chose Ruby to learn. What are some types of coding I might be asked to do as part of an evaluation at a potential employer? For example on my interview... Employer: "So, you think you can be a programer!" Me: "Why yes sir, I believe I would be a fitting addition to your team" Employer: "Well that remains to be seen! To demonstrate your basic knowledge please code insert evaluation project here What would you have me do to demonstrate my skills?

    Read the article

  • SQL Why is prefixing column names considered bad practice?

    - by P.Brian.Mackey
    According to a popular SO post is it considered a bad practice to prefix table names. At my company every column is prefixed by a table name. This is difficult for me to read. I'm not sure the reason, but this naming is actually the company standard. I can't stand the naming convention, but I have no documentation to back up my reasoning. All I know is that reading AdventureWorks is much simpler. In this our company DB you will see a table, Person and it might have column name: Person_First_Name or maybe even Person_Person_First_Name (don't ask me why you see person 2x) Why is it considered a bad practice to pre-fix column names? Are underscores considered evil in SQL as well? Note: I own Pro SQL Server 2008 - Relation Database design and implementation. References to that book are welcome.

    Read the article

  • calling methods if previous call success

    - by New Developer
    I my c# program I have to perform 5 steps(tasks) sequentially. basically these five should execute one after the other only the previous task performed is success. Currently I have done it in following style. But this is not very good code style to follow. var isSuccess=false; isSuccess=a.method1(); if(isSuccess) isSuccess=a.method2(); if(isSuccess) isSuccess=a.method3(); if(isSuccess) isSuccess=a.method4(); if(isSuccess) isSuccess=a.method5(); How can I re factor this code. What is the best way I can follow?

    Read the article

  • What are the parameters that let businesses choose .NET or Java over other languages/frameworks? [on hold]

    - by Othman
    Some of the biggest enterprise applications such as HR software, Sales and ERP, are written in .NET or Java. Also, Governments online solutions such as paying parking fines, and universities courses registration systems are all in .NET or Java. On the other hand, Facebook, Google and Wikipedia, are not using .NET or Java so much (Google uses at least both Java and Python). Twitter also was using Ruby for a long time, as well as Python. These websites work on much more data and at larger scales in terms of users and performance than any enterprise applications, yet still these companies have chosen a different path. What are the parameters driving such decisions? Note This question is not about why do you prefer x over y! or why those people are using this. The question is primarily asking about the parameters that makes Java Or .NET becomes better suite in enterprise applications based on Performance, Reliability, Scalability etc.

    Read the article

  • Should back end processes be included in use cases in requirements document?

    - by bizso09
    We're writing a requirements document for our client and need to include the use cases of the system. We're following this template: ID Description Actors Precondition Basic Steps Alternate Steps Exceptions Business validations/Rules Postconditions In the Basic Steps section, should we include steps that the system performs in the back end or should we only include steps that the user directly interacts with? Example: Basic Steps for Search 1: User goes to search page User enters term User presses search System matches search term with database entries System displays results vs Basic Steps for Search 2: User goes to search page User enters term User presses search System displays results

    Read the article

  • is requiring a video player download acceptable

    - by wantTheBest
    Our site currently is going to require our users to download a player to view videos they will want to view on our site. The videos get uploaded by users from various sources (smartphones in 3gp format for example). However most people have Flash on their machines. I am trying to 'make a gentle stand' and tell the team that requiring a download of a video player is not acceptable. My thinking is this: instead of allowing people to upload 3gp and other formats then re-serving the exact format on REQUESTs from our site's users we will instead use a video converter such as FFMpeg to convert every uploaded video to FLV for viewing on flash. so when a user requests to view one of the videos on our site -- boom they probably already have Flash installed so we just play the video in their Flash player. I feel serving up FLV flash video is best. Does it ring true that requiring, say, a 3gp player download just to view a video is the wrong approach?

    Read the article

  • Using T[1] instead of T for functions overloaded for T(&)[N]

    - by Abyx
    The asio::buffer function has (void*, size_t) and (PodType(&)[N]) overloads. I didn't want to write ugly C-style (&x, sizeof(x)) code, so I wrote this: SomePacket packet[1]; // SomePacket is POD read(socket, asio::buffer(packet)); foo = packet->foo; But that packet-> looks kinda weird - the packet is an array after all. (And packet[0]. doesn't look better.) Now, I think if it was a good idea to write such code. Maybe I should stick to unsafe C-style code with void* and sizeof? Upd: here is another example, for writing a packet: SomePacket packet[1]; // SomePacket is POD packet->id = SomePacket::ID; packet->foo = foo; write(socket, asio::buffer(packet));

    Read the article

  • what to learn & practice to be a programmer (not for Job) [closed]

    - by Hanan N.
    I want to be able to write programs (tools) for my own needs, and I want to tell the computer what to do, not to be strict to what other's told it to do. I have a good knowledge of Linux & Windows, I am also have a bit of Python in my head, but I think that I need some guidance on what languages or practices to invest my time into in order to be able to tell the computer what to do. I know there is a lot of posts similar to what I am asking here, but there is a big difference and so I have opened this thread. My goal isn't to work at a company, so I don't need to learn how to work as a team, I don't need to learn as much how to maintain my code nor how to write it nice for others to understand (I know it is important but not for me). I don't want to learn it as a profession, but as a serious hobby (you can call it hacking as long as you know what cracking means), my end goal is to be able to understand other's people code, and to write my own code with a lot of deep knowledge of the ins and outs of the computer system in order to gain from it as much as I can. So in essence my questions are: How can I understand, in depth, computers, in order to exploit them as much as I can? What should I learn? Are there any books that I should read? Is it possible to do this, or do I have to learn every aspect of programming?

    Read the article

  • Human powered document processing

    - by Denivic
    I want to create a "Human powered document processing" website in a formed as a gaming mechanic that will prompt volunteers to weed out text missed by OCR (optical character reader) The challenge is that the OCR output often contains errors and omissions, which hampers searches. Manual corrections is needed to weed out these mistakes, so that the text become machine readable.. I want to accomplish these task by splitting dull repetitive task into tiny microtasks and distributing then to register contributors on my site. I want to digitalized pedigree books into an online database. I want to recreate recaptcha and http://www.digitalkoot.fi/en/splash all roll into one.. Help!!!! What programing languages is best needed to create a recapctch type website and is there any open source project similar to this I can use? I need a some directions to begin this project, so all and any help will be great.

    Read the article

  • Any good reason to open files in text mode?

    - by Tinctorius
    (Almost-)POSIX-compliant operating systems and Windows are known to distinguish between 'binary mode' and 'text mode' file I/O. While the former mode doesn't transform any data between the actual file or stream and the application, the latter 'translates' the contents to some standard format in a platform-specific manner: line endings are transparently translated to '\n' in C, and some platforms (CP/M, DOS and Windows) cut off a file when a byte with value 0x1A is found. These transformations seem a little useless to me. People share files between computers with different operating systems. Text mode would cause some data to be handled differently across some platforms, so when this matters, one would probably use binary mode instead. As an example: while Windows uses the sequence CR LF to end a line in text mode, UNIX text mode will not treat CR as part of the line ending sequence. Applications would have to filter that noise themselves. Older Mac versions only use CR in text mode as line endings, so neither UNIX nor Windows would understand its files. If this matters, a portable application would probably implement the parsing by itself instead of using text mode. Implementing newline interpretation in the parser might also remove some overhead of using text mode, as buffers would need to be rewritten (and possibly resized) before returning to the application, while this may be less efficient than when it would happen in the application instead. So, my question is: is there any good reason to still rely on the host OS to translate line endings and file truncation?

    Read the article

  • Career change: from programming into more human-oriented area [closed]

    - by Art
    I have been a software developer for approximately 9 years, starting with part-time work during my graduation year at uni. During these years I worked for number of companies, sometimes changing places twice or three times a year. They say it takes 10 years to reach 'expert' level, and while I don't think I am an expert by any measure and I have certainly met lots of people who are more knowledgeable, smarter and more focused than I am, I think I can safely say that I had my fair share of the whole programming trade and would like to move on to something else. Psychology and behaviour was always something I was interested in, especially the practical, applicable bits of it. Recently I've been to some communication skills training and I realised that I have been missing out on the great deal of fun stuff - how people work and communicate, especially in subconscious, non-verbal area. Currently I am thinking of making a career change - ideally to move somewhere my technical skill would still be beneficial in some shape or form, or at least could serve as a bridge while I am transitioning there, you know, the whole gradual, bit-by-bit approach versus swim-or-drown one. I would like to hear your thoughts on this matter and to learn from you what are the possible transitions I can take.

    Read the article

  • Game programming : C# or C++?

    - by Chronix
    Ok so the faq says : "What language should I learn next? (Unless you have a specific requirement and don't know which language meets that requirement.)" so I guess this is not against the rules of questions. So, I've decided what I really want is to do Game Programming. So the question is, as a 18 years old who wants to learn self taught programming, what is the most suited programming language between C# and C++? ( I should state that I don't care about unix because I believe windows will be still the most used os ) I know the basics of C++, but none about C#. I know that C++ has more tutorials guides, dlls and stuff like that, while C# doesn't, but it's far easier to learn and to use for a single person to develop a program, but I've read even if the program is obfuscated it's easy to get the source of it. Well, I kinda want to focus on basic a lot first, I must say that I have a preference for C++ just because it feels more suitable to me, but I must consider that I work alone.. So even if I like it, it may not be the best thing to do. I am really not sure of which one to go for, i've read a lot of threads on various websites and it looks like C# is becoming more popular than C++. But yeah, that said, I also specified I want to do Game Programming. So I need to know some better points than just "C# is easier because of .net memory handling while C++ isnt" because I couldn't just find it. I hope the thread won't be closed because I've read the faq and I have a specific requirement. Thank you, and if you need more details you're free to ask me, but I think just by saying game programming and that I'd work alone should be enough!

    Read the article

  • Is it a bad idea to ask an interviewer what the greatest strength and weakness of their development team is?

    - by epignosisx
    I was wondering if this was a good question to ask a possible employer when interviewing for a developer position: What is the greatest strength and weakness of your development team? We all get this question when we are in an interview, so why not ask them in return? I think it is a very good question because we could find out about the team, and how this strength or weakness could affect us, but I don't want to annoy the interviewer. Is there any downside to asking this question when interviewing for a developer position?

    Read the article

  • Is it viable to become a contract programmer straight out of college?

    - by M G
    I have a Bachelor of Science in Computer Science and four months research experience designing and implementing a research project. I realize this is highly dependent on my skill set - which includes C, C++, Java, Python, and SQL. I feel I have an advantage in two ways: I am young and am not afraid to work overtime. I am willing to take lower pay to gather a client base/experience, and work nights/weekends to get a few projects under my belt. This may be cliche, but I feel that I can learn new technologies quicker than most. At the very least, I am not a slow study. With this being said, is it viable for me to become a contract programmer? Or do I need the 10+ year skill set that most contractors bring to the table?

    Read the article

  • Is it possible to keep only one Database for both web and desktop applications?

    - by B4NZ41
    I'm experiencing a trouble with my business model, let me explain better. I'm developing a software for 1 year and few months, it's for the food industry, more exactly a software to: Delivery, Take Way, Table Reservation, POS, Accounts Payable and Receivable, Prints(receipt), Kitchen Monitors Orders, Customers Orders Control and Fiscal Area. Well, I had separated the software mainly in two areas, one is web area and the other is desktop area (Used by Admins only) and local installed. 1 - Web Area (Basically do the follow:) Show Catalog with the products Customers Make Orders Customers Pay for the Orders etc ... as mentioned above 2 - Desktop Area Manage Orders Manage Customers Manage Suppliers Manage Accounts Payable and Receivable etc ... as mentioned above The web area is hosted in an online web server (scripts and database are online). The Desktop area is hosted locally in a Linux machine with a local database and local scripts files. My question is: Is it possible to keep only one Database for both applications? If YES, please what is the best approach? Follow my technical specification environment Database: Actually I have two databases working and I would love to keep only one. Operating System: Linux (Kernel 2.6.X and above) or Windows (XP and above) For the Web Area Apache, PHP, Python, Java Script, Shell Script and MySQL. For the Desktop Area: PHP-GTK2, Apache, PHP, MySQL and Shell Script.

    Read the article

  • Laser range finder, what language to use? Beginner advice

    - by DrOnline
    I hope this is the right place. I am a programming beginner, and I want to make a laser range finder, and I need advice about how to proceed etc. In a few weeks I will get a lot of dirt cheap 3-5V lasers and some cheap usb webcams. I will point the laser and webcam in parallel, and somehow use trigonometry and programming to determined distance. I have seen online that others made done it this way, I have purposefully not looked at the details too much because I want to develop it on my own, and learn, but I know the general outline. I have a general idea of how to proceed. The program loads in a picture from the webcam, and I dunno how images work really, but I imagine there is a format that is basically an array of RGB values.. is this right? I will load in the red values, and find the most red one. I know the height difference between the laser and the cam. I know the center dot in the image, I know the redmost dot. I'm sure there's some way to figure out some range there. TO THE POINT: 1) Is my reasoning sound thus far, especially in terms of image analysis? I don't need complete solutions, just general points 2) What I need to figure out, is what platform to use. I have an arduino... apparently, I've read it's too weak to process images. Read that online. I know some C I know some Python I have Matlab. Which is the best option? I do not need high sampling rates, I have not decided on whether it should be automated or whether I should make a GUI with a button to press for samples. I will keep it simple and expand I think. I also do not need it to be super accurate, I'm just having fun here. Advice!

    Read the article

  • Using "public" vars or attributes in class calls, functional approach

    - by marw
    I was always wondering about two things I tend to do in my little projects. Sometimes I will have this design: class FooClass ... self.foo = "it's a bar" self._do_some_stuff(self) def _do_some_stuff(self): print(self.foo) And sometimes this one: class FooClass2 ... self.do_some_stuff(foo="it's a bar") def do_some_stuff(self, foo): print(foo) Although I roughly understand the differences between functional and class approaches, I struggle with the design. For example, in FooClass the self.foo is always accessible as an attribute. If there are numerous calls to it, is that faster than making foo a local variable that is passed from method to method (like in FooClass2)? What happens in memory in both cases? If FooClass2 is preferred (ie. I don't need to access foo) and other attributes inside do not change their states (the class is executed once only and returns the result), should the code then be written as a series of functions in a module?

    Read the article

  • Best practice- handling images on website

    - by Steve
    I am porting an old eCommerce site to MVC 3 and would like to take advantage of design improvements. The site currently has product images stored in 3 sizes: thumbnail, medium (for display in a list) and expanded for a zoomed look. Right now we are having to upload 3 separate images that are sized exactly right, provide 3 different names that match what the site expects, etc., it is a pain. I'd like to upload just 1 file, the large one, then let the site reduce it to needed sizes, and I'd like the flexibility to change the thumbnail and list sizes depending on user preferences, form factor (e.g. mobile, iPad, desktop), etc. so might need many copies of the same image. My question is should the image be reduced then saved several times upon upload and if so what is a good storage/naming convention? The other idea is to store just the single image but resize it programmatically before serving it to the client. Has anybody done this and what are the tradeoffs besides a few more machine cycles? How do you pass a temporary image in memory to the client (there is no URL)?

    Read the article

  • Programs and memory consumption [closed]

    - by cobie
    I have a 4gb ram macbook pro but I still run out of memory when I have chrome and a few other light weight applications open such as multiple windows of macvim. These programs are written in C/C++ so technically should be memory efficient but why do they suck up all these memory. is it just bad engineering or graphical user interfaces because I have read about incredible feats performed in software dev back in the early computing days with very limited memory but now it just feels like the applications expand to fill all my memory.

    Read the article

  • How to write a real time data acquisition program [closed]

    - by Tosin Awe
    I have to write a program in assembly language that will monitor temperature continuously, and I have no idea where to begin. The temperature must be displayed in BCD format, and the high and low set points will be programmed into the system. if the set points are exceeded then an alarm will be indicated. The low point is 20 degrees Celsius, and the high point is 24 degrees Celsius. Can somebody give me some hints on how to complete this task?

    Read the article

  • Algorithm for grouping friends at the cinema [closed]

    - by Tim Skauge
    I got a brain teaser for you - it's not as simple as it sounds so please read and try to solve the issue. Before you ask if it's homework - it's not! I just wish to see if there's an elegant way of solving this. Here's the issue: X-number of friends want's to go to the cinema and wish to be seated in the best available groups. Best case is that everyone sits together and worst case is that everyone sits alone. Fewer groups are preferred over more groups. Sitting alone is least preferred. Input is the number of people going to the cinema and output should be an array of integer arrays that contains: Ordered combinations (most preferred are first) Number of people in each group Below are some examples of number of people going to the cinema and a list of preferred combinations these people can be seated: 1 person: 1 2 persons: 2, 1+1 3 persons: 3, 2+1, 1+1+1 4 persons: 4, 2+2, 3+1, 2+1+1, 1+1+1+1 5 persons: 5, 3+2, 4+1, 2+2+1, 3+1+1, 2+1+1+1, 1+1+1+1+1 6 persons: 6, 3+3, 4+2, 2+2+2, 5+1, 3+2+1, 2+2+1+1, 2+1+1+1+1, 1+1+1+1+1+1 Example with more than 7 persons explodes in combinations but I think you get the point by now. Question is: What does an algorithm look like that solves this problem? My language by choice is C# so if you could give an answer in C# it would be fantastic!

    Read the article

  • What significant lost steps forward in side tracks should be revived in the main stream of software?

    - by C.W.Holeman II
    In reading Alan Kay's question on Significant new inventions I was coming up with answers that were not new ideas but old ones that have been passed by in the main stream. So, what significant lost steps forward that happened in a side track should be revived in the main stream of software? These would be ideas that worked well in their context but a different context appeared and became dominant the context, even though it lacked the idea. It maybe that in the new dominant context one just cannot do something or it requires great effort to be exerted.

    Read the article

  • If you had the power to remove one thing in your daily job, what would it be?

    - by Pierre 303
    With nearly 60 answers to this question it's highly likely that your answer has already been posted. Please don't post an answer unless you have something new to say This can be a particular task you find not very useful, a manager behavior, your canteen, some security rule you must comply to, tools they want you to use, commuting, schedules, a customer, a technical problem you encouter all the time, anything, ... please be creative ;) One answer per thing. Explain why. Please vote up on other answers if appropriate.

    Read the article

  • Is it reasonable to insist on reproducing every defect before diagnosing and fixing it?

    - by amphibient
    I work for a software product company. We have large enterprise customers who implement our product and we provide support to them. For example, if there is a defect, we provide patches, etc. In other words, It is a fairly typical setup. Recently, a ticket was issued and assigned to me regarding an exception that a customer found in a log file and that has to do with concurrent database access in a clustered implementation of our product. So the specific configuration of this customer may well be critical in the occurrence of this bug. All we got from the customer was their log file. The approach I proposed to my team was to attempt to reproduce the bug in a similar configuration setup as that of the customer and get a comparable log. However, they disagree with my approach saying that I should not need to reproduce the bug (as that is overly time-consuming and will require simulating a server cluster on VMs) and that I should simply "follow the code" to see where the thread- and/or transaction-unsafe code is and put the change working off of a simple local development, which is not a cluster implementation like the environment from which the occurrence of the bug originates. To me, working out of an abstract blueprint (program code) rather than a concrete, tangible, visible manifestation (runtime reproduction) seems like a difficult working environment (for a person of normal cognitive abilities and attention span), so I wanted to ask a general question: Is it reasonable to insist on reproducing every defect and debug it before diagnosing and fixing it? Or: If I am a senior developer, should I be able to read (multithreaded) code and create a mental picture of what it does in all use case scenarios rather than require to run the application, test different use case scenarios hands on, and step through the code line by line? Or am I a poor developer for demanding that kind of work environment? Is debugging for sissies? In my opinion, any fix submitted in response to an incident ticket should be tested in an environment simulated to be as close to the original environment as possible. How else can you know that it will really remedy the issue? It is like releasing a new model of a vehicle without crash testing it with a dummy to demonstrate that the air bags indeed work. Last but not least, if you agree with me: How should I talk with my team to convince them that my approach is reasonable, conservative and more bulletproof?

    Read the article

  • Would md5 hashes allow detection of synced files?

    - by codpursue
    We have to develop our own file management system in Java web application. We need to sync files between our main server and client severs and find out whether all the client server has all the latest version of files. Our files are in pdf, doc and xls format they changes every now and then as and when it is required. What we are thinking of using MD5 checksum to find out hashcode of files on Main server and store it in database. Same would be there in Client Servers database. After comparing records on database we would come to know whether client servers are synced or not. Please suggest if there are any better ways to do the same.

    Read the article

< Previous Page | 492 493 494 495 496 497 498 499 500 501 502 503  | Next Page >