Search Results

Search found 7513 results on 301 pages for 'actual'.

Page 28/301 | < Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >

  • Multiple database accesses or one massive access?

    - by DudeOnRock
    What is a better approach when it comes to performance and optimal resource utilization: accessing a database multiple times through AJAX to only get the exact information needed when it is needed, or performing one access to retrieve an object that holds all information that might be needed, with a high probability that not all is actually needed? I know how to benchmark the actual queries, but I don't know how to test what is best when it comes to database performance when thousands of users are accessing the database simultaneously and how connection pooling comes into play.

    Read the article

  • How far should an entity take care of its properties values by itself?

    - by Kharlos Dominguez
    Let's consider the following example of a class, which is an entity that I'm using through Entity Framework. - InvoiceHeader - BilledAmount (property, decimal) - PaidAmount (property, decimal) - Balance (property, decimal) I'm trying to find the best approach to keep Balance updated, based on the values of the two other properties (BilledAmount and PaidAmount). I'm torn between two practices here: Updating the balance amount every time BilledAmount and PaidAmount are updated (through their setters) Having a UpdateBalance() method that the callers would run on the object when appropriate. I am aware that I can just calculate the Balance in its getter. However, it isn't really possible because this is an entity field that needs to be saved back to the database, where it has an actual column, and where the calculated amount should be persisted to. My other worry about the automatically updating approach is that the calculated values might be a little bit different from what was originally saved to the database, due to rounding values (an older version of the software, was using floats, but now decimals). So, loading, let's say 2000 entities from the database could change their status and make the ORM believe that they have changed and be persisted back to the database the next time the SaveChanges() method is called on the context. It would trigger a mass of updates that I am not really interested in, or could cause problems, if the calculation methods changed (the entities fetched would lose their old values to be replaced by freshly recalculated ones, simply by being loaded). Then, let's take the example even further. Each invoice has some related invoice details, which also have BilledAmount, PaidAmount and Balance (I'm simplifying my actual business case for the sake of the example, so let's assume the customer can pay each item of the invoice separately rather than as a whole). If we consider the entity should take care of itself, any change of the child details should cause the Invoice totals to change as well. In a fully automated approach, a simple implementation would be looping through each detail of the invoice to recalculate the header totals, every time one the property changes. It probably would be fine for just a record, but if a lot of entities were fetched at once, it could create a significant overhead, as it would perform this process every time a new invoice detail record is fetched. Possibly worse, if the details are not already loaded, it could cause the ORM to lazy-load them, just to recalculate the balances. So far, I went with the Update() method-way, mainly for the reasons I explained above, but I wonder if it was right. I'm noticing I have to keep calling these methods quite often and at different places in my code and it is potential source of bugs. It also has a detrimental effect on data-binding because when the properties of the detail or header changes, the other properties are left out of date and the method has no way to be called. What is the recommended approach in this case?

    Read the article

  • Provisioning a New SQL Server Instance – Part Two

    So how should you install and configure SQL Server 2012 properly? Glenn Berry completes his two-part series by explaining the steps needed to complete the preparation and do the actual installation. Keep your database and application development in syncSQL Connect is a Visual Studio add-in that brings your databases into your solution. It then makes it easy to keep your database in sync, and commit to your existing source control system. Find out more.

    Read the article

  • What snapshot software do you recommend? [closed]

    - by Charbel
    Possible Duplicate: What screenshot tools are available? Hi, is there a snapshot software for Ubuntu? Something like SnagIt? The idea is many times I have to take the snapshot and edit the image to crop to my region of interest, SnagIt does that automatically and very nicely. Another feature is to take a snapshot of text (that can't be otherwise copied - like text in a photo) and then parse it into actual text document using an OCR technology. Thanks

    Read the article

  • Password protected website

    - by danie7L T
    I need to add a user authentication page before the actual homepage of the website. In Joomla! I just set the website offline and the offline page loads in place of the homepage but then it's automatically throwing a "503 Service Temporarily Unavailable" error which I would like to avoid. That's why I would like to know the other ways to load an authentication page before the homepage. NB: I'm using Apache servers if that's relevant Thank you

    Read the article

  • Avoid pagination pages from appearing higher than real content in SERPS

    - by WordPress Developer
    I have a gallery-type website that has about 20k of pages and naturally it uses pagination. However, sometimes /page/2 appears higher in search results than /post/201339 for example. I'd like to give emphasis to the actual content (posts, videos, whatever the site is about) and not on pages that merely list this content in a paginated matter. What is the best way to avoid this issue? Maybe a NOINDEX,FOLLOW meta tag on the paginated pages?

    Read the article

  • How to appear very professional on my first freelance project?

    - by iamserious
    I need some help on appearing very professional on my first freelance project. How / What should you do to achieve this? Background: I started as a full time web developer 18 months ago, two promotions later I am now a senior software engineer. I've never had any problems with designing / developing / coding a complicated system. I thought I could use some help for Christmas and I started bidding for a project and now I have one - from a very reputable lawyers association in London. I have no problem dealing with the actual implementation of the system, but I have no idea how to appear professional throughout the whole process. About the project: This lawyers association are starting a distant training courses and in addition to having a website to show off all their clients etc, they want a students area where their students would log in, download course materials allocated to them etc.. and an admin section where they assign courses to students / create new ones / upload materials etc.. Questions breakdown: 1) How should I start with the requirement gathering? Is using scrum a good idea, or should I use something like Volere Template - and what should I do with it? should I submit a copy to the client etc.. 2) How often should I meet the client? Would once a fortnight would be good? 3) What are the processes / protocols that I need to follow so that they would be satisfied with me and think that I am very professional 4) How much should I charge for the product? 5) How should I get a quote / contract / receipt for the whole project? 6) What are the steps that professional freelancers go through, during the life cycle of a project? My Research so far Looking at How much should I charge doesn't help.. I live in London, zone 1, though I have no idea how much a project of this size would cost. Help on this would be appreciated. How to be professional articles talks about the work / time management etc and not the actual process, what would real people do etc.. it's like academia theory, but not practical. If this needs revising, please let me know, do not close it because of whatever reason, I will edit the question or details to fit the needs. Thanks for reading a lengthy question.

    Read the article

  • Book Readers As Envisioned Circa 1935

    - by Jason Fitzpatrick
    This early 20th century sketch showcases the future of books; thankfully the actual delivery of the concept proved to be a bit more lap-friendly. The sketch is from the April, 1935 issues of Everyday Science and Mechanics and presents a vision of book consumption that, thankfully, came to pass in a much more compact fashion that doesn’t require swapping rolls of film. [via Boing Boing] How To Properly Scan a Photograph (And Get An Even Better Image) The HTG Guide to Hiding Your Data in a TrueCrypt Hidden Volume Make Your Own Windows 8 Start Button with Zero Memory Usage

    Read the article

  • Is multiple domain names and links from same IP causing poor search engine rankings?

    - by John
    I have an ecommerce website which is not doing so well in Google. I am trying to improve this of course, and am looking at some possibilities for why it isn't doing well. The website has four domain names, all of which have been indexed by Google. A few months ago I applied 301 redirects to any requests for two of the domain names so now it is down to two domain names (one is a .net, the other is a .com.au, the others were .net.au and .com). I prefer to use my main domain name (the .com.au), but one of the names has been around for a long time and has more inbound links. According to a PageRank tool, both are PR2. It is a Classic ASP site and up until recently had a lot of querystring parameters. In the last week or so I added URL rewriting so there is now no parameters for most pages. I don't do 301 redirects from the old URLs but instead I add the META canonical tag indicating the preferred new URL. At the same time I redesigned the site and improved title tags, META descriptions, and H tags but it hasn't been long enough yet for Google to index many of these yet. I also looked at what pages Google has indexed and strangely it has some strange pages in the index, there are a lot of pages which are actual keyword searches (more a bunch of random letters than an actual word). What I mean is that it is as if they had typed in something to search for in my search box - there are no links to pages like this and the only way of getting this is to type something in to the search box). So I added a META robots tag with noindex,nofollow anytime that I render pages like this. Years ago I set up a fake price comparison site which lists all my products and links back to my site. It has a different keyword rich domain name but is on the same server and same IP address. It's a completely different layout but does have the same product categories and product descriptions (although I have stripped formatting out of them so they are not identical except in text). I also have a few blog sites which again are on the same server/IP and all have advertising for the website. My questions are: What should I do with the multiple domains, just use one, or continue with two or more? Should I add 301 redirects, not just the META canonical tag? Any idea about Google indexing my search results page, and did I do the right thing with the META robots tag? Is the fake price comparison site likely to be causing problems? Are all the links to the site from other domain names but the same IP address likely to be causing problems? Thanks for any help. Sorry for so many questions in one.

    Read the article

  • Can anybody recommend an application for laying out algorithms on a mac? [closed]

    - by Scotty
    In my intro to software development class I'm using a program called raptor which really helps me when I'm mapping out semi-complex algorithms and programs. The thing that I like about raptor is that when I'm flowcharting an algorithm, I can run it like an actual program and raptor steps through the code block by block. Unfortunately, raptor is only available on windows and when I'm at home I spend most of my time on Mac OS X. Is there any programs available for mac that help flowchart and step through algoritms?

    Read the article

  • What is SEO and Why Does it Matter?

    SEO, or Search Engine Optimization, is the act of optimizing your website for search engines like Google, Bing, Yahoo and others. Back in 2008 Google's blog mentions 1 trillion unique URLs, that's incredible and according to the blog that is the number of sites they were aware of and that the actual number is infinite!

    Read the article

  • ubuntu image size 732 mb - too big for cd

    - by memius
    i have an old pc that can't handle a boot stick install, so i have to create an actual, old fashioned boot cd. however, the image size for ubuntu 12.04 is 732mb, which is too large for cds, which can hold only 700mb. the maintainers of ubuntu 12.04 say the image size will never go over 700mb, and indeed, the download size seemed to be 689mb. Brasero says it won't burn the cd because the file is too big what's going on?

    Read the article

  • Is Microsoft Bing Trying To Kill Open Office?

    <b>Katonda:</b> "Microsoft Bing has many flaws, but this one seems to be the most outrageous one. If you try to search for OpenOffice on Bing, it will not show you the actual OpenOffice.org website but will show pages from random websites like OpenOffice.com or other non-OpenOffice.org websites."

    Read the article

  • Will New Horizons have to bailout?

    - by TATWORTH
    At http://pluto.jhuapl.edu/news_center/news/20121016.php, there is an interesting post about the challenge facing the New Horizons as to whether to allow the spacecraft to remain on the current trajectory which will take it between Pluto and the orbit of Charon, the closest in known moon of Pluto. Given that the current round-trip light time is 6 hrs 53 minutes, a decision to go for a bail-out fly-past must be taken some 10 days in advance of the actual fly-past.

    Read the article

  • Does OO, TDD, and Refactoring to Smaller Functions affect Speed of Code?

    - by Dennis
    In Computer Science field, I have noticed a notable shift in thinking when it comes to programming. The advice as it stands now is write smaller, more testable code refactor existing code into smaller and smaller chunks of code until most of your methods/functions are just a few lines long write functions that only do one thing (which makes them smaller again) This is a change compared to the "old" or "bad" code practices where you have methods spanning 2500 lines, and big classes doing everything. My question is this: when it call comes down to machine code, to 1s and 0s, to assembly instructions, should I be at all concerned that my class-separated code with variety of small-to-tiny functions generates too much extra overhead? While I am not exactly familiar with how OO code and function calls are handled in ASM in the end, I do have some idea. I assume that each extra function call, object call, or include call (in some languages), generate an extra set of instructions, thereby increasing code's volume and adding various overhead, without adding actual "useful" code. I also imagine that good optimizations can be done to ASM before it is actually ran on the hardware, but that optimization can only do so much too. Hence, my question -- how much overhead (in space and speed) does well-separated code (split up across hundreds of files, classes, and methods) actually introduce compared to having "one big method that contains everything", due to this overhead? UPDATE for clarity: I am assuming that adding more and more functions and more and more objects and classes in a code will result in more and more parameter passing between smaller code pieces. It was said somewhere (quote TBD) that up to 70% of all code is made up of ASM's MOV instruction - loading CPU registers with proper variables, not the actual computation being done. In my case, you load up CPU's time with PUSH/POP instructions to provide linkage and parameter passing between various pieces of code. The smaller you make your pieces of code, the more overhead "linkage" is required. I am concerned that this linkage adds to software bloat and slow-down and I am wondering if I should be concerned about this, and how much, if any at all, because current and future generations of programmers who are building software for the next century, will have to live with and consume software built using these practices. UPDATE: Multiple files I am writing new code now that is slowly replacing old code. In particular I've noted that one of the old classes was a ~3000 line file (as mentioned earlier). Now it is becoming a set of 15-20 files located across various directories, including test files and not including PHP framework I am using to bind some things together. More files are coming as well. When it comes to disk I/O, loading multiple files is slower than loading one large file. Of course not all files are loaded, they are loaded as needed, and disk caching and memory caching options exist, and yet still I believe that loading multiple files takes more processing than loading a single file into memory. I am adding that to my concern.

    Read the article

  • What to expect during an interview with a senior development executive?

    - by Umanga
    I passed first two technical interviews at a global e-commerce company for the position of senior software engineer. I was told that there are two more interviews, one with a senior development executive and another with a person from human resources (HR). What kind of questions I should expect during the interview with the senior development executive? Is is technical, high level architecture related, etc.? During HR interviews, is it ok to ask about the work-life balance and actual working hours?

    Read the article

  • Why do password strength requirements exist? [migrated]

    - by Bozho
    Password strength is now everything, and they force you to come up with passwords with digits, special characters, upper-case letters and whatnot. Apart from being a usability nightmare (even I as a developer hate it when a website requires a complex password), what are the actual benefits of having strong passwords (for website authentication)? Here are the prerequisites of a system that handles authentication properly: store passwords using bcrypt (or at least use salt+hash) - hard-to-impossible to find the original password when an attacker gets the database lock subsequent password attempts with a growing cooldown - no brute-force via the site

    Read the article

  • How to handle brightness trouble on dell inspiron i14R-2265?

    - by den-javamaniac
    Hi. Recently I've installed ubuntu 10.10 32-bit on my dell inspiron i14r-2265, but it looks like brightness control is not working. I can change it "programmatically" (FN + brightness key), though actual screen brightness shows no effect. I've tried this advice but it didn't work for me. I actually have no idea how it works (if hardware is not responsive to software, the way I see it), so can someone suggest a solution please?

    Read the article

  • Insanity&ndash;Day 1

    - by D'Arcy Lussier
    Some people do those posts about “Here’s what I’m going to do to change my life”. I don’t really like those. I’m a “don’t tell me what you’re going to do, tell me what you’re doing/have done” type of guy. So while I could say how I’m going to change my life and be healthier and happier and thinner in 60 days, I’m just going to tell you how I’m progressing through the Insanity workout. Insanity is a workout-in-a-box. It’s a collection of DVDs that you work out to. Nothing new here, except that the program is intense. It’s core tenet of the program is intensity – you workout at an elevated heart rate for 3 minutes with a short break in between, as opposed to traditional interval training where you go hard for a short time and recover for a few minutes. The other aspect of it is commitment. There is a timetable to follow – you workout 6 days a week with one rest day. This isn’t meant to be a “pick it up whenever you feel like it” type of program. The videos themselves are kind of…I don’t want to say low quality, but not as polished as I expected. Maybe that’s what they were going for. By that I mean they show shots of cameramen and the production equipment during the workout. Otherwise, it’s the Insanity leader Shawn T. leading a group of pretty fit folks through this gruelling workout. And ultimately, those little production nit-picks are irrelevant compared to the actual workout. Holy crap. I haven’t done an aerobics class in like…ever. And watching the video before my actual workout, I thought “That doesn’t look too hard”. Believe me, it is. No weights, no machines, just various exercises done in circuits and with increasing speed. By the end of the workout, I was drenched. So that was day one. Some stats just so I can track it: I’m 286 lbs at my last weigh-in a week ago or so. Should be interesting to see what 60 days of this does! D

    Read the article

  • Is it possible to measure if someone is a 'good' programmer? [closed]

    - by Alex Angas
    Possible Duplicate: How Can I Know Whether I Am a Good Programmer? There are a number of questions here about recognising or considering someone as a good/bad programmer. These are all subjective. What I'd like to know is if there is a way to measure this. I realise there will and should be a subjective element to it. But is it also possible to have some actual numbers to back up (or contradict) such an assessment?

    Read the article

  • SEO - Search Engine Optimization Tips and Techniques

    SEO can be broken down into 2 main categories, on-page optimization, and off-page optimization. On-page optimization involves the changes that are made on your actual site, like through a CMS such as MODx. Changing heading tag, title tags, alt tags, etc., are all examples of on-page optimization. Off-page involves anything and everything that isn't actually done on your site, such as social media, blogs, forums, etc.

    Read the article

  • Ubuntu 12.04 image size 732MB, will it fit a standard CD?

    - by memius
    I have an old computer that can't handle a boot stick install, so i have to create an actual, old fashioned boot CD. However, the image size for Ubuntu 12.04 is 732MB, which is too large for a CD, which can hold only 700MB. The maintainers of Ubuntu 12.04 say the image size will never go over 700MB, and indeed, the download size seemed to be 689MB. Brasero says it won't burn the CD because the file is too big, what's going on?

    Read the article

  • Should I charge for travel expenses as a contractor?

    - by Keith
    Here's a question for all my fellow contractors - I'm paid quite handsomely for my normal contracted hours (any overtime is billed at the same rate) but do you think it fair to bill for travel time to the other end of the country (regional office) when this takes place outside of the normal working day (or overlapping into the evening) as well as actual time holed up in a hotel room when you get there, ready for a normal working day the next day, along with the return journey? Petrol is claimed normally (nominal rate) and hotel is covered by the company I contract for.

    Read the article

< Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >