Search Results

Search found 8219 results on 329 pages for 'more or less'.

Page 76/329 | < Previous Page | 72 73 74 75 76 77 78 79 80 81 82 83  | Next Page >

  • What program do you use to write technical documentation?

    - by Tatu Ulmanen
    I'm writing an architecture/technical documentation for an inhouse project and I'm becoming more and more frustrated with Microsoft Word as I seem to use most of my time getting things to align correctly. Word has horrible usability issues but I didn't come here to rant, rather I'd like to know whether there exists something more suitable for the job. The point of this documentation is to provide a "big picture" of the system, and as such I'm not generating documentation from source code but rather writing it all by hand. Are there any alternatives to Microsoft Word for writing properly formatted technical documentation with less frustration?

    Read the article

  • "Best fit" to avoid reuse of object instances in a collection

    - by Simon
    Imagine I have a collection of object instances which represent activities for a user to undertake. Dependent on user attributes, I have to randomly select instances to present activities to the user. For some users, I need to present more activities to them than there are available activities in which case, I want to use the following algorithm. If all available activities have already been presented to the user, then re-select a "used" activity, selecting the earliest presented activity ordered by frequency of use. In other words, try to reduce repetition and where repetition is unavoidable, use the instances which have been repeated less often and were presented furthest back in time. Before I go on to code that algorithm, I wondered if there is some existing pattern I can re-use? [EDIT] "Furthest back in time" is not relevant as I will pass the algorithm an ordered collection of used instances where the first entry is the first presented.

    Read the article

  • What is the most reliable session storage in PHP: Memcache, database or files?

    - by user1179459
    What is the best and most safest way to handle PHP sessions. Is the best way to store sessions in: Database (more reliable, but high bottleneck, slow speed, not good for high database usage websites)? Memcache (super fast, but distributed more security problems, chances of loosing data when the server restarted and chances of loosing data when the cache is full)? Files (default option, I guess slow since it reads and writes from file I/O, less security, etc). Which method is the best? What are the problems and good things of each of those approaches?

    Read the article

  • Facebook connect vs. OpenID

    - by digit1001
    I just started working on a new project that has a general login feature. One suggestion in a meeting was to look into Facebook Connect or OpenID as an alternative. I'm curious if there's one that has less of a learning curve, or if they can both be used on the same site. Also, when you use either, do you have them initially create the account and just get a verify True/False back that you then use to set up a local user account? I what about forgotten passwords? I'm kind of curious as to best practices for integrated this type of login with a "traditional" one where you store the user info yourself. Thanks, D

    Read the article

  • What are the benefits of Android way of "saving memory" - explicitly passing Context objects everywhere?

    - by Sarge Borsch
    Turned out, this question is not easy to formulate for me, but let's try. In Android, pretty much any UI object depends on a Context, and has defined lifetime. It also can destroy and recreate UI objects and even whole application process at any time, and so on. This makes coding asynchronous operations correctly not straightforward. (and sometimes very cumbersome) But I never have seen a real explanation, why it's done that way? There are other OSes, including mobile OSes (iOS, for example), that don't do such things. So, what are the wins of Android way (Activities & Contexts)? Does that allow Android applications to use much less RAM, or maybe there are other benefits?

    Read the article

  • Why is the superblock last mount time in the future?

    - by user69541
    Future time: Since installing Ubuntu 12.04, I regularly have to fun fsck and reset my clock because it shows a 'future time' : Err.Msg. "superblock last mount time is in the future by less than a day, probably due to hardware clock being incorrectly set." FIXED. According to what I have read, it looks like I'm in the right place to get an answer. Following are my feeble attempts to rectify this annoyance: mjh@mjh-desktop:~$ sudo service ntp start [sudo] password for mjh: ntp: unrecognized service mjh@mjh-desktop:~$ mjh@mjh-desktop:~$ sudo apt-get install update Reading package lists... Done Building dependency tree Reading state information... Done E: Unable to locate package update mjh@mjh-desktop:~$ sudo service ntp start ntp: unrecognized service mjh@mjh-desktop:~$ Suggestions? Matthew

    Read the article

  • Unity: Render 2D textures on a 3D object's face

    - by www.Sillitoy.com
    I am not familiar with 3D graphics and I'd like to know what is the right way to render some 2D figures on different points of a wider face of a 3D object. My 3D object is just a cube representing a poker table. I have 2D png for players placeholders and I'd like to render these figures on the 3D object where needed. An alternative solution would be to render the whole face with a big picture containing all the placeholders figures. However it would be a waste of memory and thus less efficient. What do you suggest me?

    Read the article

  • How to read data from a large number of files in a folder? [closed]

    - by Gary Dhillon
    I seem to be having some trouble figuring out a solution for a problem. See the thing is, my code is supposed to read a lot of data from a bunch of files. I've been thinking of two different approaches: 1) the first one seems simpler, I ask the user if they would like to examine the next file or just quit out of the program.( I believe this is simpler and would take less time to run through.) 2)It reads through all the files and outputs the results for each of them, and then a shared result for all of them.( I think this would be better for what I've been asked to do and it saves the user some hassle.) If anyone can tell me how to code either of these in C++, I would be very grateful. Here is a sample of the file: 0 -- 19 weight 0 -- 20 weight I use this to determine density and possibly ignore the weights which is a number.

    Read the article

  • Assembly as a First Programming Language?

    - by Anto
    How good of an idea do you think it would be to teach people Assembly (some variant) as a first programming language? It would take a lot more effort than learning for instance Java or Python, but one would have good understanding of the machine more or less from "programming day one" (compared to many higher level languages, at least). What do you think? Is it a realistic idea, at least to those who are ready to make the extra effort? Advantages and disadvantages? Note: I'm no teacher, just curious

    Read the article

  • Which is Better: Revo Uninstaller or a Free Alternative?

    - by The Geek
    The results might surprise you! Our friends over at 7 Tutorials did some testing and found that the free IObit uninstaller application actually stacked up pretty well against the paid solution from Revo—though perhaps with a few less features. Note: we’ve always been fans of Revo Uninstaller around here, since it does have a lot of features to help clean up bad apps that the free alternatives don’t have. Either way, the 7 Tutorials article is worth a read. Roundup of Software Uninstallers – Reviewing IObit vs Ashampoo vs Revo [7 Tutorials] Use Amazon’s Barcode Scanner to Easily Buy Anything from Your Phone How To Migrate Windows 7 to a Solid State Drive Follow How-To Geek on Google+

    Read the article

  • How to overwrite Ubuntu with Windows 7?

    - by Will Cowled
    So I have a Windows DVD and it works. But when it gets to the part when it says "Upgrade" or "Custom" I click on custom and at the bottom it says cannot install over it because Windows 7 can only be installed on an NTFS drive? I know that Ubuntu formatted my partitions into one big on that's an ext4. What can I do? I know that I can maybe create a 30-50 GB partition that's an ntfs then when I go into windows I can format the Ubuntu one and combine them but I don't know how to make a partition much less make a big partition in the "GParted" program? So any ideas would be very helpful. I know how to do anything with a hard drive using the default program that comes with Windows 7 but I feel like a mouse in a maze when I open GParted.

    Read the article

  • Prevent Nautilus from generating thumbnails for video files larger than 5 MB?

    - by Lisa
    If I set in Nautilus preferences that it should generate thumbnail previews only for files larger than 10 MB, then this only works for pictures. Nautilus still keeps generating thumbnails for videos and pdf files. Even if a video file is 500 MB. It should only generate thumbnails for video files less than 10 MB as set in the preferences. Same goes for pdf. I have many pdf files that are larger than 50 MB. I don't want Nautilus to generate thumbnails for these, only for small sized pdf. How can I make Nautilus to obey the Previews preferences? Nautilus 3.4.2 Ubuntu 12.04.1, 64bit

    Read the article

  • How should I choose quadtree depth?

    - by Evpok
    I'm using a quadtree to prune collision detection pairs in a 2d world. How should I choose to what depth said quadtree is calculated? The world is made mostly of moving objects1, so the cost of dispatching the objects between the quadtree cells matters. What is the relationship between the gain from less collision checking and the loss from more dispatching? How can I strike a balance that performs optimally? 1 To be completely explicit, they are autonomous self-replicating cells competing for food sources. This is an attempt to show my pupils predator-prey dynamics and genetic evolution at work.

    Read the article

  • Is the tap-to-click issue solved

    - by AWE
    I'm just an average Joe when it comes to computing (maybe less then the average Joe) but I hate tap-to-click. In system and settings there is no touchpad tab? Is it true that this has been fixed? I'm using Dell inspiron N5110 xinput list: ? Virtual core pointer id=2 [master pointer (3)] ? ? Virtual core XTEST pointer id=4 [slave pointer (2)] ? ? PS/2 Generic Mouse id=13 [slave pointer (2)] This is really strange because Dell is one of the top manufacturers in laptops and Ubuntu one of the top distros in Linux and Canonical claims that they are working closely with Dell.

    Read the article

  • One Week on New Servers and Everything is Great

    - by Jeff Julian
    It has been a week since we moved our Geekswithblogs.net System to a new set of load balanced servers and everything has been going great.  I am so amazed at the performance of the new hardware.  On average, we only use less than 5% of the CPU at any given moments or the database and web servers.  I have seen a performance boost in page load as well, but I will have to confirm that with the statistics as they roll in.  This is all in preparation for a new community we are launching with some friends that we will be announcing shortly.  We will be launching a nice little contest for our bloggers as well. Technorati Tags: Geekswithblogs.net,Hardware

    Read the article

  • Is it better to cut and store all sprites needed from a spritesheet in memory, or cut them out just-in-time?

    - by xLite
    I'm not sure what's best practice here as I have little experience with this. Essentially what I am asking is... if it's better to get your single PNG with all your different sprites on it for use in-game, cut out every sprite on startup and store them in memory, then access the already-cut-out sprite from memory quickly or Only have the single PNG with all the different sprites residing in memory, and when you need, for example, a tree. You cut out the tree from the PNG and then continue to use it as normal. I imagine the former is more CPU friendly than the latter but less memory friendly, vice versa for the latter. I want to know what the norm is for game dev. This is a pixel based game using 2D art. Each PNG is actually an avatar's sprite sheet with each body part separated and then later joined to form the full body of the avatar.

    Read the article

  • dpkg reports error on package icaclient

    - by Photonics1
    I installed the icaclient (it's a client for Citrix) a while back. I don't exactly remember what I did to get it working but it was enervating. I had to install some old packages not even avaiable for ubuntu (12.04) and in the end I used some stuff from old rpms. Anyway the client is more or less working now but I always get a dpkg error when installing or updating something. The (translated) error message is something like: dpkg: Error while processing icaclient:i386 (--configure): subprocess installed post-installation-script return errorcode 2 I just want to tell dpkg to ignore this or remove this post-install-script but I don't know how. Thanks!

    Read the article

  • Anti-depressant and programming: does it reduce your thinking and focusing abilities ?

    - by user12358
    I wanted to ask your opinion on anti-depressants, since I took them daily for 3 years now, but I can't be sure if I'm less perfomant with them or without, since I never withdrawed. I'm still at school at the age of 25, still having some motivation problems (for example I can't get used to do something at school if I don't think it will teach me something), but I'm quite motivated to work in the video-game field, since I have some personnal projects in mind. I know C++ programming etc, I'm still learning techniques, but do you think I should try more to do my project instead of just following the work I'm assigned to ? Have you had experience with depression or anti-depressants ? How did it affect your work ? Do you think that being depressed or half-depressed can improve creativity ? Do you feel it has affected the way you end up writing your algorithms ? EDITED

    Read the article

  • Deduping your redundancies

    - by nospam(at)example.com (Joerg Moellenkamp)
    Robin Harris of Storagemojo pointed to an interesting article about about deduplication and it's impact to the resiliency of your data against data corruption on ACM Queue. The problem in short: A considerable number of filesystems store important metadata at multiple locations. For example the ZFS rootblock is copied to three locations. Other filesystems have similar provisions to protect their metadata. However you can easily proof, that the rootblock pointer in the uberblock of ZFS for example is pointing to blocks with absolutely equal content in all three locatition (with zdb -uu and zdb -r). It has to be that way, because they are protected by the same checksum. A number of devices offer block level dedup, either as an option or as part of their inner workings. However when you store three identical blocks on them and the devices does block level dedup internally, the device may just deduplicated your redundant metadata to a block stored just once that is stored on the non-voilatile storage. When this block is corrupted, you have essentially three corrupted copies. Three hit with one bullet. This is indeed an interesting problem: A device doing deduplication doesn't know if a block is important or just a datablock. This is the reason why I like deduplication like it's done in ZFS. It's an integrated part and so important parts don't get deduplicated away. A disk accessed by a block level interface doesn't know anything about the importance of a block. A metadata block is nothing different to it's inner mechanism than a normal data block because there is no way to tell that this is important and that those redundancies aren't allowed to fall prey to some clever deduplication mechanism. Robin talks about this in regard of the Sandforce disk controllers who use a kind of dedup to reduce some of the nasty effects of writing data to flash, but the problem is much broader. However this is relevant whenever you are using a device with block level deduplication. It's just the point that you have to activate it for most implementation by command, whereas certain devices do this by default or by design and you don't know about it. However I'm not perfectly sure about that ? given that storage administration and server administration are often different groups with different business objectives I would ask your storage guys if they have activated dedup without telling somebody elase on their boxes in order to speak less often with the storage sales rep. The problem is even more interesting with ZFS. You may use ditto blocks to protect important data to store multiple copies of data in the pool to increase redundancy, even when your pool just consists out of one disk or just a striped set of disk. However when your device is doing dedup internally it may remove your redundancy before it hits the nonvolatile storage. You've won nothing. Just spend your disk quota on the the LUNs in the SAN and you make your disk admin happy because of the good dedup ratio However you can just fall in this specific "deduped ditto block"trap when your pool just consists out of a single device, because ZFS writes ditto blocks on different disks, when there is more than just one disk. Yet another reason why you should spend some extra-thought when putting your zpool on a single LUN, especially when the LUN is sliced and dices out of a large heap of storage devices by a storage controller. However I have one problem with the articles and their specific mention of ZFS: You can just hit by this problem when you are using the deduplicating device for the pool. However in the specifically mentioned case of SSD this isn't the usecase. Most implementations of SSD in conjunction with ZFS are hybrid storage pools and so rotating rust disk is used as pool and SSD are used as L2ARC/sZIL. And there it simply doesn't matter: When you really have to resort to the sZIL (your system went down, it doesn't matter of one block or several blocks are corrupt, you have to fail back to the last known good transaction group the device. On the other side, when a block in L2ARC is corrupt, you simply read it from the pool and in HSP implementations this is the already mentioned rust. In conjunction with ZFS this is more interesting when using a storage array, that is capable to do dedup and where you use LUNs for your pool. However as mentioned before, on those devices it's a user made decision to do so, and so it's less probable that you deduplicating your redundancies. Other filesystems lacking acapability similar to hybrid storage pools are more "haunted" by this problem of SSD using dedup-like mechanisms internally, because those filesystem really store the data on the the SSD instead of using it just as accelerating devices. However at the end Robin is correct: It's jet another point why protecting your data by creating redundancies by dispersing it several disks (by mirror or parity RAIDs) is really important. No dedup mechanism inside a device can dedup away your redundancy when you write it to a totally different and indepenent device.

    Read the article

  • What happened to the this type of naming convention?

    - by Smith
    I have read so many docs about naming conventions, most recommending both Pascal and Camel naming conventions. Well, I agree to this, it's ok. This might not be pleasing to some, but I am just trying to get your opinion on why you name your objects and classes in a certain way. What happened to this type of naming conventions, and/or why are they bad? I want to name a structure, and I prefix it with "struct". My reason is that, with IntelliSense, I see all structures in one place, and anywhere I see the struct prefix, I know it's a "struct": structPerson structPosition another example is the enum, although I may not prefix it with "enum", but maybe with "enm": enmFruits enmSex again my reason is that in IntelliSense, I see all my enumerations in one place. Because .NET has so many built-in data structures, I think this helps me do less searching. Note that I used .NET in this example, but I welcome language agnostic answers.

    Read the article

  • How do you find partners for open source projects?

    - by static_rtti
    I've created a few open-source projects in the past, and had some success. The process was generally the following: I'd start alone, create something that works, promote it, and finally (maybe) get some contributions. I have less that to contribute to open-source now, but I still have ideas and can still code :) So I wonder if there is a place or a way to meet people (online), discuss ideas and design, and then start working together on the project? The reason I'm thinking about an online way to do this, with strangers, is that while I do have programmer friends, we only very rarely have the same needs and interests at the same time. It seems to me that it would be easier to find such in match in the online global community. Any insight?

    Read the article

  • TypeScript for Visual Studio 2012

    - by TATWORTH
    Originally posted on: http://geekswithblogs.net/TATWORTH/archive/2013/06/21/typescript-for-visual-studio-2012.aspxAt http://www.microsoft.com/en-us/download/details.aspx?id=34790, Microsoft provide a free download of TypeScript for Visual Studio 2012. The documentation site is at http://www.typescriptlang.org/It is described as TypeScript is a language for application-scale JavaScript development.TypeScript is a typed superset of JavaScript that compiles to plain JavaScript.Any browser. Any host. Any OS. Open Source.TypeScript starts from the syntax and semantics that millions of JavaScript developers know today.TypeScript compiles to clean, simple JavaScript code which runs on any browser, in Node.js, or in any other ES3-compatible environment.With TypeScript, you can use existing JavaScript code, incorporate popular JavaScript libraries, and be called from other JavaScript code.These features are available at development time for high-confidence application development, but are compiled into simple JavaScript.If you have written JavaScript, you will know why I welcome the release of version 0.9 of TypeScript as TypeScript should be a lot less frustrating to write. I suggest you go to https://typescript.codeplex.com/ and follow this very promising project.

    Read the article

  • Is ROA a specific form of doing SOA?

    - by JohnDoDo
    I have read somewhere that ROA (Resource Oriented Architecture) is SOA (Service Oriented Architecture) with specific constraints added. SOA is the abstract concept of combining discrete pieces of software and ROA is an implementation of SOA with all of the constraints of RESTful services applied to it: SOA = the concept ROA = the concept + specific implementation details I also had my share of posts saying that ROA is different than SOA, then simply fallback to statements like "ROA is REST" and "SOA is SOAP" and presenting the same more or less pertinent comparisons and differences between SOAP and REST. So just to clear up my confusion: Is ROA a specific form of doing SOA?

    Read the article

  • LINQ/LAMBDA filter query by date [on hold]

    - by inquisitive_one
    I'm trying to use LINQ to SQL to retrieve earnings data for a particular date range. Currently the table is set up as follows: Comp Eps Year Quarter IBM .5 2012 2 IBM .65 2012 3 IBM .60 2012 4 IBM .5 2011 2 IBM .7 2013 1 IBM .8 2013 2 Except for Eps, all fields have a data type of string or char. Eps has a data type of double. Here's my code: var myData = myTable .Where(t => t.Comp.Equals("IBM") && Convert.Int32(string.Format("{0}{1}", t.Year, t.Quarter)) <= 20131); I get the following error when I tried that code: Method 'System.String Format(System.String, System.Object, System.Object)' has no supported translation to SQL How can I select all Eps that has a year & quarter less than "20132" using a lambda expression?

    Read the article

  • test coverage reality

    - by iPhoneDeveloper
    I am NOT doing test driven development and I write my test classes after the actual code is written. In my current project I have a test coverage of(Line coverage) %70 for 3000 lines of Java code.(Using JUnit, Mockito and Sonar for testing) But while I feel actually I am not covering and catching %70 of the problems that can occur. So my question is in theory is that possible to have a %100 Line coverage but in reality it is meaningless because of low quality of the test code and maybe a %40 well written test code is much better than a bad %100 coverage? or we can always say line coverage more or less gives the percentage of all covered issues?

    Read the article

< Previous Page | 72 73 74 75 76 77 78 79 80 81 82 83  | Next Page >