Search Results

Search found 6798 results on 272 pages for 'continuous backup'.

Page 169/272 | < Previous Page | 165 166 167 168 169 170 171 172 173 174 175 176  | Next Page >

  • Review before or after code commit, which is better?

    - by fifth
    Traditionally we performed code review before commit, I had an argument with my colleague today, who preferred code review after commit. First, here's some background, we got some experienced developers and we also got new hires with almost zero programming practice. we'd like to perform fast and short iterations to release our product. we all team members locate at same site. The advantages of code review before commit I've learned, mentor new hires try to prevent errors, failures, bad designs in early developing cycle learn from others knowledge backup if someone quits But I also got some bad experience, like low efficiency, some changes may be reviewed over days hard to balance speed and quality, especially for newbies some guy felt distrust As to post-review, I just knew little about this, but the most thing I worried about is the risk of losing control, people never review. Any opinions?

    Read the article

  • SSC Clinic: Can Implementing "Optimize for Ad Hoc Queries" Boost Performance for the SQLServerCentral.com and Simple-Talk.Com SQL Servers?

    With the introduction of the instance-level option “optimize for ad hoc workloads” in SQL Server 2008, DBAs have a tool to deal with a problem known as plan cache pollution, or plan cache bloat. It’s often caused when one-time use ad hoc queries are sent to SQL Server from Object-Relational Mapping (ORM) solutions, such as LINQ, NHibernate, or Entity Framework. The problem can prevent SQL Server from using its available memory optimally, potentially hurting performance. Get smart with SQL Backup ProGet faster, smaller backups with integrated verification.Quickly and easily DBCC CHECKDB your backups. Learn more.

    Read the article

  • Why doesn't cron complete my script?

    - by brickinthewall
    I have a backup script (rsync via ssh) which is run by cron (configured in /etc/crontab) 0 2 * * * root /bin/bash --login /opt/aebackup/sshbackup.sh If I run it as logged in root like following it runs prefectly fine. root@server:~# /opt/aebackup/sshbackup.sh If I run it via cron it would just stop after a while (not always on the same task in the script.. it seems pretty random, like the process is killed at some point randomly) Does anyone have an idea why my cron would do that? thanks for anything... I'm desperate!

    Read the article

  • Which folders need to be backed up for migration in Joomla?

    - by Devdatta Tengshe
    I'm helping someone update & migrate their old website, built on the Joomla framework. Currently it is running on Joomla 1.5.8 which is an ancient version. I've convinced them to upgrade Joomla to at least 2.5 I have already made a backup of the database. Most links I have seen talk of backing up the entire public_html folder (The website runs on a shared host). But in my fresh Joomla installation there are several folders that are in the public_html folder. So which of the folders in the public_html folder are from the content of the website, and which are of the old Joomla framework? I'm afraid that I might overwrite files of the new Joomla framework with the old framework, if copy all the files and folders into the new installation.

    Read the article

  • The 2012 JAX Innovation Awards

    - by Janice J. Heiss
    A new article, now up on otn/java, titled “The 2012 JAX Innovation Awards” reports on  important Java developments celebrated by the Awards, which were announced in July of 2012. The Awards, given by S&S Media Group, aim to, "Reward those technologies, companies, organizations and individuals that make outstanding contributions to Java." The Awards fall into three categories: Most Innovative Java Technology, Most Innovative Java Company, and Top Java Ambassador. In addition, a finalist who did not win an award receives a Special Jury prize, "in acknowledgement of their unique contribution and positive impact on the Java ecosystem."The winners were: JetBrains for Most Innovative Java Company; Adam Bien as Top Java Ambassador; Restructure 101, created by Headway Software, as Most Innovative Technology; and Charles Nutter, Special Jury award. Each winner received a $2,500 prize. The five finalists in each category were invited to attend the JAX Conference in San Francisco, California. This year's winners each received a $2,500 prize. JetBrains Fellow, Ann Oreshnikova, listed her favorite JetBrains innovations: * Nullability annotations and nullability checker* CamelCase navigation and completion* Continuous Integration in grid (on multiple agents), in TeamCity* IntelliJ Platform and its language support framework* MPS language workbench* Kotlin programming languageWhen asked what currently excites him about Java, Adam Bien, winner of the Java Ambassador Award, expressed enthusiasm over the increasing interest of smaller companies and startups for Java EE. “This is a very good sign,” he said. “Only a few years ago J2EE was mostly used by larger companies -- now it becomes interesting even for one-person shows. Enterprise Java events are also extremely popular. On the Java SE side, I'm really excited about Project Nashorn.”Special Jury Prize Winner, Charles Nutter of Red Hat, remarked that, “JRuby seems to have hit a tipping point this past year, moving from ‘just another Ruby implementation’ to ‘the best Ruby implementation for X,’ where X may be performance, scaling, big data, stability, reliability, security, and a number of other features important for today's applications. Check out the complete article here.

    Read the article

  • Slow transfer to external USB3 hard drive

    - by JMP
    Trying to backup data from hard drive before reloading windows following some issue with its load. Having trouble with the file transfer to a USB3/2 external hard drive NTFS. Getting transfer speed of about 116.7kB/sec. In other words its taking about 5 hours to transfer 1.4GB. I've got about 80GB to go. So the transfer is going to take 11days. Seems a little on the slow side. Am I missing something? Is there a way to make this faster. No issue with the external drive transferring this amount in windows. But don't have that option at the moment.

    Read the article

  • Switching my legacy desktop back to Windows XP from Windows 7

    - by Kevin Shyr
    I was happy with Windows 7 at the beginning, until I started to add in the peripherals.  Windows 7 was never able to recognize any of my PCI video card (I know, I know, we should be in the DVI age). Anyway, I went through another 4 days of trouble setting my computer up with dual monitor in XP (also did a bunch of other things like getting rid of my sound card and taking the computer off RAID. Kind of feel stupid to put the computer on RAID in the first place because now I can have 2 drives: double the page files program seems to run faster Microsoft Sync toy 2.1 takes care of my backup needs (Thank god they solved the network drive issue) As of last night, the system is running beautifully.  I still have a laptop with Windows 7, but even that is in dual boot mode.

    Read the article

  • Ubuntu 12.04 wont install on Macbook

    - by user92325
    I've installed Ubuntu on my Macbook before but something went wrong with the updater. So, I had to backup my HDD and format it and reinstall OS X Lion. But here's the thing: Ever since I re-installed my OS I've been trying to get Ubuntu back on the HDD. I partitioned it to 40 GB and set the file system to Ext4. I also recently created a swap partition too and it seems to install correctly. After i installed rEFIt it just has this cute little penguin sitting there. I rebooted and tried to go back to Ubuntu and the penguin still shows up but this time a black screen pops up and it asks me to insert a bootable device and press any key. I'm not sure why this is happening. This is probably the 5th time I've tried to install it. I've even used a different Ubuntu ISO but it still won't boot after the installation.

    Read the article

  • Oracle Technology Network July 2012 Special Offers

    - by programmarketingOTN
    Oracle Technology Network July 2012 New offers are here!  Manning is offering 41% off The Well Grounded Java Developer and Oracle Press has added a discount on the eBook for Effective MySQL: Backup and Recovery.  To learn more and get the discount codes/links please go to the OTN Member Discount page.Lets not forget the other GREAT offers still going on as well - Packt Publishing Offers -  25% off - Oracle SOA Infrastructure Implementation Certification Handbook (1Z0-451 Oracle BPM Suite 11g Developer's cookbook Pearson Offers - 35% off Java Applications ArchitectureApress Offers - 40% off Beginning Database Design ENDS July 6th!Murach Offers - 30% off Oracle SQL and PL/SQL Safari Books Online - 10-day free trial + 20% off unlimited access to Safari Books Online for 6 monthsOracle Store Discounts - Save 10% on Your Next Purchase from the Oracle Store!

    Read the article

  • Which folders need to be backedup for migration in Joomla?

    - by Devdatta Tengshe
    I have never used Joomla before, so this might be a noobish question. I have tried searching, but haven't found anything relevant. I'm helping someone update & migrate their old website, built on the Joomla framework. Currently it is running on Joomla 1.5.8 which is an ancient version. I've convinced them to upgrade Joomla to atleast 2.5 I have already taken the database backup. Most links I have seen talk of backing up the entire public_html folder (The website runs on a shared host). But in my fresh Joomla installation there are several folders that are in the public_html folder. So which of the folders in the public_html folder are from the content of the website, and which are of the old Joomla framework? I'm afraid that I might overwrite files of the new Joomla framework with the old framework, if copy all the files and folders into the new installation.

    Read the article

  • Neural network input preprocessing

    - by TND
    It's clear that the effectiveness of a neural network depends strongly on the format you give it to work with. You want to preprocess it into the most convenient form you can algorithmically get to, so that the neural network doesn't have to account for that itself. I'm working on a little project that (surprise!) is going to be using neural networks. My future goal is to eventually use NEAT, which I'm really excited about. Anyway, one of my ideas involves moving entities in continuous 2D space, from a top-down perspective (this would be a really cool game AI). Of course, unless these guys are blind, they're going to be able to see the world around them. There's a lot of different ways this information could be fed into the network. One interesting but expensive way is to simply render a top-down "view" of things, with the entities as dots on the picture, and feed that in. I was hoping for something much simpler to use (at least at first), such as a list of the x (maybe 7 or so) nearest entities and their position in relative polar coordinates, orientation, health, etc., but I'm trying to think of the best way to do it. My first instinct was to order them by distance, which would inherently also train the neural network to consider those more "important". However, I was thinking- what if there's two entities that are nearly the same distance away? They could easily alternate indexes in that list, confusing the network. My question is, is there a better way of representing this? Essentially, the issue is the network needs a good way of keeping track of who's who, while knowing (by being inputted) relevant information about the list of entities it can see. Thanks!

    Read the article

  • I cannot play some WMA files

    - by Lucio
    I did a backup of several music files from my older Windows XP system. Now I can play all the .MP3 files but not all the .WMA. There are some kinds of WMA files that can be reproduced without problems, may refute this in the picture below. The right file can be played, the left isn't. I have long been looking different sites, Q&A from here, installing many packages but not luck. An example is this answer. What can I do? My system: Ubuntu 11.10 32b. Player: Banshee & VLC

    Read the article

  • Programming my first C++ program

    - by Jason H.
    I have a basic understanding of programming and I currently learning C++. I'm in the beginning phases of building my own CLI program for ubuntu. However, I have hit a few snags and I was wondering if I could get some clarification. The program I am working on is called "sat" and will be available via command line only. I have the main.cpp. However, my real question is more of a "best practices" for programming/organization. When my program "sat" is invoked I want it to take additional arguments. Here is an example: > sat task subtask I'm not sure if the task should be in its own task.cpp file for better organization or if it should be a function in the main.cpp? If the task should be in its own file how do you accept arguments in the main.cpp file and reference the other file? Any thoughts on which method is preferred and reference material to backup the reasoning?

    Read the article

  • Installing packages into local directory?

    - by Gili
    I'd like to install software packages, similar to apt-get install <foo> but: Without sudo, and Into a local directory The purpose of this exercise is to isolate independent builds in my continuous integration server. I don't mind compiling from source, if that's what it takes, but obviously I'd prefer the simplest approach possible. I tried apt-get source --compile <foo> as mentioned here but I can't get it working for packages like autoconf. I get the following error: dpkg-checkbuilddeps: Unmet build dependencies: help2man I've got help2man compiled in a local directory, but I don't know how to inform apt-get of that. Any ideas? UPDATE: I found an answer that almost works at http://askubuntu.com/a/350/23678. The problem with chroot is that it requires sudo. The problem with apt-get source is that I don't know how to resolve dependencies. I must say, chroot looks very appealing. Is there an equivalent command that doesn't require sudo?

    Read the article

  • U1 music shows unknown artist, how can I make it recognize m4a files?

    - by bisi
    Hello all, I am in the process of uploading my music library as a backup to U1, but I figured, why not also enjoy Ubuntu 1 music on my iPhone? With a few difficulties to start, the upload is now in progress, but I've noticed that there is a huge percentage of files in the unknown artist folder, and I believe it is all of my m4a files. They play fine, but without any information. Coming from an iTunes background, and having bought the majority of my music on the iTunes store, I wonder how I could make this work, easily? I am on Maverick (afaik), but About Ubuntu shows 11.04. I use Banshee as my music manager, and I monitor my sync using Ubuntu one preferences, ubuntuone-indicator and magicicada. The total file size of my music folder is 38.9GB. Thank you for your help!! And apologies if I couldn't find a thread where this was already covered...

    Read the article

  • Recommended formats to store bitmaps in memory?

    - by Geotarget
    I'm working with general purpose image rendering, and high-performance image processing, and so I need to know how to store bitmaps in-memory. (24bpp/32bpp, compressed/raw, etc) I'm not working with 3D graphics or DirectX / OpenGL rendering and so I don't need to use graphics card compatible bitmap formats. My questions: What is the "usual" or "normal" way to store bitmaps in memory? (in C++ engines/projects?) How to store bitmaps for high-performance algorithms, such that read/write times are the fastest? (fixed array? with/without padding? 24-bpp or 32-bpp?) How to store bitmaps for applications handling a lot of bitmap data, to minimize memory usage? (JPEG? or a faster [de]compression algorithm?) Some possible methods: Use a fixed packed 24-bpp or 32-bpp int[] array and simply access pixels using pointer access, all pixels are allocated in one continuous memory chunk (could be 1-10 MB) Use a form of "sparse" data storage so each line of the bitmap is allocated separately, reusing more memory and requiring smaller contiguous memory segments Store bitmaps in its compressed form (PNG, JPG, GIF, etc) and unpack only when its needed, reducing the amount of memory used. Delete the unpacked data if its not used for 10 secs.

    Read the article

  • Is it OK to push my code to GitHub while it is still in early development?

    - by marco-fiset
    I have some projects that are in a very early development state. They are nowhere nearing completion but I do host them (as public repos) on GitHub because: I have multiple computers and I want access to my code everywhere I want a backup for my code I want it to be easy if someone wants to collaborate in some way I use GitHub Issues as a poor man's project management software Is it OK to publish a project on GitHub even when it is very early in the development? I am a bit concerned about someone to come by and say OMG this is total BS, this code is so bad! while looking at unpolished/still in development/not tested code. What are your practices when you start new public projects? Do you wait until you have something substantial to show or you create a bare repo directly on GitHub and start from there? I used GitHub throughout this post but this applies to every code hosting service out there.

    Read the article

  • What do I need to learn to decide on rename/recompile source package names because of company rebranding?

    - by Roberto Linares
    My company is currently at a rebranding process and the brand names have been used in the sources' package names but these names are only visible to developers who maintain this code so nobody from project management is really interested in changing them considering also that it would imply the recompiling of several old components. What factors do I need to consider when deciding on a change like that? I don't know if I should worry about legal issues or not and if so, how to address this with project management. More background details. I have all the sources and dependencies but since the company rebranding, other development areas have adopted some of the code that needs package name-changing so I cannot take the decision only by myself so I don't make everyone else's code to crash with my core components and I cannot change other areas' code without the permission of those areas' users so yes, my concern is more political than technical. I am going try to coordinate the involved it areas to make the change anyway, since it seems to be the best approach.   Unfortunatelly in my company there's no continuous integration build server so we build our code manually on demand and to get something to production I have to justify the change (even just the package name changing) to QA with an user requirement and some other bureaucratic documentation so that's why I was hesitating the change in first place.

    Read the article

  • Ask the Readers: How Do You Give an Old Laptop a New Life?

    - by Jason Fitzpatrick
    That powerhouse laptop you bought back in 2006 can’t compete with the sleek ultrabook you just unboxed–but that doesn’t mean you should ship it to the dump. How do you give an old laptop a new lease on life? Whether you tear it apart and rebuild it into something brand new, put it on night duty as a backup station, or install a lightweight Linux distro before passing it on to your relatives, we want to hear all about your tools and methods for keeping old laptops out of the junk bin. However big or small your repurposing project, sound off in the comments below with your tips, tricks, and tools. Make sure to check back in on Friday for the What You Said roundup to see how your fellow readers revitalize their old laptops. How To Delete, Move, or Rename Locked Files in Windows HTG Explains: Why Screen Savers Are No Longer Necessary 6 Ways Windows 8 Is More Secure Than Windows 7

    Read the article

  • New hard drive means re-install everything

    I recently found a good deal on a 7200 RPM 500GB latpop drive and decided to upgrade myself to that. Ive been operating on a 150GB drive for a long time and hadnt been cleaning up old projects, encodings, etc. so it was running thin on space. This was a cheap upgrade (< $90) so I opted for it. I finally got a moment last evening to pop it in and start the fresh installs of everything. Doing so reminded me of how long this process takes for us geeky people. Luckily I do a backup first using my...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Quickly Investigating What's in the Tables of SQL Server Databases

    From SQL Server Management Studio it's hard to look through the first few rows of a whole lot of tables in a database. This is odd, since it is a great way to get quickly familiar with a database. Phil tidied up a SQL routine he uses to investigate databases quickly in a browser. He explains how to use it, how it works, and how to use it from PowerShell. Want faster, smaller backups you can rely on?Use SQL Backup Pro for up to 95% compression, faster file transfer and integrated DBCC CHECKDB. Download a free trial now.

    Read the article

  • Advice and resources on collaborative environments

    - by Tjaart
    I need some advice on collaborative software environments. More specifically, I am looking for books and reference materials that can aid me in understanding team and code structures and the interactions thereof. In other words books, blogs or white papers explaining: Different strategies for structuring teams that share common code between each other but have distinct individual functions? To summarise my question I would like to know what would be a good source of knowledge if I were to set up teams in an organisation that shared code but each unit still remained autonomous. I have done some research on this subject and explored: code review tools, distributed VCS, continuous integration tools, Unit testing automation. The tough part about implementing these tools are to determine where a good place would be to start, which tools are low hanging fruit, which tools or methods provide higher success rates. If someone asks me about code quality reference I point them to Code Complete. I am looking for an equivalent guide on software team structures and tools to make this equation work better. I realise that this question is quite vague but it arose as "we need to share code between teams without breaking each others stuff and causing management headaches and reams of red tape" The answer is definitely not simple and requires changes on many levels, hence the question. If the question is too vague please vote to close or delete. I would accept any good starting point as an answer.

    Read the article

  • Command line option to check which filesystem I am using?

    - by j-g-faustus
    Is there a command that will show which file system (ext3, ext4, FAT32, ...) the various partitions and disks are using? Similar to how sudo fdisk -l lists information about disks and partitions? Update Accepted the "mount" answer as mount works without specifying filesystem type (commenting out the relevant entries in fstab, if any): $ sudo mount /dev/sdf1 /mnt/tmp $ mount | grep /mnt/tmp /dev/sdf1 on /mnt/tmp type ext3 (rw) Found another option in ubuntuforums - blkid: # system disk $ sudo blkid /dev/sda1 /dev/sda1: UUID="...." TYPE="ext4" # USB disk: $ sudo blkid /dev/sdf1 /dev/sdf1: LABEL="backup" UUID="..." TYPE="ext3" # mdadm RAID: $ sudo blkid /dev/md0 /dev/md0: LABEL="raid" UUID="..." TYPE="ext4" Thanks for your help!

    Read the article

  • SQL Server Luxembourg User Group

    Come join the SQL Server Luxembourg UG for free training and networking on June 27th at 5:30pm. Soren Nielsen will take a Deep Dive into SQL Server 2012’s “Always On” High Availability technology. This will be followed by Vern Rabe of the SQL User Group in Portland, Oregon, presenting “Data Types - Think You Know It All? Think Again”. Want faster, smaller backups you can rely on?Use SQL Backup Pro for up to 95% compression, faster file transfer and integrated DBCC CHECKDB. Download a free trial now.

    Read the article

  • iTunes 7 can't find iPod in PlayOnLinux using Ubuntu 12.04 Precise?

    - by Endia
    I'm using 12.04 and the latest version of PlayOnLinux to run iTunes 7. I have my iPod 4th gen plugged in and the computer recognizes that it's mounted but it's not showing up in iTunes for me to do anything with it. I know there have been a lot of issues with being able to access your iDevices in iTunes from Linux when you're using a compatibility layer like PlayOnLinux or Wine (both of which I have), so it's recommended that I use Rhythmbox to back up my music etc. Which is fine, because that's part of what I need to do, but my primary need in trying to access my iPod in iTunes is that I need to remove my backup password so that I can jailbreak my iPod with the new Absinthe jailbreak. If there is some other way I can do this without iTunes you know of, I'd love to hear it. If not, what can I do?

    Read the article

< Previous Page | 165 166 167 168 169 170 171 172 173 174 175 176  | Next Page >