Search Results

Search found 26798 results on 1072 pages for 'difference between detach attach and restore backup a db'.

Page 684/1072 | < Previous Page | 680 681 682 683 684 685 686 687 688 689 690 691  | Next Page >

  • hg unshelve not working

    - by shanebonham
    Our team is just getting started with Mercurial. One of the first things we've started to play with is hg shelve. Locally, I have no problem shelving changes. It all works perfectly from what I can tell. However, when I try to unshelve, I get the restoring backup files message, but when I run hg diff, there are no changes, and my changes are missing from the code. If i do hg unshelve -i I can see the diff, but again, trying to unshelve seems to have no effect. I've been trying to test it with some very simple changes that shouldn't be a problem in terms of conflicts, e.g. adding a test comment. I should note that I've tried hg unshelve -f after which it says unshelve completed but again, my changes are not restored. Any ideas what I am doing wrong? If it matters: Mercurial Distributed SCM (version 1.5.1+20100405)

    Read the article

  • Started with a local git repo now I want to push my changes to a remote server

    - by Eliseo Soto
    Hi, I started a new project and created a local git repo with "git init" and now I have a few branches and everything works great. However since my webhosting company offers git hosting (if you're curious https://support.eapps.com/index.php?_m=knowledgebase&_a=viewarticle&kbarticleid=203) I'd like to push my entire repo to their servers to have a backup in the cloud in case something bad happens to my local repo. How can I make the remote repo the "origin" since the repo was started locally? Hope my question makes sense. Thanks, a Git newbie.

    Read the article

  • Account for cancelling an update process

    - by Aerus
    I'm trying to develop the logic of updating my (Java) application but I'm stuck on a particular problem. The user can, at any time, be it on purpose (by clicking the Cancel button) or unwillingly (system/power failure) cancel the update process. If the updater was busy downloading files or unpacking files and thus didn't make any changes to the main application yet, this can be easily recovered by re-downloading the files and continuing the update process like normally. But, how can i account for changes made to the main application (i.e. a file was only partially replaced, some files were corrupted, etc.) that result in a main application that can't be launched anymore? Is it possible to rollback the changes made during the update without having to keep a full backup of my application or without doing a full reinstall? Are there any standards for updating an application that can help me with this?

    Read the article

  • Should I rebuild table indexes after a SQL Server 2000 to 2005 database migration

    - by Joe T
    I'm tasked with doing a SQL Server 2000 to 2005 migration. I will be doing a side-by-side migration. After restoring from a backup I plan to do the following: ALTER DATABASE <database_name> SET COMPATIBILITY_LEVEL = 90; DBCC CHECKDB(<database_name>) WITH NO_INFOMSGS DBCC UPDATEUSAGE(<database_name>) WITH NO_INFOMSGS exec sp_updatestats ‘resample’ Should I rebuild table indexes before using DBCC UPDATEUSAGE and sp_updatestats? Have I missed anything obvious that should be executed after a migration? All help would be warmly up-voted. Thanks

    Read the article

  • How to find that 'runas' execution finished?

    - by Radek
    I use ruby 1.9.3p194 (2012-04-20) [i386-mingw32] on Windows7 To do mySQL backup I run runas /savecred /user:yogurt\administrator "cmd.exe /k mysqldump --user=#{dbuser} --password=#{dbpassword} #{dbname} > #{dump}" - mysqldump must be executed as administrator. I do not run my ruby scripts under administrator account. runas starts new cmd.exe and ruby doesn't wait for it to finish. Dump process takes about one minute to finish. After that I zip the dump file and delete it. But I have to make sure that the dump process already finished before I do any other action on that file. Right now I use sleep(60) that works but I wonder if there any better more systematic solution.

    Read the article

  • Using SVN alone or in small workgroups - workflow approach?

    - by Industrial
    Hi everybody, I have spent some months working on a web application and we're come close to production stage. It's soon time to expand the development group with 1-3 people on this project. I have not too much experience on working with SVN, but It's obviously the choice for a big part of the larger companies out there, so I am guessing that the pros of SVN without a doubt outweights the time spent on commit/check ins / check outs etc. The workflow seems to become a bit more complicated with SVN, and even though I have read Version Control with Subversion by O'Reilly Media and I am not sure yet if it's overkill to use SVN for any reasons besides backup when developing alone or in a small (1-3 people) workgroup? How do you do it? What's your workflow with version control while working alone or in small workgroups? Thanks!

    Read the article

  • best place to store .net assemblies

    - by stackoverflow
    To give a scenario, let us simply assume an engine that loads plug-ins and exposes features in the plug-in. User 1 uploads a plug-in which allows implements Act 1 User 2 uploads a plug-in which allows implements Act 2 A plug-in in this case is .net assembly. Now in this scenario, if we have to store all the assemblies - what would be the best place? Also, the plug-in would require to be versioned so execution can happen of a particular version. Further considering the plug-in engine is installed on multiple machines or on same machine as different instances (similar to sql server). Would a centralized database (sql server 2005) with a table to store all the assembly be a good idea (centralized backup etc.,) (assembly size would be around 500kb to 1MB)

    Read the article

  • Converting files to blob to save in database

    - by jimbo
    I have an old user site that we are currently updating. The site's users have previously uploaded profile pics that are stored in a directory. I am now wanting to keep these profile images in the database, (easy to backup all data) but am having trouble working out how to do it. There are a lot of tutorials talking about do ing this from files that have just been uploaded and using the tmp name etc, but how can I re-create with files that have already been uploaded? I have looked into $data = file_get_contents($filename); Which seems to create a binary file, but doesn't seem to save in database with: mysql_query("UPDATE profiles SET company_logo = mysql_real_escape_string('".$data."') WHERE id = 1 ") or die(mysql_error());

    Read the article

  • Moving from dedicated to shared cpanel - any scripts to do all / some of the install tasks ?

    - by mbbcat
    Hi, I have a few hundred phpld sites to move - each has its own cpanel, ( & the target may have shared cpanel) & I can do a full cpanel backup on the original server, but I don't have whm on the current host - the backups are fairly easy to organize but the installs so far means picking through files & setting up db's & mail etc by hand - I am thinking there ought to be an easier ie scripted way to do the installs or at least some parts - can anyone please suggest something ? I would like to migrate the stats at the same time Thanks M

    Read the article

  • phpMyAdmin "No database selected" MySQL

    - by user1751660
    I downloaded a MySQL backup file and promptly imported into MAMP's phpMyAdmin. I got this return: Error SQL query: -- -- Database: `mysql` -- -- -------------------------------------------------------- -- -- Table structure for table `columns_priv` -- CREATE TABLE IF NOT EXISTS `columns_priv` ( `Host` CHAR( 60 ) COLLATE utf8_bin NOT NULL DEFAULT '', `Db` CHAR( 64 ) COLLATE utf8_bin NOT NULL DEFAULT '', `User` CHAR( 16 ) COLLATE utf8_bin NOT NULL DEFAULT '', `Table_name` CHAR( 64 ) COLLATE utf8_bin NOT NULL DEFAULT '', `Column_name` CHAR( 64 ) COLLATE utf8_bin NOT NULL DEFAULT '', `Timestamp` TIMESTAMP NOT NULL DEFAULT CURRENT_TIMESTAMP ON UPDATE CURRENT_TIMESTAMP , `Column_priv` SET( 'Select', 'Insert', 'Update', 'References' ) CHARACTER SET utf8 NOT NULL DEFAULT '', PRIMARY KEY ( `Host` , `Db` , `User` , `Table_name` , `Column_name` ) ) ENGINE = MYISAM DEFAULT CHARSET = utf8 COLLATE = utf8_bin COMMENT = 'Column privileges'; MySQL said: #1046 - No database selected I did not alter the .sql file at all. Any hints on how i can get this puppy going locally? Thanks!

    Read the article

  • disable download of my paid app in Android

    - by Boy
    I have a paid app in the store which will remove the ads in another app when it is installed on that device. Now I want to remove this 'remove ads' app, as I want to have an in-app payement for this for instance (or maybe I just keep the ads version only). But the problem is, if I unpublish the app, people who bought it will not be able to download it again when they get a new phone or reset their phone. How to I keep the app in the Play Store, but prevent people from buying it? Is this possible? My backup plan is: make the app cost 10.000 euro's and put in the message that this app should not be bought anymore. But I don't like that...

    Read the article

  • Compressing digitalized document images

    - by Adabada
    Hello, We are now required by law to digitalize all the financial documents in our company and submit them to evaluations every 3 months. Since this is sensitive data we decided to take matters into our own hands and build some sort of digital data archiver. The tool works perfectly, but after 7 months of usage we are begining to worry about the disk space used by these images. Here some info on the amount of documents digitalized: 15K documents scanned and archived per day, with final PNG size of +- 860KB: 15 000 * 860 kilobits = 1.53779984 gigabytes 30 days of work per month: 1.53779984 gigabytes * 30 = 46.1339952 gigabytes Expectation of disk space usage after 1 year: 46.1339952 gigabytes * 12 = 553.607942 gigabytes So far we're at 424 gigabytes of disk space used, without counting backup. We're using PNG as image format, but I would like to know if anyone have any advice on a better compression algorithm for images or alternative strategies for compressing the PNG's even more or even better ways to archive images as to save disk space. Any help would be appreciated, thanks.

    Read the article

  • Splitting string into array upon token

    - by Gnutt
    I'm writing a script to perform an offsite rsync backup, and whenever the rsyncline recieves some output it goes into a single variable. I then want to split that variable into an array upon the ^M token, so that I can send them to two different logger-sessions (so I get them on seperate lines in the log). My current line to perform the rsync result=rsync --del -az -e "ssh -i $cert" $source $destination 2>&1 Result in the log, when the server is unavailable ssh: connect to host offsite port 22: Connection timed out^M rsync: connection unexpectedly closed (0 bytes received so far) [sender] rsync error: unexplained error (code 255) at io.c(601) [sender=3.0.7]

    Read the article

  • SVN always getting errors when I commit (Subclipse)

    - by jax
    I have setup svn on my server and Subclipse at home. I am the only developer and am mainly using it for the backup and versioning features. Everytime I commit my changes I get eighter: Out Of date errors or Tree conflicts Sometimes I even delete files and they don't delete on svn, in a directory hierarchy only the very last item will delete so I have to delete each folder one at a time. How do I avoid these errors in the future? Update: Another problem I am having is that sometimes eclipse seems to sync with the server so that when I refactor a filename it goes off to the server and does something and makes me wait, which is annoying. And for clarity, this is a typical operation: I might change a filename, move a file to a different folder then change the contents of a file. I select the 'Team menu' and click 'commit'. Then I get all these errors above.

    Read the article

  • Robocopy Invalid Parameters

    - by Drue
    I'm using the batch command robocopy to copy my website source files from my flash drive to my dad's laptop for a backup. I've ran it as an administrator and I get an invalid parameter error message. Here's what I have so far: @echo off robocopy "G:\xampp-portable\htdocs" "W:\xcopytest" /e pause>nul I want to use %cd% instead of "G:\xampp-portable" Also, my .bat file is in G:\xampp-portable. UAC is set to normal and I'm using Windows 7. [EDIT] I tried a similar code in 'W:\' under a test folder and it worked fine. I think the problem is the drive to drive. [/EDIT]

    Read the article

  • Drawing a waveform in C#

    - by user488792
    Hi! I want to be able to display a WaveForm in C#, along with some simple features such as zooming and selection. I already have the data as a short[] of amplitude values. However, I am an amateur when it comes to hardcoding GUI. I have already found a possible helper class WaveFormClass that may help me achieve this but as a backup, I want to learn how to manually do it. So may I ask for some methods and possibly some links that will help? Thanks!

    Read the article

  • Empty files generated from running `mysqldump` using PHP

    - by alex
    I keep getting empty files generated from running $command = 'mysqldump --opt -h localhost -u username -p \'password\' dbname > \'backup 2009-04-15 09-57-13.sql\''; command($command); Anyone know what might be causing this? My password has strange characters in it, but works fine with connecting to the db. I've ran exec($command, $return) and outputted the $return array and it is finding the command. I've also ran it with mysqldump > file.sql and the file contains Usage: mysqldump [OPTIONS] database [tables] OR mysqldump [OPTIONS] --databases [OPTIONS] DB1 [DB2 DB3...] OR mysqldump [OPTIONS] --all-databases [OPTIONS] For more options, use mysqldump --help So it would seem like the command is working.

    Read the article

  • What is the purpose of Process class in Java?

    - by Nitesh Panchal
    Runtime objRuntime = Runtime.getRuntime(); String strBackupString = "mysqldump -u " + userName + " -p" + password + " " + dbName; Process objProcess = objRuntime.exec(strBackupString); This is used for backup of database. But what exactly happens? Can anybody make me explain, what is the purpose of Runtime and Process class? Is this class used to act as if we are typing command from command prompt? Then what should i pass to objRuntime.exec() if i want to open notepad? And is the command executed as soon as we call exec method? If yes, then what purpose does Process serve here? I really can't understand these two classes. Please make me understand. Thanks in advance :)

    Read the article

  • Error in SQL. Can not find it.

    - by kmsboy
    Error in SQL. Can not find it. DECLARE @year VARCHAR (4), @month VARCHAR (2), @day VARCHAR (2), @weekday VARCHAR (2), @hour VARCHAR (2), @archivePath VARCHAR (128), @archiveName VARCHAR (128), @archiveFullName VARCHAR (128) SET @year = CAST(DATEPART(yyyy, GETDATE()) AS VARCHAR) SET @month = CAST(DATEPART(mm, GETDATE()) AS VARCHAR) SET @day = CAST(DATEPART(dd, GETDATE()) AS VARCHAR) SET @weekday = CAST(DATEPART (dw, GETDATE()) AS VARCHAR) SET @hour = CAST(DATEPART (hh, GETDATE()) AS VARCHAR) SET @archivePath = 'd:\1c_new\backupdb\' SET @archiveName = 'TransactionLog_' + @year + '_' + @month + '_' + @day + '_' + @hour + '.bak' SET @archiveFullName = @archivePath + @archiveName BACKUP LOG [xxx] TO DISK = @archiveFullName WITH INIT , NOUNLOAD , NAME = N'?????????? ??? ??????????', SKIP , STATS = 10, DESCRIPTION = N'?????????? ??? ??????????', NOFORMAT

    Read the article

  • read variables from file

    - by TonyVipros
    Hi, I'm trying to write a shell script to backup websites, I've got it to go through each directory in vhosts and create a new directory vhosts/{siteurl}/private/backups/ and gzip up the site into there. However, I also want it to read a file from private that will contain database name, user and password. So I can then use these to perform a mysqldump to private/backups/. The problem is that I've never written a shell script before so I've go no idea how to read variable data into my script, or how I should format the file to be read. Hope that's enough information.

    Read the article

  • How to nightly update a copy folder of main website?

    - by HollerTrain
    I have a main website, then I have a backup of the main site in another folder on the server to be used if main site goes down I can push traffic to my copy of the main site. With that said I want to nightly update the copy site with the updated files made to the main site from the prior day's updates to text files and new images. What is best method of doing this automatically every 24 hours? I assume a cron script could be created for this need? Any help would be greatly appreciated.

    Read the article

  • .net File.Copy very slow when copying many small files (not over network)

    - by Guavaman
    I'm making a simple folder sync backup tool for myself and ran into quite a roadblock using File.Copy. Doing tests copying a folder of ~44,000 small files (Windows mail folders) to another drive in my system, I found that using File.Copy was over 3x slower than using a command line and running xcopy to copy the same files/folders. My C# version takes over 16+ minutes to copy the files, whereas xcopy takes only 5 minutes. I've tried searching for help on this topic, but all I find is people complaining about slow file copying of large files over a network. This is neither a large file problem nor a network copying problem. I found an interesting article about a better File.Copy replacement, but the code as posted has some errors which causes problems with the stack and I am nowhere near knowledgeable enough to fix the problems in his code. Are there any common or easy ways to replace File.Copy with something more speedy?

    Read the article

  • What is the importance of the .DFSFolderLink

    - by Swift
    I am currently building a new DFS namespace setup. My folder E:\CommonStuff\ is already shared on the fileserver. And now it is also shared as \DOMAIN\CommonStuff I got 3 questions: I see .DFSFolderLink records in all folders I create in the DFS Managment console. But all folders that was in E:\CommonStuff allready does not contain these records. Are the .DFSFolderLink records important? Does it make any difference if I create subfolders "the old way" in the E:\Commonstuff\ or if I do it within the DFS Managment console?

    Read the article

  • Install server 2012 on HP ML110 G7 with B110i controller, no disks found

    - by Molotch
    I'm trying to install Server 2012 on a HP ML110 G7 with a B110i controller and four non hot-swappable SATA drives. I just can't get the Server 2012 boot disk PE environment to find any disks. I have downloaded the latest SPP (Service Pack for Proliant 2012.10) and flashed the BIOS. I have tried two different HP drivers for B110i and Windows X64, 6.18.0.64 and 6.18.2.64 to no avail. I have tried setting the controller to both AHCI and legacy mode in the BIOS, no difference. HP:s SmartStart disc for G7 servers only support installation of up to Windows Server 2008R2. HP:s installation instructions for Server 2012 Essentials says boot from the windows disk and use the storage drivers found on the SPP (I can't find any storage drivers on the SPP disk).

    Read the article

  • RAID 0 Volatile Volume Cache Mode configuration

    - by SnippetSpace
    I discovered that in IRST there is an option to set a cache mode for my 3 ssd raid 0 array. I've read the documentation by Intel and have some questions: Are there any overall benefits/risks from enabling cache mode? As I'm on a laptop, would write back be recommended? I read it increases chance of data loss on power interruption. What is the difference between how windows handles data integrity and the intel driver? Read only mode seems to have the benefit of faster reads, does it have any downsides? Thanks for your help guys!

    Read the article

< Previous Page | 680 681 682 683 684 685 686 687 688 689 690 691  | Next Page >