Search Results

Search found 33569 results on 1343 pages for 'sql backup and restore'.

Page 121/1343 | < Previous Page | 117 118 119 120 121 122 123 124 125 126 127 128  | Next Page >

  • esx backup usb disk

    - by maruti
    my server has a disk error, unfortunately RAID-0. So i am planning to boot it off CD (partedmagic) and copy the VMs to a USB disk. File system is VMFS (esxi4) once the damaged disk is replaced back could the data be restored back? this server has two datastores this bad disk belongs to store-1. please suggest any better ways or tools. thanks in advance

    Read the article

  • Mirror/Backup from SSH/SFTP to Windows

    - by Andrew Russell
    What I am trying to do is mirror a directory (recursively) from a server I can SSH/SFTP to, to a Windows machine. I want to do this as part of a script, so it can be automated. I only want to copy new or modified files. I don't want to have to download all the files every time the script runs. In other words, I'm trying to get the equivalent of RoboCopy /MIR that will work using SFTP as a source. What would you recommend?

    Read the article

  • How to backup a folder and it's subfolders to CD/DVD

    - by Maximus
    If I've got a folder structure that I'd like to split so that it could be burned to multiple CDs or DVDs, what's the best tool out there to accomplish this? The folders/files have to be natively accessible once the split is done (no compression or physical splitting that would require a later join), and the folder structure should still remain intact as it spaces discs. Any ideas?

    Read the article

  • Robocopy with local catalog of remote data for incremental backup

    - by Bill
    I am currently using robocopy to an extremely slow destination. The compare between source and destination files can take a while to run through. Since the destination will never change (apart from the robocopy changes), is there any program that will work similarly to robocopy, but have a local list of what files (attributes and timestamps) the destination has, to compare with? I know there are expensive solutions which may do this, but I'm looking for something free if possible. Hopefully this makes sense.

    Read the article

  • Cisco Configuration backup with Windows Script.

    - by Jeff
    We have a client with a lot of Cisco Devices and we would like to automate the backups of these devices through telnet. We have both 2003 and 2008 servers and ideally use tftp to back it up. I wrote this: Set WshShell = WScript.CreateObject("WScript.Shell") Dim fso Set fso = CreateObject("Scripting.FileSystemObject") Dim ciscoList ciscoList = "D:\Scripts\SwitchList.txt" Set theSwitchList = fso.OpenTextFile(ciscoList, 1) Do While theSwitchList.AtEndOfStream <> True cisco = theSwitchList.ReadLine Run "cmd.exe" SendKeys "telnet " SendKeys cisco SendKeys "{ENTER}" SendKeys "USERNAME" SendKeys "{ENTER}" SendKeys "PASSWORD" SendKeys "{ENTER}" SendKeys "en" SendKeys "{ENTER}" SendKeys "PASSWORD" SendKeys "{ENTER}" SendKeys "copy startup-config tftp{ENTER}" SendKeys "(TFTP IP){ENTER}" SendKeys "FileName.txt{ENTER}" SendKeys "exit{ENTER}" 'close telnet session' SendKeys "{ENTER}" 'get command prompt back SendKeys "{ENTER}" SendKeys "exit{ENTER}" 'close cmd.exe On Error Resume Next WScript.Sleep 3000 Loop Sub SendKeys(s) WshShell.SendKeys s WScript.Sleep 300 End Sub Sub Run(command) WshShell.Run command WScript.Sleep 100 WshShell.AppActivate command WScript.Sleep 300 End Sub But the problem with this is the sendkeys are sent to the console session, I'm trying to find a solution that would not require a user to be logged in. Does anyone have any ideas? I have some knowlage of VBS, PowerShell and a pretty good grasp on batching.

    Read the article

  • Backup solutions for Rackspace Cloud Sites

    - by Daniel A. White
    What options do I have for backing up the content from a Rackspace Cloud Sites including files and databases? I know they have cron jobs, but I am not sure what options I have when it comes to that. Here are some of the things the Cron jobs they support. http://cloudsites.rackspacecloud.com/index.php/How%5Fdo%5FI%5Fschedule%5Fa%5Fcron%5Fjob%3F

    Read the article

  • Backup drivers for a device in Windows 7

    - by Moses
    If I can't find a driver on the internet when I prep for a reinstall of Windows 7, I figure if the driver is already installed, why not back it up and reinstall later? The only method I know of is going through Device Manager Device Properties Driver Details. This gives me a nice list of all the files, but then I need to go manually retrieve all of them. For devices with a lot of driver files, this is inefficient. Is there a better way to accomplish this?

    Read the article

  • How can I copy files from a DVD skipping corrupt files?

    - by ujjain
    I have so far tried the following software solutions without success. Windows Explorer Copy (default) TeraCopy Roadkil's Unstoppable Copier Unstoppable Copier stops copying as a whole, it keeps continuing the same file after 20 minutes, despite any settings. I have spent much time Googling, but the one program ("Unstoppable Copier") that people recommended, did not to the trick as it keeps choking on the same file despite a setting of "Fasteset Data Recovery", "Maximum Retries 0" and "Undamaged Files First". The back side of the DVD has some scratches. What options do I have? I have access to both Ubuntu and Windows 7.

    Read the article

  • Multi Gateway and Backup Routing on a cisco router

    - by user64880
    Hi all, I have a 2611 Cisco Router with only one Fastethernet port Now I have two internet gateways. I want to config my router as when primary routing fails second routing automatically start to route all my packets. When I set 2 IP route command in my router then I check I see it work well but when peer IP on primary routing is down it can not change to second routing until I remove first route command.In the following I write my setting. How can I set it? interface FastEthernet0/0 ip address 81.12.21.100 255.255.255.248 secondary ip address 62.220.97.14 255.255.255.252 ip route 0.0.0.0 0.0.0.0 62.220.97.13 ip route 0.0.0.0 0.0.0.0 81.12.21.97 100 Cheer, Kamal

    Read the article

  • Collect and Backup Photos from Multiple Photographers

    - by Graviton
    I have a few photographers working under me ( well, not exactly under me, but I say it anyway just for illustration purpose), so they shoot a lot of photos, the problem is that they all store their photos on their hard disk, and I have no way to retrieve them unless I pass them an USB and ask them to fill it with their photos. Very labor intensive and inefficient! Is there any other better ( more automated) ways of doing this? For the fear of losing the resolution, I don't really favor a online-synch approach, because I think all the photos uploaded to any website will suffer a resolution loss, which is the last thing I want. Is there a better idea? Edit: Being artistic as they are, I can't guarantee that they all use PC and Windows; so the software must at least be able to run on Mac.

    Read the article

  • Boot to VHD backup plan

    - by Josh Barker
    I have a machine that I just reinstalled Windows and all of my applications onto... what a chore that is. I want to totally and completely avoid this from now on by creating an image. My first thought was to see if it possible to copy a VHD file when you are booted into it since I am using Windows 7 Ultimate as boot-to-vhd (without a parent machine). Is this possible and if so, how could I accomplish this? Keep in mind, this is my personal machine and I'm trying to keep things inexpensive (a good script would work). Thanks, Josh

    Read the article

  • backup and file server for 50+ TB of data

    - by a-bomb
    our office wants to build a new server to handle our data, over the last 10 years our data was stored on CDs, DVDs, HDDs but now they want all of it in one place that is attached to the network for everybody in the office to access it. the data is 20TB new data and the rest is old, the important now is to store these 20tb and gradually store the other 30tb over time. so what is the best solution to do ? we thought of getting an hp server and connect it to an external enclosure that either had tape drives or HDDs (we haven;t decided yet) or to get a NAS server and connect it to the hp server. what should we do because this is new for us ...

    Read the article

  • Backup linux to ftp server

    - by Alakdae
    What do you use for backups to ftp server? I've tried the setup with Amanda and virtual tapes on the ftp server mounted with Curlftpfs and I'm not satisfied with it. I just don't feel confident about Amanda. Also I cannot use anything that uses rsync on the ftp mounted filesystem because it only creates the directories and doesn't create files as it cannot execute "mkstemp". I've been thinking about Bacula but I can't find any good HOWTO for it.

    Read the article

  • PST backup with Volume Shadow Copy Service

    - by NoMadMan
    I was asked to implement the task of backing up 35 PST files ranging from 800Mb to 2000Mb. Windows XP and Windows 2000 workstations are assigned to the users and we have a Windows 2000 domain controller we use to back up files on 3x 500Gb external hard drives. I found several methods from applications to scripts. Local or remote applications would be my last resort. I came across this script based on Volume Shadow Copy Service. CopyWithVss I wanted to know if there would be a problem if the path had spaces. Would mounting the destination path of each PST folder with a drive letter be more practical? My concern with mounting option is that i would eventually run out of letters since I have 35 and possibly more workstations to back up. Lastly, can someone give me an example of CopyWithVss if it were run on a production network? The script is a bit cryptic even after reading several times. Where in the script do I enter the source and the destination? I'm a Mac user so please excuse my ignorance with Windows platform.

    Read the article

  • Deploying a Windows 7 image, which way is the fastest?

    - by MatF
    I captured an image of a basic Windows 7 installation with some modifications using imagex. Before the image was captured, I ran sysprep generalizing and selecting to enter OOBE after it's done. Which way would be the fastest to deploy that image again? Using imagex /apply. Or naming the image install.wim and put it in the source folder of a normal installation (on a bootable USB device), running a normal setup afterwards. Currently I have only tried the second approach. However I just found out about the imagex way and wondered if it would be faster. Or are the even more methods that would be better?

    Read the article

  • Reinstall Windows on mass computers

    - by user791022
    At work there are about 40 DELL PC's - half of them are Windows Vista and other half are Windows 7. Some of the DELL PC's are 220s and 230s model. If I want to re-install Windows then I can use a Recovery/Recovery disk on any Dell PC's (220s or 230s). It take too long to complete the setup, for example - after re installing the Windows then I need to install the drivers, update windows, install IE9, create default users account and so on. It take about 3 hour to complete setup! I am looking for a solution Where I can create an Image of windows (with driver and completed setup) and then I use same Image it on any PC's at work? If one of the PC need reinstalling - I would like to reinstall via network (LAN) by grabbing the Image? Is there something that before booting into windows - computer check something via network which can download the Image?

    Read the article

  • Using Windows Azure storage for backup

    - by Bruno
    I am currently looking at Windows Azure blobs as an option for backing up archive data. I want to be able to upload files from an external windows machine via the internet but I don't know enough about Windows Azure storage to make a decision. Some of the questions I have are How do I upload the files. Is there a client application, can I use robocopy? Would it be fast enough? i.e. Could I download or upload 1TB of data in a week? Is it secure? Hopefully someone smarter than me can help me :-)

    Read the article

  • Schedules Folder Backup

    - by Junaid Saeed
    i have some folders in C drive on which i work on daily basis and the data in them is very critical.. so every night when i shutdown my PC i copy - paste - overwrite existing files these folders to a separate location... so that of the system crashed or something bad happens.. i will be able to easily format C and all i cannot move these folders from C drive because these folders include C:\wamp\www\ of WAMP server and such folders... is there a tool on which i can schedule that everyday at X time these folders will be backuped to 'Y' path

    Read the article

  • Lazy linux backup system?

    - by Alex
    Ok, so I want to say Time Machine, but that's note exactly what I'm looking for. I want to set up a system that will regularly (hourly?) back up the /home directory of our machine. time Machine style things are naturally preferable since they save space by only saving the changes, but honestly, this is important enough that I can suffer some waste. Any ideas?

    Read the article

  • Backup, Migrate or Clone Failing CentOS 4 (LVM)

    - by Hegelworm
    I've been running a BlueQuartz CentOS 4 system (Nuonce.net distro) for a few years now and although the hard drive (Deskstar) has always been a bit noisy, on a few recent occasions I've heard it having trouble spinning up. Basically, I want to clone this drive to a similar sized one (80 Gig). I've spent many hours reading upon dd, dd_rescue, rsync, clonezilla and LVM mirroring yet the sheer number of options and nightmarish accounts has left me frozen - unable to make an informed decision as to how to start. I've made a few attempts. dd failed after about 2 hours, as, although the drives appeared to be identical on the surface (ATA Seagate Barracudas, Thai not Chinese), the destination drive is slightly smaller. My most recent attempt involved using a Debian CD to format the new drive and then rsync-ing everything over and editing the new drive's grub and fstab to reflect the changes. No joy here either as I hadn't chosen LVM when partitioning the destination drive and it wouldn't boot. As you can probably tell, I'm out of my depth here and a panic-invoking mixture of caution and frustration has prompted me to sign up here. The server itself, although not strictly a production environment, has a very specific installation of Festival, LAME and ffMpeg and provides the back-end for a Text-to-Speech jQuery plugin that I've built over the last 2 years. I'm also planning to rebuild the whole TTS system on Debian as the existing CentOS system still has PHP4 etc. For now though, I'd really like to just shift everything over to a new drive. As this is my first post, please feel free to lay any house rules on me that I might've overlooked; I've been hovering around StackOverflow for a while now but have only just signed up. Many thanks. Update: Thanks for your responses so far - it's much appreciated and makes me feel a little more confident when I can double-check things here. I had the idea of doing a fresh install of CentOS (from the original disk) on the new drive so the partitions and LVM were all set up correctly (after disconnecting my source drive to prevent painful mistakes). I then booted into rescue mode from the same CD, and, to avoid a conflicting label, changed the /boot partition's label using e2label to /bootnew. I then changed the VolGroup name using lvm vgrename from VolGroup00 to VolGroup001. I could then boot with both drives in. After mounting the new drive (via its VolGroup001 alias) into /newhd, I rsync-ed over everything I could to the new drive, using -avr switches and backslashes. Like mentioned here. I then disconnected my original source drive again, booted from the liveCD again, changed back the boot partition label from /bootnew to /boot using e2label and then renamed the VolGroup back to VolGroup00. I then rebooted and it went through the familiar start-up routine only to not find a host of files in proc, usr, lib, var etc. The boot did complete but there were lots of red 'FAILS'. I could log in with my existing creds, but the network was kaput, I couldn't startX (desktop GUI) and there were also a few (a lot) of error messages pertaining to iptables. Back to square one. I naively thought I'd nailed it. Shall I just buy a bigger hard drive and attempt the dd route? I've read that this can mess with LVM setups and there's the added risk of working on two unmounted drives at once with a low-level tool. Thanks again.

    Read the article

  • Backup with bash and rsync...

    - by Roger
    Is there a way to auto-rename an existing file on the receiver? For example: if filename already exists, it auto-rename filename to something like filename_001, filename_002 and so on.... So far all I have is this: $ rsync -rh --progress --stats --exclude '.thumb' \ --update --perms /origin /destination By the way, I know rsync has --ignore-existing to "skip updating files that exist on receiver", but I guess what I need would be something like --rename-existing.

    Read the article

  • Setting up logging for a remote backup script

    - by Brian Dainis
    So I wrote up a short script that I am planning to run via a cron job daily to package up my site files and send them to a remote location. I also plan to incorporate DB dumps, but I have not gotten that far yet. My issue today however is that Im am uncertain how to log the output of each command for errors, warnings, or other pertinent information the command may output. I would also like to install sometype of fail safe so if something goes horribly wrong the script will stop dead in its tracks and notify me via email or something. Ok the email thing is not as critical, but would be nice. Does anybody have any ideas for that? Here is what I have so far. By the way, both servers are CentOS 6.2 running standard LAMP. #!/bin/sh ################################# ### Set Vars ################################# THEDATE=`date +%m%d%y%H%M` ################################# ### Create Archives ################################# tar -cf /root/backups/files/server_BAK_${THEDATE}.tar -C / var/www/vhosts gzip /root/backups/files/server_BAK_${THEDATE}.tar ################################# ### Send Data to Remote Server ################################# scp /root/backups/files/server_BAK_${THEDATE}.tar.gz user@host:/home/bak1/ftp/backups/ ################################# ### Remove Data from this Server ################################# rm -rf /root/backups/files/server_BAK_${THEDATE}.tar.gz

    Read the article

< Previous Page | 117 118 119 120 121 122 123 124 125 126 127 128  | Next Page >