Search Results

Search found 15087 results on 604 pages for 'copy constructor'.

Page 334/604 | < Previous Page | 330 331 332 333 334 335 336 337 338 339 340 341  | Next Page >

  • SQL Server 2000 need to prevent logons whilst performing a backup for a side by side migration

    - by pigeon
    I'm looking for a way to prevent logons from occurring in order to take a full backup of a Database to migrate from its current SQL Server 2000 instance to a new SQL 2005 instance. A friend of mine suggested running a script which would put the DB into a rollback state. Not being a DBA my DDL is very poor and running a script that I don't understand may not be the best idea. One option which might be easier is to simply detach and copy, to the new server. Any suggestions would be greatly appreciated.

    Read the article

  • Incremental backup up from a remote ftp box to a Windows server

    - by user65712
    I need to backup a website to a Windows server every week and I only have access to remote FTP. I'd like to use an incremental backup program so that I can just copy the files every week on a schedule and not worry too much about the size of the backups becoming an issue. Unfortunately, I can't find a Windows program that will automatically make incremental backups of specific FTP folders and files, as most programs are designed to backup to FTP, not from it. Are there any applications that can do this? I also have a Ubuntu 10.04 box I could use to relay the site to the Windows server if I needed to run Linux programs, but I would prefer a Windows-only solution over a Linux/Windows one, and a combined Linux/Windows solution over not having it work at all.

    Read the article

  • Forgot to unmount/eject external hard drive, lost moved files. Mac OS X

    - by balupton
    So I was using my Mac with my external hard drive connected via USB. I moved about 10 GB of data to it (via drag and drop while holding down the Command key to move the files rather than to copy them). They moved to the drive all right, but as I was having some issues and the Finder crashed after the transfer, I was unable to eject the volume and later everything froze so I had to do a hard restart (hold the power button). When I remounted the volume (plugged the external hard drive back in) it no longer had any of the files which I moved onto it. As it was a lot of data, how can I recover these files?

    Read the article

  • Windows 7 - sharing file - slow seek time?

    - by progtick
    I do not want to copy the files. I simply have video files (.flv) on one computer and I would like another computer user to watch them without copying. Playback is fine, but seek time (as in if you move the cursor to skip some portion of the video) takes forever! I thought wireless speed might be the culprit, so I wired the two computer. Maybe I saw some improvement but still so bad. It's 1 Gbps! (I know real speed will vary, but before I monitor real speed and such, do I have reasonable issue? Or am I bound to have very slow seek time?) What is going on? I must mention some of these files are huge!

    Read the article

  • rsync for coping file

    - by vinayrks
    I am migrating my old server to new server . I used this server for hosting website . first I tried sftp but due to huge number of files and connection time out , it simply didn't work . then I tried rsync .rsync working good , but only problem I am facing it updating file very nicely & fastly but do not copy new files please help me . because still i need to transfer lots of file. I am using this command : rsync -anv -e ssh oldserver:/path/ /path

    Read the article

  • Free Cloud Mind Map Solution

    - by Zekta Chan
    As a Software Engineer we had a lot of discussion on SE design. Although we sync every document on the development process on Google Doc, mind map just didn’t fit in Google doc yet. The best we can do is to store a copy and share it online. To me, Mind map is an ir-replaceable piece of tools (yet) as agile note taking tools. And an eFormat is even greater than a paper one, due to the portability and extensibility. Does anyone have a good solution on “cloud-sharing” mind map? (We are using FreeMind at the moment) Thanks

    Read the article

  • Proper way to re-image Windows 7?

    - by Alec
    I had a driver completely fail on me so I have to restore my computer from a system image backup. I used an installation DVD to run the Re-Image utility on there, but after 8 hours of "preparing the image" to be restored, it began restoring to my hard drive. After 12 hours there, it was 5-8% complete. I figured I must have done something wrong or it started doing something wrong. So I installed a fresh copy of Win-7 and ran the utility from there. It's going at the same snail's pace. I still think I must be doing something incorrectly - I don't see how it could possibly take so long, I could probably manually flip the on my hard drive and be done before that utility. Am I doing this correctly or is there something else I should be doing? Edit: In case my hardware is relevant: Win 7 64 bit Core i7 8 GB Ram 640GB Internal 1TB External connected via eSATA I had approximately 400GB of data on my computer before it crashed.

    Read the article

  • Linux WD30EZRX WD Green HDD & Blacx Duet 5G Usb

    - by Adam
    I have connected up an WD30EZRX WD Green HDD to a Thermaltake Blacx Duet 5G USB dock in Ubuntu 12.04. Every thing seems fine except when the HDD idles it seems to have error ls: reading directory .: Input/output error after a while and is only fixed by unmounting and remounting the drive as root. I have the following line in /etc/fstab UUID=AAF670E9F670B6E3 /media/3TB ntfs defaults,user,auto 0 0 I have noticed that it seems to go between /dev/sdc2 and /dev/sdd2 devices on remount. I did copy 1TB last night without issue in 1 sitting. But after x mins of idle it has remount issue. Any tips/suggestions on how to proceed would be appreciated. Spent most of the night googling and all its done is made me sad. Edit (tried as suggested): root@mediaserver:/media/3TB# sudo hdparm -B 255 -S 253 /dev/sdd2 /dev/sdd2: setting Advanced Power Management level to disabled HDIO_DRIVE_CMD failed: Input/output error setting standby to 253 (vendor-specific) APM_level = not supported Seems as if that didn't help with this particular drive.

    Read the article

  • What does "Endcode #816" stand for in Canon iR-ADV C5035 printer logs?

    - by Odin Kroeger
    We are using the Canon iR-ADV C5035. Today a worker has reported that some Word documents cannot be printed via that device (others do). I could confirm that and found that the Canon iR-ADV C5035 logs showed an "Endcode" "#816" for every print job that was not printed (whereas other print jobs show an "Encode" "OK"). Unfortunately, neither the manual that came with the Canon copy machine nor the web seems to know what this "Endcode" means. Canon support has been informed, but I'm still waiting for them to call me back. Does anyone know how I could find out what this "Endcode" stands for? Or, preferably, where I can look up the meaning of such Canon "Endcodes" in the future? Thanks in advance for any hints.

    Read the article

  • Disable MathML output of eLyXer

    - by Gryllida
    eLyXer is a standalone LyX to HTML converter. In the resulting file, equations are formatted as MathML, and the file itself starts with an XML tag. This causes two problems: LibreOffice does not read the XML file (it can read HTML files, but not XHTML). I am unable to copy and paste the equations into a document editor such as LibreOffice with the goal of subsequent conversion into .doc, because .doc files do not support MathML. The eLyXer help page mentions an option to only use simple math, but there is no option to set math equations to output as images. And I already set Document Settings Output Math equations Format: images in LyX, which presumably is saved in the lyx document somewhere. A web search did not come up with any solutions.

    Read the article

  • Deployment Workbench no longer available after PXE boot

    - by Patrick
    Our build process revolves around windows Deployment Workbench. Unfortunately this was setup by someone who is no longer with the company, and no-one has ever dared/needed to make any changes. The other day it stopped working. It turns out that one of our build guys started thinking about changing some stuff in it, clicked something and now it no longer works (He is saying now that he right clicked on the 'LAB' entry in 'Deployment points' and hit 'Update', which took some time to run through apparently). The job has fallen on me to resolve and frankly I'm not sure what I'm doing. I was wondering if someone with more experience than me can provide some pointers as to troubleshooting cos I'm feeling quite a lot in the dark here. On the server I have Deployment Workbench up and running (MMC snapin) version 3.0. There is a WDS service that appears to be running ok, as does the tFTPd service. Nothing specific to this in event logs. From the client side; PXE boot works and gets you to the Win PE launch, and it has the correct company logo as the background (proving to me that its loading win PE from the network). WPEINIT runs, and asks for domain credentials, here the team simply put User/Pass/Domain in the boxes and click ok. Normally the build would kick off. Instead they get an error message saying that the \NATBLU01\Distribution$ share isn't available. Checking \NATBLU01\Distribution$ shows that its there and accessible over the network. Security/permissions seem ok, even 'ANONYMOUS LOGON' has read access to that share so I don't see that being a problem. Digging the trace files from C:\MININT\SMSOSD\OSDLOGS\ after an attempt to run the build I can see an error saying much the same - <![LOG[Validating connection to \\NATBLU01\Distribution$]LOG]!><time="16:42:14.000+000" date="03-15-2012" component="LiteTouch" context="" type="1" thread="" file="LiteTouch"> <![LOG[FindFile: The file OSDConnectToUNC.exe could not be found in any standard locations.]LOG]!><time="16:42:14.000+000" date="03-15-2012" component="LiteTouch" context="" type="1" thread="" file="LiteTouch"> <![LOG[The network location cannot be reached. For information about network troubleshooting, see Windows Help.]LOG]!><time="16:42:24.000+000" date="03-15-2012" component="LiteTouch" context="" type="3" thread="" file="LiteTouch"> <![LOG[ERROR - Unable to map a network drive to \\NATBLU01\Distribution$.]LOG]!><time="16:42:24.000+000" date="03-15-2012" component="LiteTouch" context="" type="3" thread="" file="LiteTouch"> BDD.LOG shows much the same. Full copies of the .LOG files can both files be found here : BDD.LOG LITETOUCH.LOG I can get to a command prompt from the Win PE that boots from PXE, however there isn't any network stuff there. IPCONFIG returns nothing so none of the tests I would usually run resolve anything. I'm at a loss frankly. I did wonder if I could perhaps start a new build process but if the change to the DeploymentWorkbench has knocked it offline I don't think I'm going to be able to create a new deployment. Failing that; we do have a deployment point labeled type 'Media' which appears to be a DVD ISO image of one of the builds, but its dated 2008, is it possible to export the network build to .ISO and build from DVD? We are looking at new hardware to run this from anyway (for the impending Windows 7 roll out) so a temporary work round isn't going to be too much of a problem. All assistance is appreciated! EDIT : OK. Got it working again. Solution was close to Newmanth's idea. The problem was that our PE image didn't appear to be connecting the network. I had an older copy of the PE boot.WIM on a stick that I had been using for other purposes. I booted that and correctly got a network connection. Showed a correct internal IP and could ping out etc etc. However I was still getting the same errors in all the logs and in when wpeinit was running. What I did seperately was to update the PE image that DeploymentWorkbench was pushing out to display a different back ground. I wanted to prove that I was working in the correct place. Turns out that I wasn't. I went and looked at the other deployment stuff we had on this machine, Windows Deployment Services was installed and although all the install images are off line the boot image was online, so I uploaded the copy from my stick to that. Booted straight off. And fixed. Working. Yay! For anyone stumbling across this in the future you may find that although your deployment images are located in the DeploymentWorkbench, the Win PE boot image you are launching from is located in the associated Windows Deployment Services images.

    Read the article

  • Server hard disk read speed and client download speed, is there a connection? [closed]

    - by Mywiki Witwiki
    Ok so a client's download speed is only as fast as a server's upload speed, and vice versa. Based on the answers to this post: Does upload speed depend upon download speed of the server? In other words, the data transfer rate between the two computers is only as fast as the speed of the "bottleneck". Let's pretend the two computers are in two different networks and both have 100Mbps internet connection. Ben wants a copy of a file in Mark's computer hard disk with 30Mbps read speed. Does this mean that Ben can download the file at a speed of around 30Mbps only, despite having an internet connection faster than 30Mbps?

    Read the article

  • Can I mark a folder as mountpoint-only?

    - by Collin
    I have a folder ~/nas which I usually use sshfs to mount a network drive on. Today, I didn't realize the share hadn't been mounted yet, and copied some data into it. It took me a bit to realize that I'd just copied data into my own local drive rather than the network share. Is there some way to mark in the system that this folder is supposed to be a mount point, and to not let anyone copy data into it? I tried the permissions solution here: How to only allow a program to write to a directory if it is mounted?, but if I don't have write access I also can't mount anything to it.

    Read the article

  • Is there an Alternative to TextMate's mate and rmate for Windows?

    - by TiernanO
    As part of the upcoming TextMate 2 release, there will be a new feature called rmate, which will allow you to edit files from a remote machine (Linux/Unix/OSX) via SSH using your local copy of TextMate... Is there something similar for Windows? I know i could use CyberDuck, find the file i want to edit, download locally, work on it and then re-upload, but rmate looks like you just type rmate on the remote server, and text mate pops up with the file. (i have not tried since i am not a TextMate owner) Is there something similar for Windows?

    Read the article

  • Windows 7 x64 CD to MP3 ripper

    - by marc_s
    I'm trying to find a good, simple CD-to-MP3 ripper to copy my physical CD's to my computer's hard disk. I'm running Windows 7 Professional x64 - and that appears to be a major problem. All the "usual" free- and shareware tools I've tried (CD-to-MP3, Visual MP3 and quite a few more) seem to have trouble with either 64-bit Windows, or with my HP CD/DVD built into my HP Compaq Elite 8100 machine. Does anyone have any good recommendations? I don't want to install a monster like iTunes - something really clean, small, simple would be fine. Free- or Shareware - if it works reliably and with good quality output, I'll be happy to register! Any hints are welcome!

    Read the article

  • pull sql query execution location from either the sql server or IIS

    - by jon3laze
    I am working on restructuring the database for a project that has hundreds of classic asp pages. I need to be able to find out which pages are executing which queries so that I can analyze the data. I am hoping there is some way to accomplish this without having to manually open each asp page and copy/paste the queries into a spreadsheet. I would imagine this should be something I could pull from possibly logs? Any info is appreciated. IIS 7 MSSQL 2008 R2 Windows Web Server 2008 build 6001

    Read the article

  • Using rsync to take backup of folder

    - by Ali
    Hi, I have a server (Linux) with NAS which is mounted as folder "mount" I have website in "public_html" folder. I want to take backup of website in mount folder automatically at certain intervals for e.g. every hour. I read that there is something called "rsync" which is used to make two folders sync. And it doesn't copy all files every time and instead matches if the file has been changed and then only update changed files. How do I use it to make automatic backups? I have root access to server. Thanks

    Read the article

  • Using a script that uses Duplicity + S3 excluding large files

    - by Jason
    I'm trying to write an backup script that will exclude files over a certain size. If i run the script duplicity gives an error. However if I copy and paste the same command generated by the script everything works... Here is the script #!/bin/bash # Export some ENV variables so you don't have to type anything export AWS_ACCESS_KEY_ID="accesskey" export AWS_SECRET_ACCESS_KEY="secretaccesskey" export PASSPHRASE="password" SOURCE=/home/ DEST=s3+http://s3bucket GPG_KEY="gpgkey" # exclude files over 100MB exclude () { find /home/jason -size +100M \ | while read FILE; do echo -n " --exclude " echo -n \'**${FILE##/*/}\' | sed 's/\ /\\ /g' #Replace whitespace with "\ " done } echo "Using Command" echo "duplicity --encrypt-key=$GPG_KEY --sign-key=$GPG_KEY `exclude` $SOURCE $DEST" duplicity --encrypt-key=$GPG_KEY --sign-key=$GPG_KEY `exclude` $SOURCE $DEST # Reset the ENV variables. export AWS_ACCESS_KEY_ID= export AWS_SECRET_ACCESS_KEY= export PASSPHRASE= When the script is run I get the error; Command line error: Expected 2 args, got 6 Where am i going wrong??

    Read the article

  • I used my ntfs formatted hard drive, copied files from a mac using a bridge program, but now i can't

    - by Marc
    I used my ntfs formatted hard drive, copied files from a mac using a bridge program, but now i can't see any of the files on my pc help? I brought it from home, (NTFS HD) found that it was only read only.. so i got a bridge program (don't remember) and was able to copy files from the mac to the hard drive. I took it home and went to get into those files, and they do not even show up.. it copied, and did everything it was supposed to do to let me know it was working, but now i can't find any of those files. am I FUBAR? :(

    Read the article

  • How to FTP multiple folders to another server using mput in Unix?

    - by Mircea
    I am logged in on a server (using Putty). from there I'm connecting using FTP to another server. I want to copy several folders from the first server to the second server using mput. Like: ftp mput folder1 folder2 folder3 But I get "folder1: not a plain file."...and so on. Each of these folders have subfolders, files (some binary, some not). How can I accomplish what I want without zipping the stuff and then transfer? Thanks.

    Read the article

  • Can't Access a file I downloaded since I don't have permissions yet I am admin?

    - by chobo2
    Hi I downloaded some file and everytime I try to copy it or move it somewhere I get these errors Windows cannot access the specified device, path , or file. You may not have the appropriate permissions to access the item. You need permission to perform this action You require permission from "user(where this is my pc name/myusername)" to make changes to this file. Yet I am an admin and it is my only account on this computer and it saying to me I need permissions to open it up from "chobo2" user yet that's the account I am logged in as. I don't get it. How do I disable this stupid message so I can access every file I want. I am using windows 7 ultimate 64bit. As far as I know I have full permissions set so I don't know why I am getting this.

    Read the article

  • Backup Client/Server Software that Syncs only the Delta?

    - by Urda
    I have a co-located server, and a desktop computer. I push small things and large ammounts of small files (like my iTunes) into a JungleDisk cloud. If a few files change there, no big deal, the file gets re-uped. For larger files JungleDisk backup isn't helpful. Things like movies and VMware images that change a lot, but I want backed up. Just not to JungleDisk since that would cost me even more money. I am looking for a product, closed or open source (preferably open source) that will sync the change, or delta, to my personal server on a schedule. That way I can keep a copy of my larger things, without paying JungleDisk a ton more since they are in the range of many Gigabytes. Right now these few items are backed up over FTP, and take forever. Both the client and server are windows environments.

    Read the article

  • Download a website that requires log-in with HTTtrack Copier

    - by H.Moss
    Hi guys! I have been researching of how to download content of a site that requires username and password. This is actually harder than I thought it would be. I tried to use both HTTtrack Copier and followed the instruction below, but it's not working! Q: I can not access several pages (access forbidden, or redirect to another location), but I can with my browser, what's going on? A: You may need cookies! Cookies are specific data (for example, your username or password) that are sent to your browser once you have logged in certain sites so that you only have to log-in once. For example, after having entered your username in a website, you can view pages and articles, and the next time you will go to this site, you will not have to re-enter your username/password. To "merge" your personnal cookies to an HTTrack project, just copy the cookies.txt file from your Netscape folder (or the cookies located into the Temporary Internet Files folder for IE) into your project folder (or even the HTTrack folder)

    Read the article

  • Checking for orphaned snapshots - ESXi5

    - by Tim Alexander
    So we had some issues with our passive mail node over the weekend doing vmtools updates and to resolve a problem we had to revert to a snapshot and then reseed all the databases across. All in all everything seemed fine, the server works and CCR copy status is running fine. I used the "Delete All" option this morning to remove the snapshot and the process according to vCenter has completed with no errors and no "Needs Consolidation" flag. This all seems fine until I check the Datastore that holds the VM on our SAN and I can clearly see snapshots that are pretty big [see attached image]. These do not seem to be changing size and the data modified is around the time the works were started for the vmtools update. Does this possibly mean that at some stage, possibly during reversion or hard resetting of the VM, that they have become orphaned? Are there any methods to check orphaned status of snapshots? We are running ESXi5.0 Update 1 with storage provide by an EMC SAN. Enterprise plus is the license level.

    Read the article

  • wget not converting links

    - by acrosman
    I am trying to mirror a fairly large site (20,000+ pages) prior to a major overhaul. Basically, I need a backup before cutting over to the new one in case we forgot something we need (we'll have about 1,000 pages at launch). The site is run on a CMS that I cannot easily extract usable data from, so I'm trying to make the copy with wget. My problem is that wget does not appear to be actually converting links, despite the presence of --convert-links or -k in the command. I've tried a couple of different combinations of flags, but I haven't been able to get the output I need. Most recent failed attempt was: nohup wget --mirror -k -l10 -PafscSnapshot --html-extension -R *calendar* -o wget.log http://www.example.org & I've also included the --backup-converted, and --convert-links instead of -k (not that it have mattered). I've done it with and without -P and -l, again no that they should matter. Results in files that still have links like: http://www.example.org/ht/d/sp/i/17770

    Read the article

< Previous Page | 330 331 332 333 334 335 336 337 338 339 340 341  | Next Page >