Search Results

Search found 15055 results on 603 pages for 'volume shadow copy'.

Page 364/603 | < Previous Page | 360 361 362 363 364 365 366 367 368 369 370 371  | Next Page >

  • Does RAID 1 protect against corruption?

    - by Shaun
    Does Raid 1 protect against data corruption? For example, let's say that I am keeping all of my important files on a NAS that uses 2 disks in a RAID 1. If one hard drive has some kind of internal problem and the data becomes corrupted, does the RAID recognize this automatically and correct it using data from the other good disk? Could it even know which copy is the good one? Does RAID 5 protect against corruption? I know that RAID is not a backup solution. I am trying to figure out how to make sure that I am not backing up corrupt data!

    Read the article

  • Hidden characters inserted after pipe (|) followed by a space

    - by nifty
    Very often, on my Mac, when I use the pipe (|) character followed by a space character, an invincible character will be inserted in between. This is especially annoying when using the terminal, as it makes commands invalid. If I type the following in iterm2, I often get the following: ls | cat zsh: command not found:  cat If I hit the up-arrow-key to get my previous command, and then remove and reinsert the space between | and cat, the command will work. When I copy paste the working and non working commands into a file, like this: non-working: ls | cat working: ls | cat and open it in Hex Fiend it shows the following: non-working: ls |¬†cat working: ls | cat I've also experienced the same kind of issue in SublimeText2 using the square brackets ([]) followed by a space. So I don't believe its an issue with iTerm2.

    Read the article

  • import macintosh thunderbird emails to windows email

    - by Ryan
    My mac laptop died. The hard drive is good. I am able to pull what I need off my mac and copy it to windows using an open source tool (I forgot the name of it). I need my old mac emails. I was using the macintosh thunderbird for emails. How do I load those emails into thunderbird, outlook, or any other windows email tool? I did google this and there were some brief explanations, but they did not work. The vast majority of posts are about going from windows to mac.

    Read the article

  • Using 64 bit wuauclt from 32 bit command prompt

    - by Tim Brigham
    I have a script that for legacy reasons needs to run inside a 32 bit command shell. This script also includes references to certain core windows binaries - most notably wuauclt but others as well - which are not accessible by default within the 32 bit environment. This script is being run in several locations including many windows 7 and server 2008 r2 boxes. I'm aware of the possibility to copy files from the system32 to syswow64 in order to get around this. Is there any better method - something along the lines of adding an entry to the path variable - which will allow me to fall back to these 64 bit binaries from within a 32 bit script?

    Read the article

  • How to integrate Thunderbird with SpamAssassin running on the server?

    - by haimg
    I'm trying to integrate SpamAssassin running on a server with Thunderbird. Basically I need to be able to select several emails in Thunderbird and send them back to SpamAssassin for training, either as spam or ham. I tried several approaches: Tried "Report Spam" plugin, which is able to send message back to server either as an email attachment or via HTTP post. However, the plugin is rather buggy... Does not support sending several messages at once, "report as ham" is not working at all, etc. Wanted to make a custom button that will copy selected messages to a separate IMAP folder (I could create "LearnAsSpam" and "LearnAsHam" folders in IMAP that will get processed automatically on server), but don't even know how to approach this in Thunderbird, don't want to learn Thunderbird extention authoring... Server-side, I'm prepared to do some custom programming or integration needed (can receive a message via HTTP / SMTP / whatever), my stumbling block is Thunderbird... So, how can I send emails from Thunderbird back to SpamAssassin running on email server for Bayesian training, with as few keystrokes as possible?

    Read the article

  • How do I keep multiple copies of Outlook in sync when using RPC over HTTP?

    - by Don
    I use Outlook 2007 at work with our Exchange 2003 server. I just setup my home system with Outlook 2007 so that I could use the RPC over HTTP to access Exchange without having to use a VPN. It works fine. I can get mail, send mail, etc. What it doesn't seem to be doing is staying in sync. For example, I read a few messages at home, moved them into different folders from the Inbox, etc. That all seemed fine. When I login to my work machine and look at the copy of Outlook there, the mail is still unread and nothing has been moved. Am I missing something simple here? I would have to assume that my home machine should be telling Exchange where these messages belong and that they've been read. Both machines are running Windows 7, if that matters. Ideas?

    Read the article

  • RoboCopy Log Files missing whitespace?

    - by TwystNeko
    So I'm working on a script to use RoboCopy to copy a bunch of files, and log what it's copied. It works reasonably well, except for the logfiles. They tend to look like this: C:\Users\Tech\Documents\desktop.initechscan1.jpgtechscan2.jpgtechscan3.jpgtechscan4.jpgtechscan5.jpgwsus.jpgwsus2.jpgC:\Users\Tech\Documents\My Music\C:\Users\Tech\Documents\My Pictures\C:\Users\Tech\Documents\My Videos\C:\Users\Tech\Documents\My Digital Editions\ As you can see, the log seems to be missing all whitespace and separators between entries. Is there something I can do to fix this? It's kind of frustrating. The commandline I'm using is this: C:\Users\Tech\Desktop>Robocopy.exe C:\Users\Tech\Documents c:\Temp /e /l /b /xj /xf ntuser.* desktop.ini *.lnk /np /njh /log:migratedfiles.txt /v I have the /l in there since I'm debugging, and it's the easiest way to keep from copying everything a million times.

    Read the article

  • How do I perform commands in another folder, without repeating the folder path?

    - by Valter Henrique
    Is there a clever way to do copy and move operations or a command to duplicate a file, without having to do a cd, then mv after, at the same folder? For example, I have to run the following: mv /folder1/folder2/folder3/file.txt /folder1/folder2/folder3/file-2013.txt Note that the directory to where I'm moving the file is the same, but I have to put the whole path again and sometimes it gets annoying. I'm curious to know if there's another way to do that without having to put the whole path again, because the operation would be done in the same path.

    Read the article

  • Download a file via HTTP from a script in Windows

    - by Jason R. Coombs
    I want a way to download a file via HTTP given its URL (similar to how wget works). I've seen the answers to this question, but I have two changes to the requirements: I'd like it to run on Windows 7 or later (though if it works on Windows XP, that's a bonus). I need to be able to do this on a stock machine with nothing but the script, which should be text that could be easily entered on a keyboard or copy/pasted. The shorter, the better. So, essentially, I'd like a .cmd (batch) script, VBScript, or Powershell script that can accomplish the download. It could use COM or invoke IE, but it needs to run without any input, and should behave well when invoked without a display (such as through a telnet session).

    Read the article

  • Checkout repo from SVN but use local files to populate

    - by aidan
    I have an SVN server on our development server, and I release to our production server using rsync. It not ideal, but it's worked so far. Anyway, I've finally got the SVN client installed on the production server and I want to start using that to copy files from development to production. My problem is this, I don't want to check all the data out of development when I already have it on the production server. Is there a way to "checkout" a repository, but use the files that are already on the production server (and force it to assume they are the head versions for example)? Thanks.

    Read the article

  • Installing W2K8 R2 on a Dell Poweredge 2850

    - by DerekT
    I'm a server novice and have been given a Dell 2850 (PERC 4e/Di). It has 3 blank HDs that I think are configured as RAID 5. I am trying to install W2K8 R2 SP1 on it. It doesn't have a DVD reader so I created an install USB. This works fine until it's time to copy files to the HD. It can't see a HD. There is an option to browse for drivers at this stage. Any idea what drivers I need and where to download them? I downloaded this driver RAID_DRVR_WIN_R227150.EXE but this failed with invalid signature. Thanks to dyasny for the link to LSI-LOGIC_MULTI-DEVICE_A00_R227150.exe.

    Read the article

  • Ubuntu 12.04 Server: permissions on /var/www for newly copied files

    - by Abe
    I ran the following commands to set up ACL on the /var/www folder in my Ubuntu 12.04 Server: sudo usermod -g www-data abe sudo chown -R www-data:www-data /var/www sudo chmod -R 775 /var/www I downloaded Wordpress using wget in my /var/www folder and unzipped the downloaded file: cd /var/www wget http://wordpress.org/latest.zip mv latest.zip wordpress.zip unzip wordpress.zip I created a new database and user in mysql and attempted to run the setup process through the web interface. When I enter the configuration info in wordpress I run into the following error message: Sorry, but I can't write the wp-config.php file. When I run ls -la, I see that the files are owned by my user abe, but they are part of the group www-data. Would I have to run the chmod command every time I copy new files to /var/www? sudo chmod -R 775 /var/www

    Read the article

  • Trigger ZFS dedup one-off scan/rededup

    - by Jake Wharton
    I have a ZFS filesystems which has been running for some time and I recently had the opportunity to upgrade it (finally!) to the latest ZSF version. Our data doesn't scream dedup but I firmly believe based on small tests that we could gain anywhere from 5-10% of our space back for free by utilizing it. I have enabled dedup on the filesystem and new files are slowly being dedupified but the majority (95%+) of our data already exists on the filesystem. Short of moving the data off-pool and then recopying it back, is there any way to trigger a dedup scan of existing data? It doesn't have to be asynchronous or live. (And FYI there isn't enough room on the pool to copy the entire filesystem to another and then just switch the mounts.)

    Read the article

  • Transparently cache files from a network drive in Linux

    - by Vadim
    We have a Linux server that reads files from a network drive and processes them. In a common scenario, a user will log in and access the same files over and over again. The size of the files varies but the larger ones can be around 50+ Mb. The files seldom change. I was wondering if it's somehow possible to transparently cache the files. I don't want (or can) change the program the reads the files, nor do I control the protocol by which the files are accessed. I just want something to detect that I access a certain path, copy the file locally (if needed) and then read the file from the local drive. I've read about Bcache but can't figure out if it's what I need. Do you have any suggestions? Thanks, Vadim.

    Read the article

  • VMware ESXi 4 On-Disk Data Deduplication - possible and supported?

    - by hurikhan77
    Environment: We are running multiple web, database, and application servers which usually share a pretty common installation (gentoo linux) and similar configuration in VMware ESXi 4. The differences are usually only some installed features or differing component versions. To create a new server, I usually choose the most similar (by features) running server, rsync a copy of it into freshly mounted filesystems, run grub, reconfigure and reboot. Problem: Over time this duplicates many on-disk data blocks which probably sums up to several 10's of gigabytes. I suppose if I could use a base system as template with the actual machines based on top of that, only writing changed blocks to some sort of "diff image", performance should improve (increased cache hit rate) and storage efficiency should increase (deduplicated storage space). This would be similar to what ESXi already supports for RAM deduplication (page sharing). Question: Is there any way to easily do this on ESXi 4? I already share the portage tree via NFS but this would not work for the rootfs.

    Read the article

  • Bootable SD card still has small memory, even after formating

    - by Inazuma
    I have an SD card which I used to run my RaspberryPi. I wanted to update the copy of raspbian on it, so I formated the card using the software from www.sdcard.com. I followed all the instructions correctly, however the size of my SD card didn't go back to it's default. It is a 4gb SD card, which after it's spell in the RaspberryPi had shrunken to 52mb, which I understand is normal. After formatting, the size rose to 3.69gb. This means that there is not enough space to install a new OS, so how can I make my SD card 4gb again? Any help would be much appreciated!

    Read the article

  • Software to cache a web application for use offline

    - by littlecharva
    My boss quite regularly has to demo our web application to clients in a situation with no wifi available and sketchy 3G access, quite often the 3G lets him down. I have considered setting a copy of our server up in a virtual machine on his laptop so he could demo it offline, but I fear this will just introduce more headaches when he forgets how to boot the VM up. What I'd ideally like is an app that records you logging into a web app, saves copies of all the pages and ties the links and buttons you click up to offline copies of the pages it saves. So you could run through the demonstration you're going to give and have it cache the pages. When you then click the same buttons and links in offline mode it will present the relevant offline pages. Does such a thing exist? Can anyone recommend any alternative solutions to this problem? Thanks, Anthony

    Read the article

  • Secure copying (file transfer) between two Linux servers in the same datacenter (Linode)

    - by MountainX
    I have two Linodes in the same data center. I want to copy files from one to the other each night or on demand (for about the next month, until this project is finished). So I'm thinking about using rsync. My question is how do I set up the two Linode servers to communicate via private IP addresses securely? Both servers are SSH hardened, they use denyhosts and have a fairly restrictive iptables setup. I know I need to first assign private IP addresses to each server, then configure static networking according to this guide. What is next? What SSH or iptables settings are needed to allow these two servers to communicate? What further info do I need to supply in this question? I'm looking for a basic step-by-step guide for how to do this.

    Read the article

  • Offline Outlook 2007 global address book slow to update

    - by munrobasher
    Outlook 2007 talking to an Exchange 2007 server. Usual set-up of personal contacts and a site wide global address book. Distribution lists are often created by IT in the global address book but it sometimes takes days for them to appear in the local offline copy. Performing a manual download of the address book doesn't help. Problem doesn't occur with non-offline mode of Outlook 2007. Any ideas? Server or client side issue? Cheers, Rob.

    Read the article

  • Pull row of data from one place in spreadsheet to another

    - by bookcasey
    I have a list of names and their assorted properties. I am trying to find a way to quickly move an entire row's worth of data into a different sheet so I can manipulate it. Currently I can accomplish this by copy and pasting the entire row, but this is not elegant nor dynamic (the number in blue does not update with this method): My ideal solution: A name is typed into the "Name" column. The spreadsheet program pulls the associated row from the "main" sheet into the new sheet. I am currently using Numbers.app, but am curious if this is possible in Excel, OpenOffice Calc, Google Docs, or any other spreadsheet program. Thank you.

    Read the article

  • Fake domain doesn't resolve when offline

    - by Fletcher Moore
    I have a flimsy grasp of DNS. Nonetheless, in order to install a local development copy of Wordpress MU, I needed to create a fake domain, which I called local.dev. It and all subdomains simply resolve to 127.0.0.1. Apache then directs to the correct folder. I installed PowerDNS, and got it working properly with a MySQL backend. I didn't feel comfortable, but since it worked, I didn't ask any more questions. The bizarre thing is it requires an internet connection to resolve correctly, and now I need to use it offline. If I am offline, Chrome provies the error: Error 105 (net::ERR_NAME_NOT_RESOLVED): The server could not be found. If you need more information, I am happy to provide it.

    Read the article

  • How to choose my own filename format for subscribed podcast files?

    - by meomaxy
    I subscribe to several podcasts where the filenames of the downloaded mp3 files have no particular pattern to them. When I copy the directory of accumulated mp3 files into my mp3 player, the files play in alphabetical order. What I really want is to play the files chronologically by release date. I currently use iTunes on Windows XP to download the files. What I do now is manually rename the files, adding the date in YYYYMMDD format to the start of each filename so that an alphabetical listing of files will correspond to their chronological order when I listen to them later in PocketTunes on my Palm Centro. Is there some way to get the release date into the filename automatically? If so, I could automate or possibly skip the file renaming step. I would switch from iTunes to something else if that would solve my problem. The file creation time on my local disk isn't a reliable indicator because sometimes I download a few days worth of content at one time, and the files don't necessarily get downloaded in chronological order.

    Read the article

  • Can someone correct me about Windows 2008 DFS implementation

    - by cwheeler33
    I have 2 two Windows 2003 file servers using DoubleTake at the moment. They are in an 2003 AD domain. And each server has it's own disk set. It is time to replace the servers... I want to use Windows 2008 64bit Ent. I was thinking of using DFS-R to replace DoubleTake. The part I'm not sure about is clustering. Do I need to have shared storage if each server has a copy of the data? I want to have the data available to the same share name, so maybe I don't fully understand how DFS is set up. I currently have about 6TB of data. I expect to grow by 3TB a year on these file servers. Any resources/books that could teach me would be good to know as well.

    Read the article

  • Notepad/Edit equivalent for Linux command line

    - by Jason Kester
    I'm looking for a simple text editor that I can use from the command line in Linux to edit files. I'm used to editing files in windows, so I'm looking for something with the same keyboard interface. That means: SHIFT+Arrow Keys/PGUP/PGDN to select text CTRL+C, CTRL+X, CTRL+V to copy/cut/paste And that's pretty much it. Surprisingly I'm having a tough time finding something like this. Vi/emacs are naturally out. Nano comes close, but has its own non-standard cut/paste/select keyboard shortcuts. Surely this thing exists somewhere. Thanks in advance for pointing me in the right direction.

    Read the article

  • Subversion error: (405 Method Not Allowed) in response to MKCOL

    - by Sergio del Amo
    I am getting the following error while trying to commit a new directory addition. svn: Commit failed (details follow): svn: Server sent unexpected return value (405 Method Not Allowed) in response to MKCOL request for '.... I have never seen this error before. How can I fix this problem? Solution I managed to solve the problem: Delete the parent's directory of the folder giving the problem. Do SVN Update. A folder with the same name as the new one already existed in repository. Delete this folder. SVN commit. Copy the new folder, schedule for addition and SVN commit.

    Read the article

< Previous Page | 360 361 362 363 364 365 366 367 368 369 370 371  | Next Page >