Search Results

Search found 962 results on 39 pages for 'tar'.

Page 2/39 | < Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >

  • "Stream" tar action from remote server through Putty?

    - by Pekka
    I am accessing a shared hosting account through Putty / SSH. The account is pretty full, 300 MB below its quota. I need to make a full backup of all data present on the account. Creating a bzip2 file using tar cjf archive.tar.bz2 directory/* fails because there is not enough space for the tar. Does anybody know a way to create a tar file and "stream" it to a local file on my PC? Through putty or any other SSH tool?

    Read the article

  • tar a directory and only include certain file types

    - by Susan
    Following the instructions here: http://linuxdevcenter.com/pub/a/linux/lpt/20_08.html I'm aiming to tar up a directory but only want to include .php files from that directory. Given the aforementioned instructions, I've come up with this command. It creates a file called IncludeTheseFiles which lists all the .php files, then the tar is supposed to do it's job only using the files listed in IncludeTheseFiles find myProjectDirectory -type f -print | \ egrep '(\.[php]|[Mm]akefile)$' > IncludeTheseFiles tar cvf myProjectTarName -I IncludeTheseFiles However, when I run this it doesn't like the I include option? tar: invalid option -- I

    Read the article

  • How do I tar dot files but not dot directories

    - by bjackfly
    The following tar command will exclude all dot files and dot directories. tar -cvzf /media/bjackfly/bkup/bkup.gz --exclude '.*' --one-file-system /home/bjackfly In my case I want the dot files to be backed up in the home directory (.vimrc, .bashrc) etc. but not the dot directories /.config /.cache /.eclipse etc. Any Linux gurus with a command for this, or do I need to run a find into a tar or do two different tar commands which is non-ideal? One for dot files in the home directory and one for everything else?

    Read the article

  • Tar exclude not working?

    - by Andrew Fashion
    tar -cvf file.tar --exclude=thumbs/ \ --exclude=uploads_event/ \ --exclude=uploads_forum/ \ --exclude=uploads_admin/ \ --exclude=uploads_userpoints/ \ --exclude=uploads_group/ \ --exclude=up_old/ \ --exclude=uploads_user/ \ --exclude=uploads_wall/ \ directory_to_tar/ Do I have it wrong? Trying to tar entire directory but exclude all those directories and any files in those folders completely.

    Read the article

  • Ubuntu: Tar doesn't work correctly

    - by Phuong Nguyen
    It looks like tar is having some problems. I installed Ubuntu 10.04 alpha a few of weeks ago. After that, this alpha version is so terrible that I must switched back to 9.10. So, I backed up all of my profiles data (/home/my_user_name) to my_user_name.tar.gz Here what I did: (in U10.04) Open Nautilus and goto /home/my_user_name Press Ctrl+H to view all hidden files. Press Ctrl+A to select all files Right click and choose [Compress...] Now, when I have set up Ubuntu 9.10 again, I extract the tar using Extract Here command from Nautilus. Funny things happened: Instead of extracting to current folder, the archive manager create a folder named my_user_name and put the extracted content into it. All of the files that I placed directly under /home/my_user_name doesn't get extracted All of the directories that started with . (dot) is not extracted. I wonder if there is any incompatibility between tar in U10.04alpha and U9.10 that cause the problem? Now every of my email, basket data is long gone. I'm freaking right now. Is there anything I can do to get the tar return back my data?

    Read the article

  • Forcing tar to create an empty archive

    - by snostorm
    I'm trying to use tar to tar files before transfer, so I can keep the entire file path rather than losing it along the way. However, when I try to tar an empty folder, it tells me that it is cowardly refusing to create an empty archive. I want to keep the empty folder on the other end, but don't want to put anything else into the archive to make it non-empty. Is there any way to do this?

    Read the article

  • Tar dereference only 1 level

    - by Bart van Heukelom
    I use the following pseudo-script to create a TAR of my installed software mkdir tmp ln -s /path/to/app1/bin tmp/app1 ln -s /and/path/going/to/the-app-2 tmp/app2 tar -c --dereference -f apps.tar tmp I need the --dereference option here to follow the links I just made in tmp. The reason I make the links in the first place is to store the directories with a different name in the archive than they have on the filesystem. Until now it has worked fine. However, I now have the situation that /path/to/app1 also contains links, and those I don't want to follow. Is this possible with some changes to the tar command? Or do I need to completely switch around the way I build the archive?

    Read the article

  • Corrupt tar (Resulting folder smaller than packed file)

    - by Om3g4
    Hi everybody, I have an issue with a tarball created on a SuSe 10.3 Server version. The .tar file has a size of 6.5 GB but if I untar it under Ubuntu 9.10 the resulting folder only has a size of 1.5 GB. commands used: tar cvf for packing, tar xvf for unpacking. Perhaps somebody knows how this can be fixed, would be great, cheers

    Read the article

  • Not possible to extract tar.gz archive in Centos

    - by Petuga
    I have just uploaded tar.gz archive to my Centos server from my home Windows 7 computer. Before that, it was extracted from .zip archive. When I try to extract it: tar -xvf file.tar.gz gunzip file.tar.gz It always give me error: "not in gzip format" The archive is not corrupted, because it is possible to extract it in Windows with Winrar. The archive is quite big, more than 100 MB. I don't know, what's the problem. Could you help me please? Thanks. Peter

    Read the article

  • extracting files from tar

    - by shantanuo
    tar -xvf company_raw_2012-03-16.tgz --directory=/root/test --strip-components=4 I am using the following tar option to remove the leading directories and it is working as expected. --strip-components NUMBER strip NUMBER of leading components from file names before extraction It works only when I know that there are going to be 4 sub-directories. I have tar files and I do not know if there will be 2, 3 or 4 folders inside. How do I strip the entire path and extract files in the given "directory" path.

    Read the article

  • How to make TAR to not save the directory hierarchy

    - by Nerian
    system("tar -czf #{RAILS_ROOT}/tmp/export-result #{RAILS_ROOT}/tmp/export") When I decompress the resulting file I get app/c3ec2057-7d3a-40d9-9a9d-d5c3fe3ffd6f/home/tmp/export/and_the_files I would like to just get: export_result/and_the_files How do I change my TAR call to accomplish this? solution: system("tar -czf #{RAILS_ROOT}/tmp/export.tgz --directory=#{RAILS_ROOT}/tmp export/")

    Read the article

  • Why /home folder is missing from my backup archive created by tar?

    - by Konstantin Pereyaslov
    So I'm doing full backup of my VPS using the following command (as root, of course): tar czvf 20120604.tar.gz / Everything seems to be fine, all files seem to appear in the list. The size of archive is 6 Gb and gunzipped version is 11 Gb which includes /home, because I totally have 11 Gb of data on VPS. But when I try actually to unpack archive, or open it using mc or WinRAR, there's no /home folder. And WinRAR tells 20120604.tar.gz - TAR+GZIP archive, unpacked size 894 841 346 bytes. It can't be WinRAR's bug, because when I type tar xzvf 20120604.tar.gz, /home folder isn't unpacked either. Why is /home folder missing from my archive? And what can I do to include it there? tar --version outputs the following: tar (GNU tar) 1.15.1

    Read the article

  • Is there a way to diagnose a tar file when, on extract, files are missing but no errors are given?

    - by ljvillanueva
    I have tar files where I archive about 250 files, each about 80 Mb, without compression. In a few cases tar is only returning some of the files. For example, when doing an extract of the file using: tar -xvf 356.tar I got only 103 files, when it should return 255 files, but tar does not give me an error. Furthermore, the tar archive is 15.8 Gb while the extracted folder is just 6.4 Gb. The tar files were created using tar -cvf 356.tar 356 where 356 is the name of the folder. All the steps where done in the same machines, under Ubuntu 6 and newer. Any ideas if there is a way to recover the files that are not being extracted?

    Read the article

  • Using find and tar with files with special characters in the name

    - by Costi
    I want to archive all .ctl files in a folder, recursively. tar -cf ctlfiles.tar `find /home/db -name "*.ctl" -print` The error message : tar: Removing leading `/' from member names tar: /home/db/dunn/j: Cannot stat: No such file or directory tar: 74.ctl: Cannot stat: No such file or directory I have these files: /home/db/dunn/j 74.ctl and j 75. Notice the extra space. What if the files have other special characters? How do I archive these files recursively?

    Read the article

  • tar incremental backup is backing everything up, every time when used on the Dropbox directory

    - by Cyclic
    I made an incremental backup about 10 months ago (on Jan 27, 2013), creating a .snar metadata file. Now, when I try to make an incremental backup using tar --create --file=dropbox_incremental_1.tar --listed-incremental=dropbox_0.snar Dropbox the command just re-backs up everything. I'm not an expert at Unix timestamps, but I noticed that virtually all of my directory timestamps are way more recent than the last time they changed. For my actual files, they look like this: Access: 2013-03-12 19:04:51.000000000 -0500 Modify: 2012-09-30 15:10:47.000000000 -0500 Change: 2013-03-12 19:04:51.306209672 -0500 The 'Modify' timestamp seems correct, but the files were definitely not changed (at least not doing anything that I know of) at the time they say they were. These files still seem to go into the incremental archive. What's happening here? Is there a way to tell tar to look at the 'modify' timestamp? Isn't this what it's supposed to be doing?

    Read the article

  • tar incremental backup is backing everything up, every time

    - by Cyclic
    I made an incremental backup about 10 months ago (on Jan 27, 2013), creating a .snar metadata file. Now, when I try to make an incremental backup using tar --create --file=dropbox_incremental_1.tar --listed-incremental=dropbox_0.snar Dropbox the command just re-backs up everything. I'm not an expert at Unix timestamps, but I noticed that virtually all of my directory timestamps are way more recent than the last time they changed. For my actual files, they look like this: Access: 2013-03-12 19:04:51.000000000 -0500 Modify: 2012-09-30 15:10:47.000000000 -0500 Change: 2013-03-12 19:04:51.306209672 -0500 The 'Modify' timestamp seems correct, but the files were definitely not changed (at least not doing anything that I know of) at the time they say they were. These files still seem to go into the incremental archive. What's happening here? Is there a way to tell tar to look at the 'modify' timestamp? Isn't this what it's supposed to be doing?

    Read the article

  • How to extract nested tar.gz files easily?

    - by StarCub
    Hi, I need to extract tar.gz a file. It's about 950mb. It has another 23 tar.gz files in it. Each of those 23 tar.gz files has one tar file in them. My questions is how I can easily extract all of them? Is there a commandline tool that I can use? The structure is like the following: foo.tar.gz +---bar1.tar.gz ¦ +---foobar1.tar +---bar2.tar.gz ¦ +---foobar2.tar +---bar3.tar.gz ¦ +---foobar3.tar +---bar4.tar.gz ¦ +---foobar4.tar +---bar5.tar.gz ¦ +---foobar5.tar +---bar6.tar.gz ¦ +---foobar6.tar | .......... | .......... | .......... | 23 of them Thanks in advance.

    Read the article

  • Estimate compressed file size in tar.gz

    - by liori
    I've got a set of .tar.gz files, which are duplicity backup files (either full backups or incremental ones). I'd like to compute which directories take the most space on backups. This will most probably be a different figure to calculating which directories take the most space on a live filesystem because I need to account for how often are files changing (and therefore taking space on incremental backups) and how compressible are files. I know that while many other archive formats store compressed files as different entities inside the archive file, .tar.gz files do not, and therefore it is impossible to get an exact amount of storage taken in the archive by a single file after compression. Are there any tools to calculate at least some estimates?

    Read the article

  • Creating diff from a tar

    - by Hulk
    There is a tar file which was created few days before, this tar file was created on the /files directory. And now /files directory has new files uploaded in it. My question is how to create another files with only the new files uploaded. Thanks..

    Read the article

  • cpio VS tar and cp

    - by Tim
    I just learned that cpio has three modes: copy-out, copy-in and pass-through. I was wondering what are the advantages and disadvantages of cpio under copy-out and copy-in modes over tar. When is it better to use cpio and when to use tar? Similar question for cpio under pass-through mode versus cp. Thanks and regards!

    Read the article

  • What command to use to unpack a .tar.gz file in Windows?

    - by Ivan
    I use to receive tar.gz files fromm colleagues but unpacking them with 7zip is not convenient as it unpacks a tar file from gz first while I usually seek to get the tar file contents (so I have to unpack the tar file explicitly next). I've tried adding tar -zxvf %P%N command to Total Commander button bar but it turned that tar -zxvf doesn't work but returns the following error (tried it in bare command prompt): tar: Cannot fork: Function not implemented tar: Error is not recoverable: exiting now I have got both GnuWin32 and cygwin installed, but I seek a solution that will work in Windows command line promt (and so will be callable from Totat Commander), not in cygwin command line promt.

    Read the article

  • cygwin tar omits directory when creating archive

    - by Dean Schulze
    I have the following directory structure books/java books/java.ebooks books/java.gwt books/java.swing books/javaee I use this command to create an archive of all the directories starting with the letters a-m tar czf books.a-m.no.AI.tgz --exclude=books/AI.and.algorithms --exclude=books/[N-Z]* --exclude=books/[n-z]* books The .tgz archive that gets created does not contain the java.swing/ directory. Am I missing something in my use of tar or is this a bug in cygwin?

    Read the article

< Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >