Search Results

Search found 38781 results on 1552 pages for 'offline files'.

Page 84/1552 | < Previous Page | 80 81 82 83 84 85 86 87 88 89 90 91  | Next Page >

  • Untar after uploading to linux from windows

    - by Miqdad Ali
    I have created tar.gz from my linux server, and I downloaded the same to my linux system and I successfully doen untar with tar -xvf package.tar.gz. And now my issue I downloaded same package.tar.gz to the windows system then uploaded to another linux server, and tried same command tar -xvf package.tar.gz. but it getting tar: This does not look like a tar archive tar: Skipping to next header tar: Exiting with failure status due to previous errors as response. I also tried filezill manual trnasfer with binary mode. How can I do the same ? Update When I directly download to the linux system its working fine. When I downloaded to the windows system and try to extract with 7zip or winrar error is getting When I download to windows and upload to linux same error getting

    Read the article

  • Design of input files reading when it comes to defaults/transformations

    - by Stefano Borini
    Suppose you have an application that reads an input file, on a language that does not support the concept of None. The input is read, parsed, and the contents are stored on a structure for later use. Now, in general you want to keep into account transformation of the data from the input, such as adding default values when not specified, or adding full path information to relative path specified in the input. There are two different strategies to achieve this. The first strategy is to perform these transformations at input file reading time. In practice, you put all the intelligence into the input parser, and your application has no logic to deal with unexpected circumstances, such as an unspecified value. You lose the information of what was specified and what wasn't, but you gain in black-boxing the details. Your "running code" needs that information in any case and in a proper form, and is not concerned if it's the default or a user-specified information. The second strategy is to have the file reader a real one-to-one mapper from the file to a memory-stored object, with no intelligent behavior. unspecified values are not filled (which may however be a problem in languages not supporting None) and data is stored verbatim from the file. The intelligence for recovery must now go into the "running code", which must check what was specified in the file, eventually fall back to a default, or modify the input properly before using it. I would like to know your opinion on these two approaches, and in particular which one you found the most frequently implemented.

    Read the article

  • Can't rename/move files from OSX that were copied from NTFS

    - by 99miles
    Hello- I recently had data recovered and it was sent back to me on what I think is an NTFS drive. I copied all the files over to a file share I have on a Linux box, that's ext4. Now I have that share mounted on my OSX machine, and I can't move or rename most of the files. However, in a couple cases I was able to rename a folder after the third try. Another time I was able to rename a folder once, but not again. All the permissions are showing up the same on the command-line -- I can't see any differences between the permissions on any of the files/folders. Any clues??? Thanks.

    Read the article

  • Excel crashes when opening Excel files from Internet Explorer

    - by Rob Z
    I have been running into some issues when opening Excel files from Internet Explorer, generally the first document or two will open fine but after that trying to open a file will cause Excel and Internet Explorer to crash to the desktop without any notifications being given. This doesn't happen for users who are running Excel 2007, but for users with Excel 2003 it may or may not happen to them. The files in question are Excel XML files and Internet Explorer 6 and Excel 2003 are being use. At this time it would not be possible to upgrade Internet Explorer, but it would be able to upgrade to Excel to version 2007 if that would resolve the issue. Overdue Update: We recently upgraded to Firefox at the office which has rendered this error a non-issue; however, it is still unresolved from the standpoint that we haven't been able to come up with an explanation to the issue. Since IE6 is still installed on the systems, a fix to the problem (or explanation of why it's happening) would be appreciated.

    Read the article

  • Apache 2.2 Windows XP Uninstall Doesn't Remove All Files

    - by AJ
    Hello, I am trying to uninstall Apache 2.2 on Windows XP. The original installation was from the binary msi distribution. The uninstall function of the msi ran successfully, but it failed to remove a couple of folders: C:\Apache2.2\conf C:\Apache2.2\logs I am unable to remove these folders manually because they contain files for which I do not have ownership. And this is the source of my confusion: why if I installed the program do I not have permission to remove it? To be clear, I do not have local admin rights (nor can I request them), yet the files in these two remaining folders are owned by Administrator. How is it that Administrator created these files (and how can I possibly remove them)? Thanks, -aj

    Read the article

  • SCO Files Motion for Judgment As a Matter of Law, or For a New Trial

    <b>Groklaw:</b> "SCO has filed its "renewed" motion for judgment "as a matter of law", with its supporting memorandum. They ask the judge to rule over the heads of the jury and decide that the jury "simply got it wrong" when it ruled that SCO didn't get the copyrights in 1995 from Novell. In the alternative, they'd like a new trial."

    Read the article

  • Aptana Ext code completion in .php files

    - by Frederik Wordenskjold
    I'm having problems getting the code-completion for the Ext 3.2 plugin to work, when working with .php files. I've also installed the php plugin for Aptana, and the same thing applies for php - I cannot access php code-completion when working with a html-file, so it seems like a general issue... It's also not possible to write Ext in .js files, which is weird... The latter case does make sense of course. But I should be able to write both php and Ext in .php files! Is this possible in any way? I have of course tested the code-completion for php in a .php file, which works. The same applies for Ext code in .html documents!

    Read the article

  • In *nix, how to determine which filesystem a particular file is on?

    - by smokris
    In a generic, modern unix environment (say, GNU/Linux, GNU/Solaris, or Mac OS X), is there a good way to determine which mountpoint and filesystem-type a particular absolute file path is on? I suppose I could execute the mount command and manually parse the output of that and string-compare it with my file path, but before I do that I'm wondering if there's a more elegant way. I'm developing a BASH script that makes use of extended attributes, and want to make it Do The Right Thing (to the small extent that it is possible) for a variety of filesystems and host environments.

    Read the article

  • Loading files during run time

    - by NDraskovic
    I made a content pipeline extension (using this tutorial) in XNA 4.0 game. I altered some aspects, so it serves my need better, but the basic idea still applies. Now I want to go a step further and enable my game to be changed during run time. The file I am loading trough my content pipeline extension is very simple, it only contains decimal numbers, so I want to enable the user to change that file at will and reload it while the game is running (without recompiling as I had to do so far). This file is a very simplified version of level editor, meaning that it contains rows like: 1 1,5 1,78 -3,6 Here, the first number determines the object that will be drawn to the scene, and the other 3 numbers are coordinates where that object will be placed. So, how can I change the file that contains these numbers so that the game loads it and redraws the scene accordingly? Thanks

    Read the article

  • What does 'http: getaddrinfo*.gaih_getanswer: got type "46"' mean

    - by koffie
    Today I got an e-mail from logcheck informing me that the following system event occurred. http: getaddrinfo*.gaih_getanswer: got type "46" Indeed, the above message occurs 4 times in /var/log/aut.log it occured twice yesterday around 6:46 am. And this morning it also occured twice around the same time. I wonder what this message means and if this message is something to worry about or if should just tell logcheck to ignore it. If it is something to worry about, then any pointers on how to fix the problem are appreciated.

    Read the article

  • How to uncompress a 9GB file in Windows FAT32

    - by Kashif
    I have a 2GB RAR file that contains a 9GB video file. I'm using a FAT32 file system. Now I want to unzip that file but after 4GB I get an error due to the FAT32 file size limit. Now I want to know that how I can extract that video? I know that one way is to convert my partition to NTFS but I don't want to follow that way. I've also tried 7-zip but that again gives error after 4GB. One other way is to split that file but I don't know how I can split a video file that is zipped. So any idea please? How can I get rid of this problem.

    Read the article

  • Can't set permissions for files on an NTFS partition

    - by ashishsony
    I remember that I was able to run a Linux .exe that was placed on an NTFS partition earlier before I installed 10.10 RC. But if I try to run it now, I can't run it as it hasn't the execution permission. The bad part is that I can't change the permissions too. I'm chmod-ding +x but no change at all with its permissions. So this seems to be a bug? Any help? Though when I put it on ext4 partition, I can set the permission. But I want to do this as I did before, right from its default NTFS location.

    Read the article

  • Migrating data from Oracle database to Pervasive .DAT files

    - by kaychaks
    The requirement is to migrate some tables with data from a Oracle database server to Pervasive database's .DAT file. Then those .DAT files will be used by a Pervasive database server. The restriction is that Oracle DB can not directly migrate to the Pervasive DB. It has to generate the .DAT files and then the new .DAT files will replace the old one for the Pervasive DB which will then use them for the new data. I was trying this task with SSIS. Exporting the Oracle table to a delimited .txt file and then creating a .DAT file from that text file. I can export the data from Oracle to .txt but I am not finding any way to migrate .txt to Pervasive .DAT? Is this the right approach? If not then please help with my problem.

    Read the article

  • Unable to install files with apt-get: "unable to locate package"

    - by Ben Casling
    I'm having issues with my ubuntu server version 12.04 installed on a HP550 laptop, when i try sudo apt-get install <programname>, e.g apache2 it will not work, saying E: Unable to locate package apache2. I have tried to look/edit the sources. but they will not work either the gedit command is broken too, i am trying gedit /etc/apt/sources.list for those wondering, is this a case of the computer network not configured properly? it downloaded a language pack easily enough in the installation though. how do i fix this? a prompt reply would be appreciated.

    Read the article

  • Backup Windows files using Ubuntu - Unable to find Win partition

    - by Siva
    I am using a Dell laptop with Windows 7, all of a sudden the HDD is not recognized by Win7. I wanted to backup the data in Win, so I made a Ubuntu 12.04.1 Live CD, and booted from it. I am using Ubuntu without installing it in my laptop. My problem is that I don’t see the Windows partitions in Ubuntu 12.04.1, b’cos of which I am unable to backup the data. Any suggestion in this regard would be very helpful.. PS: I checked the SMART status of the HDD it says 2 Bad sectors, when I attempted an extended Self-Test, I get a Read Failed message, though the Short test goes through fine. Thank You, Siva

    Read the article

  • SQL Server Files Local or NAS or SAN?

    - by Jedi Master Spooky
    I have to install a new Server with SQL Server 2008, What do you recommend, One server with Raid 10 or the Files in a NAS? What about iSCSI should I use it? What about SAN? The server has 4Gb of RAM and that database file is about 2GB. To make my self clear today the server has no RAID, I have to implement some kind of strategy so if something happend I can have my files safe, so What should I choose Local Files, NAS, SAN? What option has the most performance, what is the more secure?

    Read the article

  • What is it that automatically checks config changes (such as those in /etc) into git?

    - by Brandon
    I remember reading on the ubuntu forums some time ago about a program to automatically check configuration changes into version control for you. It was (of course) not Ubuntu-specific. I'm pretty sure it used git, though it may have been svn, or perhaps even able to work with multiple different VCSs. My Googling has turned up nothing, and I'd rather not roll my own script if someone has already done this well. Of course I could just manually check things in, but there are reasons I'd like it done automatically. (I'm actually planning to use this for my LastSession.plist file for Safari, so when the #@$%^*&! thing crashes, and I don't restore everything, and then Leopard crashes, the fact that it has such lousy session management won't mean I lose the dozens of windows with dozens of tabs I had open.)

    Read the article

  • updater stuck on downloading python files

    - by Tim
    However i am a java programmer is consider myself as a linux noob. So i could use a little help here... I am trying to update my ubuntu from version 10.04 to 12.04.1 (LTS). The downloading start and runs at around 10MB/s untill i am somewhere near 26% and the speed just drops to 0. I figured i could use the command "do-release-update" so i could see what it was doing. Again the same occured: the downloading stucked on 26% at "http://nl.archive.ubuntu.com/ubunutu/ precise/main python-qt4 1386 4.9.1-2ubuntu1". It says it downloaded 41% of that file/package. It also tries another wget on the same file every X seconds. Help? Greetings Tim, Holland.

    Read the article

  • How to move files over samba share with gnomevfs cli

    - by Allan
    Ok I am in the process of backing up my film collection to a NAS and I wanted to automate this as much as possible as I have to work at the same time. I am trying to setup a daily dump of ISO's ready to be converted overnight. I would like to do this as a cron job using gnomevfs. I have been able to connect and do an ls command successfully with gnomevfs-ls smb://user:WORKGROUP:password@media-centre/videos/ but I am having trouble setting up a mv command from a local folder to the same shared folder keep getting the Usage: gnomevfs-mv <from> <to> quote which isn't particularly informative ;) any ideas?

    Read the article

  • How to Run files and open folders as an administrater 11.10

    - by Mattlinux1
    Solved!!! Enable open as administrator in nautilus! To get started, press Ctrl – Alt – T on your keyboard to open Terminal. When it opens, run the commands below: enter code here sudo apt-get install nautilus-gksu After installing that application run copy and paste the line below and press enter. sudo cp /usr/lib/nautilus/extensions-2.0/libnautilus-gksu.so /usr/lib/nautilus/extensions-3.0/ Finally, log out and back in then go and test it, by clicking your right mouse click on the file you want to run as Admin and your see a popup menu with the words: open as administrater. Visit: http://www.liberiangeek.net/2011/12/add-open-as-administrator-to-nautilus-context-menu-in-ubuntu-11-10-oneiric-ocelot/ for help and from here. Enjoy!

    Read the article

  • Complete stack traces from Hyperic

    - by Mike Kushner
    I've setup Hyperic to run on our CI-machine, and every once in a while it reacts to some random stack trace and sends of an alert. So far so good, we've caught a lot of intermittent bugs that way. My only issue is that the alert only contains the first error line and not the entire stack trace, which requires me to access the machine and look at the logs manually. Is there any way to modify the alert message to contain more information, alternatively to include the log file in the alert mail?

    Read the article

  • Possibility of recovering files from a dd zero-filled hard disk

    - by unknownthreat
    I have "zero filled" (complete wiped) an external hard disk using dd, and from what I have heard: people said you should at least "zero fill" 3 times to be sure that the data are really wiped and no one can recover anything. So I decided to scan the disk once again after I've zero filled the disk. I was expecting the disk to still have some random binary left. It turned out that it has only a few sequential bytes in the very beginning. This is probably the file structure type and other headers stuff. Other than that, it's all zeros and nothing else. So if we have to recover any file from a zero filled disk, ...how? From what I've heard, even you zero fill the disk, you should still have some data left. ...or could dd really completely annihilate all data?

    Read the article

  • apt-get doesnt download files from NFS location

    - by Pravesh
    I have switched to unix from last 3 months and trying to understand install process and in particular apt-get. I am able to successfully install and download the packages when I configure my repository on http location in /etc/apt/sources.list file. e.g. deb http://web.myspqce.com/u/eng/rose/debian-mirror-squeeze-amd64/mirror/ftp.us.debian.org/debian/ squeeze main contrib non-free This command will download(/var/cache/apt/archive) and install the package when i use apt-get install When I change the source location to file instead of http(nfs mount point), the package is getting installed but NOT getting downloaded in /var/cache/apt/archive. deb file:/deb_repository/debian-mirror-squeeze-amd64/mirror/ftp.us.debian.org/debian/ squeeze main contrib non-free Please let me know if there is any configuration or settings that i have to make to let apt-get to both download and install package when i use (nfs)file:/ instead of http:/ in sources.list. To achieve this, I can use apt-get --downlaod-only and then use apt-get install for both download and install in two separate calls, but I want to know why package is not getting downloaded with apt-get install but only getting installed when used with file:/ in sources.list

    Read the article

  • access_log item w/out IP. Starts with "::1 - - [<date>]"

    - by Meltemi
    Looking at our Apache log I see normal requests like: 174.133.xxx.xxx - - [20/May/2010:17:36:44 -0700] "GET /index.html HTTP/1.1" 200 2004 but every so often i get a cluster of these w/out an IP address. ::1 - - [20/May/2010:18:47:21 -0700] "OPTIONS * HTTP/1.0" 200 - ::1 - - [20/May/2010:18:47:22 -0700] "OPTIONS * HTTP/1.0" 200 - ::1 - - [20/May/2010:18:47:23 -0700] "OPTIONS * HTTP/1.0" 200 - what do they mean and curious what causes them?

    Read the article

  • How to handle multiple effect files in XNA

    - by Adam 'Pi' Burch
    So I'm using ModelMesh and it's built in Effects parameter to draw a mesh with some shaders I'm playing with. I have a simple GUI that lets me change these parameters to my heart's desire. My question is, how do I handle shaders that have unique parameters? For example, I want a 'shiny' parameter that affects shaders with Phong-type specular components, but for an environment mapping shader such a parameter doesn't make a lot of sense. How I have it right now is that every time I call the ModelMesh's Draw() function, I set all the Effect parameters as so foreach (ModelMesh m in model.Meshes) { if (isDrawBunny == true)//Slightly change the way the world matrix is calculated if using the bunny object, since it is not quite centered in object space { world = boneTransforms[m.ParentBone.Index] * Matrix.CreateScale(scale) * rotation * Matrix.CreateTranslation(position + bunnyPositionTransform); } else //If not rendering the bunny, draw normally { world = boneTransforms[m.ParentBone.Index] * Matrix.CreateScale(scale) * rotation * Matrix.CreateTranslation(position); } foreach (Effect e in m.Effects) { Matrix ViewProjection = camera.ViewMatrix * camera.ProjectionMatrix; e.Parameters["ViewProjection"].SetValue(ViewProjection); e.Parameters["World"].SetValue(world); e.Parameters["diffuseLightPosition"].SetValue(lightPositionW); e.Parameters["CameraPosition"].SetValue(camera.Position); e.Parameters["LightColor"].SetValue(lightColor); e.Parameters["MaterialColor"].SetValue(materialColor); e.Parameters["shininess"].SetValue(shininess); //e.Parameters //e.Parameters["normal"] } m.Draw(); Note the prescience of the example! The solutions I've thought of involve preloading all the shaders, and updating the unique parameters as needed. So my question is, is there a best practice I'm missing here? Is there a way to pull the parameters a given Effect needs from that Effect? Thank you all for your time!

    Read the article

< Previous Page | 80 81 82 83 84 85 86 87 88 89 90 91  | Next Page >