Search Results

Search found 50510 results on 2021 pages for 'static files'.

Page 101/2021 | < Previous Page | 97 98 99 100 101 102 103 104 105 106 107 108  | Next Page >

  • copy files to Nexus 4 on Kubuntu (13.04)

    - by cerr
    I know, this has been asked a few times already but I'm still having troubles to get files copied my new Nexus 4 phone mounted on my Kubuntu 13.04 machine. I followed the guide at: http://www.pocketables.com/2013/03/how-to-mount-the-lg-nexus-4-as-a-usb-drive-in-linux.html I would only get: $ cp -r /mnt/media/Albums/A\ Perfect\ Circle\ -\ Thirteenth\ Step/ . cp: accessing ‘.’: Transport endpoint is not connected reg@regDesktopHome:/media/nexus4/Internal storage/Music$ ls -l but thereafter , mount still shows it mounted: mtpfs on /media/nexus4 type fuse.mtpfs (rw,nosuid,nodev,allow_other) What's going on here? Thanks for the assistance!

    Read the article

  • Sharing Files between Ubuntu 14.04 and Windows 8

    - by Matinn
    I have Ubuntu and Windows 8 installed on one System. I am trying to share files between these two operating systems using an NTFS Partition wich was created by Windows. I don't have trouble accessing the data on this partition from Ubuntu, however if i create a file in Ubuntu, this file doesn't show up when I boot into Windows. Does anyone know how to do this. From what I have read file sharing should work without installing any additional Software, as I am not trying to access the Linux ext4 Partition from Windows.

    Read the article

  • Add rules (filters) to ftp programs to avoid uploading certain files/folders

    - by guisasso
    i use Filezilla as my ftp client, but this question goes to any other client that could be useful. Can i (in any client) add rules (filters) to an ftp program to avoid upload of certain files or folders? For example: Expression web creates those annoying _vti_cnf folders, or, certain folders in which i have the original version of a picture without a watermark that i don't want to upload. Example, i have a folder A, that has sub folders "original" and "current", i would like to add a filter, so every time i select A to be uploaded, "original" wouldn't go, but "current would".

    Read the article

  • script shortcut to open two files in gedit as sudo

    - by Sam
    I want to double click a file on my desktop, and have two files be opened in gedit as sudo. Whenever I'm making a new website, I need to open /etc/hosts /etc/apache2/sites-enabled/000-default.conf as sudo. At the moment this means opening the terminal, runing sudo gedit then opening each file manually. I want to streamline this part of my workflow. On windows I had wrote a little bash script which worked nicely. How can I do the same in ubuntu? So far in my searches I've come across ways of adding a shortcut to the file browser, and similar things, but not exactly what I want. I have tried creating a desktop launcher, but can't see how to make it run as sudo.

    Read the article

  • Routing using Linux with 2 NIC cards

    - by Kevin Parker
    Configured Clear OS to be in Gateway mode on a machine with two NIC cards. eth0:192.168.2.0/24 with ip 192.168.2.27 which is connected to a modem and thus have internet connectivity. eth1:192.168.122.0/24 with ip 192.168.122.10 which is connected to other machines in LAN through switch. LAN machines with network 192.168.122.0 is not getting internet.How can they get internet Through Clear OS gateway.I have enabled packet forwarding in clear os using "ip_forward=1" What am i missing?.Can you please help me in this. Following are the static routing i have added: on LAN machine1 with ip address 192.168.122.11 ip route add 192.168.2.0/24 via 192.168.122.10 dev eth0 ip route show 192.168.2.0/24 via 192.168.122.10 dev eth0 192.168.122.0/24 dev eth0 proto kernel scope link src 192.168.122.11 But still 192.168.2.0/24 network is not reachable.Where can be the problem??

    Read the article

  • Converting large files in python

    - by Cenoc
    I have a few files that are ~64GB in size that I think I would like to convert to hdf5 format. I was wondering what the best approach for doing so would be? Reading line-by-line seems to take more than 4 hours, so I was thinking of using multiprocessing in sequence, but was hoping for some direction on what would be the most efficient way without resorting to hadoop. Any help would be very much appreciated. (and thank you in advance) EDIT: Right now I'm just doing a for line in fd: approach. After that right now I just check to make sure I'm picking out the right sort of data, which is very short; I'm not writing anywhere, and it's taking around 4 hours to complete with that. I can't read blocks of data because the blocks in this weird file format I'm reading are not standard, it switches between three different sizes... and you can only tell which by reading the first few characters of the block.

    Read the article

  • How to package static content outside of web application?

    - by chinto
    Our web application has static content packaged as part of WAR. We have been planning to move it out of the project and host it directly on Apache to achieve the following objectives. It's getting too big and bloating the EAR size resulting in slower deployment across nodes. Faster deployment times. Take the load of Application Server Host the static content on a sub domain allowing some browsers (IE) to load resources simultaneously Give us an option to use further caching such as Apache mod_cache apart from the cache headers we send out to browsers. We use yuicompressor-maven-plugin to aggregate and minimize JS file. My question is how do package and manage this static content out side of the web application? My current options are. New maven war project. Still use the same plugin for aggregation and compression. Just a plain directory in SVN and use YUI/Google compressor directly. Or is there a better technology out there to manage static content as a project?

    Read the article

  • Material to use for computer system cover against UV and salty air?

    - by hippietrail
    I live right next to the sea and have a large window quite close to my computer setup which allows a lot of indirect sunlight to enter. I'd like to buy or make a cover for my computer system. From visiting my usual mom & pop computer shop yesterday I got the impression these might not really exist any more. If I make my own I need a material with these qualities: Block or reduce ultraviolet light which can depolymerize plastics (the sun here in Australia is much stronger than in the northern hemisphere). Block salt-laden sea air which can oxidize USB and other connectors. Not cause static electricity when covering or uncovering. Keep dust off of course (-: My setup is a laptop plugged into a wide-screen LCD with a few external drives. So I think I'd want a largish sheet to flop over the whole desk. Are such covers commonly sold these days? What material(s) should I look for which provides the listed attributes?

    Read the article

  • How to correctly set permisions for all subfolders and files

    - by Saeid87
    I am following the guide to install TinyOS on Ubuntu 12.04 I have done up to step 3, But I am not sure if I have done the step 3 correctly. Because by doing the step 4 I get the permission error : saeid@saeid-Satellite-C660:~$ tos-install-jni /usr/bin/tos-install-jni: 13: [: =: unexpected operator Installing 32-bit Java JNI code in /usr/lib/jvm/java-6-openjdk-i386/jre/lib/i386 ... install: cannot create regular file `/usr/lib/jvm/java-6-openjdk-i386/jre/lib/i386/libgetenv.so': Permission denied Can you please tell me what would be the actual commands for step 3? What I have to replace with following lines?: /opt/tinyos-2.x files: chown -R /opt/tinyos-2.x Change the permissions on any serial (/dev/ttyS), usb (/dev/tts/usb, /dev/ttyUSB), or parallel (/dev/parport) devices you are going to use: chmod 666 /dev/ I mean how would you do those steps in your ubuntu?

    Read the article

  • Syncing Files between workgroup server and Ubuntu workstation

    - by dotdawtdaught
    Recently I have decided that I can't make Windows 8 my primary OS on my laptop as it is just too cumbersome to deal with. I am made the switch to Ubuntu and so far so good. Using Windows I have been able to cache folders on my workgroup server using a feature called "Client Side Cache" that allows me to take a copy of my personal files offline while I am in the field, then later I when I return any changes get pushed up to the server and my local cache is refreshed. This feature is completely client driven although characteristics of it (who and what can be cached, and if caching is automatic) can be controlled via a policy assigned as part of a directory membership. Can anyone suggest a linux replacement for this feature? Is there a better way of handling this?

    Read the article

  • How To Remotely Copy Files Over SSH Without Entering Your Password

    - by YatriTrivedi
    SSH is a life-saver when you need to remotely manage a computer, but did you know you can also upload and download files, too? Using SSH keys, you can skip having to enter passwords and use this for scripts! This process works on Linux and Mac OS, provided that they’re properly configured for SSH access. If you’re using Windows, you can use Cygwin to get Linux-like functionality, and with a little tweaking, SSH will run as well.What is a Histogram, and How Can I Use it to Improve My Photos?How To Easily Access Your Home Network From Anywhere With DDNSHow To Recover After Your Email Password Is Compromised

    Read the article

  • OpenWorld Presentations and Anatomy of an RTF Template w/ files

    - by mdonohue
    For those who missed it ... or those who made it and couldn't get enough, check out the presentations delivered at OpenWorld: Overview and Roadmap The Reporting Platform for Oracle Applications Best Practices and even though it wasn't presented at OpenWorld an updated version of Anatomy of an RTF Template to include documented example files  (RTF template, Sub-Template and sample XML data) so you can re-use and play with the code directly.  Huge thanks to Tim and Hok-Min who did all the hard, original work on this example loaded with tips and tricks.  

    Read the article

  • Virtual box doesnt have files machine and help tab

    - by user54986
    this is the second time im asking this question because someone had to close my question for a reason which i still dont understand. this is the original question Why is my files,machine,help tab missing in ubuntu 11.10? Most importantly ,i cannot access the machine menu that should be there but isnt. im using ubuntu 11.10 with oracle 4.1.10 heres a picture of what ubuntu should look like,or maybe i am wrong, http://i.stack.imgur.com/JgM5S.png The reason why im asking,is because id like to run a usb key inside my virtual box,but since i cannot access machine,i cant access the devices manager either. Just for your information,ive already installed the guest additions.

    Read the article

  • VirtualBox - Install Windows 7 freeze on expading files

    - by spauny
    I'm using Kubuntu 12.10. I have the latest version of VirtualBox(4.2.4) with expansions installed and guest addition also. I'm trying to install Windows 7 but it freezes at the second step: expanding files (random percent). I even tried to install Vista but is happening the same thing. Is there a bug? Do you happen to know a workaround? Or at least there is another free virtual machine I can use to install win 7?

    Read the article

  • IIS to parse php in a .dll files

    - by Agony
    The .dll files ain't the dynamic link library. That's what the client side software calls for (cannot change). Its essentially a php script that should run and return specific values. However currently it simply downloads it and that results in a failure. That's what it results in on a Apache server: [Update] NewVersion=1 UpdateFileNumber=1 UpdateFile1=update1/LPServerInfo.dat ServerNumber=1 Server1=http://88.159.116.217/ here it is on IIS: 198.24.133.74:8080/update.dll?0 renaming it to php works fine for testing - it runs and returns values. I edited the MIME and set .dll to application/x-httpd-php but that doesn't seem to work in IIS. Any solutions?

    Read the article

  • Search for odt files without indexing

    - by josinalvo
    I am looking for: a way to search inside odt files (i.e. search for contents, not name) that does not require any kind of indexing that is graphical and very user-friendly (for a relatively old person, who does not like computers much) I know that it is possible to have 1) and 2): for x in `find . -iname '*odt'`; do odt2txt $x | grep Query; done works well enough, and it's pretty fast. But I wonder if there is already a good solution that does this with a GUI (or can be adapted to do this easily)

    Read the article

  • wine: x development files not found

    - by Subin
    I downloaded the wine-1.4-rc5.tar.bz2. I ran the tools/wineinstall in terminal. After some time I get this error. checking for X... no configure: error: X development files not found. Wine will be built without X support, which probably isn't what you want. You will need to install development packages of Xlib/Xfree86 at the very least. Use the --without-x option if you really want this. Configure failed, aborting install. I am using Ubuntu 10.04. The installation won't finish. Can I fix this and make wine install ?

    Read the article

  • Copying first 1000 PDF files having single, double quotes in their name to another folder

    - by racer_ace
    I am having this folder with PDFs into it and I need to process 1000 at a time. So I need to move them into another folder, process them and delete them. For this I tried using $ find . -maxdepth 1 -type f |head -1000|xargs cp -t $destdir It gives error on single and double quotes in filename. There are thousands of files and I have no idea how many of them has these quotes in them. Can anyone help me find a solution? And I tried with the -0 option, it did not work

    Read the article

  • Convert Microsoft Word documents (.doc/x) into HTML files

    - by danie7L T
    Does anybody knows of a good application to get it done quickly and efficiently ? I bought Word Cleaner but the results are merely sufficient and I need go over all the generated html files to clean tons of useless injected tags like <strongH</strong<strongell</strong<strongo </strong<emWor</em<emld</em Most of the articles displayed on a website I manage are based on documents written on MS Word by people how has little idea of what are margins for or ordered/unordered lists, foot/end notes etc and I cannot make them use something else. Does anyone has a tip to help me handle those pages more efficiently than going over them to correct and apply my CSS style ? NB: Just for the record, using "Save as HTML DOC" in Word is faaar worst than Word cleaner

    Read the article

  • Failure to download extra data files

    - by armanke13
    After fresh install 12.04 I && updating apt and system restart, I always get this annoying message after reboot : Failure to download extra data files The following packages requested additional data downloads after package installation, but the data could not be downloaded or could not be processed. ttf-mscorefonts-installer The download will be attempted again later, or you can try the download again now. Running this command requires an active Internet connection. But if I run attempt now, it shows flashing terminal window and like nothing happen. It'll happen again when I restart system. I found someone have this problem too, but he haven't replied yet. I'm a newbie here, please help, thanks ^^

    Read the article

  • Please help with samba, can't access files from other computer

    - by user254074
    Maybe I am missing something but all I want to do is be able to access and transfer files from one folder on my main computer to my htpc. Both computers are running Ubuntu 14.04. I followed the instructions on this video but I am not sure if I need to do something on the HTPC to access my Movie folder. I went into nautilus then Browse Network Windows Network Workgroup but I get an error and it won't let me access it. Is there something I am doing wrong?

    Read the article

  • Automatically delete files after they expire

    - by Auxiliary
    I've got this idea for some time and I was wondering if anyone has seen such a feature/app in any operating system and if you haven't, what do you think about it. Where do you think I should begin? The idea is simple. I think we all have those files that are made and probably used for a few days and then are left on our disk and we never delete them or even check to see if we need them again. It'd be cool if you could right click on a file and click on "Expire in.. 3 days" for example. And the file gets deleted after 3 days. I have a great need for this and maybe some people will find it useful. I was thinking of writing a script and use the Nautilus Action project in GNOME for a start.

    Read the article

  • Isolated Network Set-Up

    - by Isaac Kleinman
    I am looking to set up a small network for a client which would be isolated from his primary network. (I'm hoping to set-up the additional network as VMs on the primary network.) I've instructed the client to request a bunch of static ip addresses from his ISP, but I'm not sure how to proceed with setting this up. What hardware will I need and how do I go about the configuration? Heavy security is not my concern. All that's really required is that web requests from the two networks be presented to the outside world with different ip addresses.

    Read the article

  • Synchronizing (updating) published files

    - by MucMug
    How can I update a published file and maintain the same URL? After saving an update of a published file on my desktop, it will automatically "synchronise" with the corresponding files at UbuntuOne (and it does). Problem is that the "new file", actually the updated file with the same name, is no longer published. Pressing the publish button results is a new URL. I now have to mail new URL's and change embedded links, as old URL will result in a failure to find the updated file (or indeed any file). I am not sure if it is a bug or a design flaw (maybe intentional?), but it seems strange to me.

    Read the article

  • Linking LLVM JIT Code to Static LLVM Libraries?

    - by inflector
    I'm in the process of implementing a cross-platform (Mac OS X, Windows, and Linux) application which will do lots of CPU intensive analysis of financial data. The bulk of the analysis engine will be written in C++ for speed reasons, with a user-accessible scripting engine interfacing with the C++ testing engine. I want to write several scripting front-ends over time to emulate other popular software with existing large user bases. The first front will be a VisualBasic-like scripting language. I'm thinking that LLVM would be perfect for my needs. Performance is very important because of the sheer amount of data; it can take hours or days to run a single run of tests to get an answer. I believe that using LLVM will also allow me to use a single back-end solution while I implement different front-ends for different flavors of the scripting language over time. The testing engine itself will be separated from the interface and testing will even take place in a separate process with progress and results being reported to the testing management interface. Tests will consist of scripting code integrated with the testing engine code. In a previous implementation of a similar commercial testing system I wrote, I built a fast interpreter which easily interfaced with the testing library because it was written in C++ and linked directly to the testing engine library. Callbacks from scripting code to testing library objects involved translating between the formats with significant overhead. I'm imagining that with LLVM, I could implement the callbacks into C++ directly so that I could make the scripting code work almost as if it had been written in C++. Likewise, if all the code was compiled to LLVM byte-code format, it seems like the LLVM optimizers could optimize across the boundaries between the scripting language and the testing engine code that was written in C++. I don't want to have to compile the testing engine every time. Ideally, I'd like to JIT compile only the scripting code. For small tests, I'd skip some optimization passes, while for large tests, I'd perform full optimizations during the link. So is this possible? Can I precompile the testing engine to a .o object file or .a library file and then link in the scripting code using the JIT? Finally, ideally, I'd like to have the scripting code implement specific methods as subclasses for a specific C++ class. So the C++ testing engine would only see C++ objects while the JIT setup code compiled scripting code that implemented some of the methods for the objects. It seems that if I used the right name mangling algorithm it would be relatively easy to set up the LLVM generation for the scripting language to look like a C++ method call which could then be linked into the testing engine. Thus the linking stage would go in two directions, calls from the scripting language into the testing engine objects to retrieve pricing information and test state information and calls from the testing engine of methods of some particular C++ objects where the code was supplied not from C++ but from the scripting language. In summary: 1) Can I link in precompiled (either .bc, .o, or .a) files as part of the JIT compilation, code-generation process? 2) Can I link in code using that the process in 1) above in such a way that I am able to create code that acts as if it was all written in C++?

    Read the article

< Previous Page | 97 98 99 100 101 102 103 104 105 106 107 108  | Next Page >