Search Results

Search found 74849 results on 2994 pages for 'file folder'.

Page 232/2994 | < Previous Page | 228 229 230 231 232 233 234 235 236 237 238 239  | Next Page >

  • Adding a file path into a formula that is typed into a another cell

    - by Adam Graham
    I have 'C:\Users\Documents...etc.......[file name.xlsx]Work Sheet'!$B:$F in cell B1 i then want to run a vlookup formula to the above file but instead of vlookup(A1,'C:\Users\Documents...etc.......[file name.xlsx]Work Sheet'!$B:$F,2,false) I want to use the cell B1 for the path. Reason is i want the master to look at multiple workbooks and i don't have time to sit and retype. Please help

    Read the article

  • how to remove background layer of djvu file

    - by Jon
    Hello, I've downloaded some files from the Internet Archive. They come in different file formats and most of the time I use pdf. However, sometimes the scans are saves in colour instead of b/w. This makes it difficult/impossible to read on a dedicated ebook reader. In that case I downloaded the djvu files as on the PC you can select which layer (color, bw,fore,back) one would like to see. Selecting the bw gives excellent results. However, the ebook reader does not has this option. The question is, how can I remove /extract a layer from the djvu file and save only this layer. So far I've tried the following two approaches: 1) select bw in the djvu viewer on the PC and printed to postscript file. Followed by a ps2pdf conversion. This works, but generates a fairly large pdf file. Sure, I can again upload it to any2djvu but it just seems to much manual work for each file. 2) I tried the shared annotation feature and said (mode bw). This works on the PC as desired but is ignored on the ebook reader as the other layers are still present. Any help or suggestions would be greatly appreciated.

    Read the article

  • Determine the percentage of a file that has been ftp'd from client to server

    - by klwillie
    I want to ftp a large file from a Windows client to a Windows server, using their IP addresses. This is on an internet independent network. While the file is transferring, I would like to determine how many bytes have been received by the server. I then would like to use this information to determine in real-time the percentage of the file that has been transferred to the server. Any recommendations as to the ftp command syntax and C# code to achieve this?

    Read the article

  • Odd IIS v6 .html file caching behavior: Ideas?

    - by samsmith
    Our loadbalancer looks for "loadbalancer.html" So today, I deleted the file, but in a browser I could still fetch it. My eyes crossed, I made sure, 3 ways, that the file was removed from the iis web site home dir. But fetching it gave a nice 200 http response, and the file. In desperation, I created a new loadbalancer.html, and put some test text in it, and then this new content was retrieved, and then I deleted this file, and THEN it finally gave the 404 as desired. How can I better understand/control this ? Thanks!

    Read the article

  • "Send to: Mail recipient" context menu not working

    - by Louis
    I have set the Mail app from Microsoft (WinRT) as the default application for mail in Default Programs. However, when I right-click on a file, and select Send to Mail recipient I get a message box saying:There is no email program associated to perform the requested action. Please install an email program or, if one is already installed, create an association in the Default Programs control panel. Is there a way to get this action working with Mail?

    Read the article

  • How to pause an m4b file

    - by Phenom
    When I play m4b files on my computer, they open with iTunes. I can stop the file, but I cannot resume the file from within iTunes. In order to pick up where I left off, I have to open the file again. How can I resume where I left off from within iTunes? Is there another program that will play m4b files and resume from where you left off?

    Read the article

  • windows printer spool gets stuck with file at 64kb from linux and mac

    - by Juan Diego
    Hi I have two printers one on a file server with windows 2003, and other with windows XP. The thing is that when i try to print from my machine, my file stays in queue for ever, it says 64kb out of whatever the file i send. I have seen similar problems with some machines that run on Mac OS X. The windows machines apparently have no problems printing. They are not connect through active directory, just the network. In the past I have seen people install non microsoft windows Printer server on windows, i dont remember the name of any of the programs. I have being googling a lot and have not found anything to replace the microsoft print spooler service, maybe i am mistaken. Everyday I have to restart the print spooler service i even created a bat file for it. I am out of ideas here.

    Read the article

  • Read ahead buffering while playing video file from optical disc

    - by Saxtus
    I was wondering if there is a program for Windows 7 (64-bit), that reads ahead the file to be played back (usually MKV files in my case) to the RAM, so the disc that the video file resides in, won't spin for the entire duration of the playback, but only every time the cache is almost empty (with big enough cache so it won't need the drive for long periods). A program that I've used in the past (called DVDIdle), was doing that universally for every video player I wanted, but they've stopped updating it 6 years ago and now it doesn't seem to work with Windows 7 (tried using compatibility mode too). The method I am using now, is to either have the drive wearing down and buzzing all the time or manually copy the entire file to HDD, SSD or RamDisk and play it from there. The closest thing I've found, is a software that slows down the drive's spin speed, but I was looking for something more convenient, automated, without waiting for an entire file to be copied before starting playback or needing any HDD space, like I've used to in the past. Thanks in advance.

    Read the article

  • How to convert an Outlook 2007 .ost file to a .pst file, so I can open it in another, local computer?

    - by ysap
    I have a couple of backed up .pst and .ost files that were used with a server that is not accessible anymore. On my local machine, I was able to Add... the .pst file to my fresh Outlook 2007 and access the emails. However, adding an .ost file does not seem to be an options in the Accounts Settings dialog. Is it possible to import the .ost file, or convert it to a .pst w/o the aid of the exchange server?

    Read the article

  • Sfsync04 file is corrupted

    - by webdreamer
    I don't know exactly what I did, but somehow the sfsync04.sys (drivers) file is corrupt, and now I can't start my windows. That's the error startup repair gives me, and it is unable to correct it automatically. I'm currently running a live version of Ubuntu, so I can access my files, and everything seems to be alright except for the sfsync04.sys file. Is there a way I can repair that file so I can avoid reinstalling Windows for now? I'm running windows 7 64 bits. The file is in windows\system32\drivers\ . I think this happened when I tried to install some legacy software, but I'm not sure.

    Read the article

  • xf86OpenConsole: Cannot open /dev/tty0 (No such file or directory)

    - by mazgalici
    root@mazgalici:~# startx X.Org X Server 1.7.6 Release Date: 2010-03-17 X Protocol Version 11, Revision 0 Build Operating System: Linux 2.6.24-28-server i686 Ubuntu Current Operating System: Linux mazgalici 2.6.18-194.26.1.el5.028stab079.2PAE #1 SMP Fri Dec 17 19:34:22 MSK 2010 i686 Kernel command line: quiet Build Date: 10 November 2010 11:25:26AM xorg-server 2:1.7.6-2ubuntu7.4 (For technical support please see ) Current version of pixman: 0.16.4 Before reporting problems, check to make sure that you have the latest version. Markers: (--) probed, (**) from config file, (==) default setting, (++) from command line, (!!) notice, (II) informational, (WW) warning, (EE) error, (NI) not implemented, (??) unknown. (==) Log file: "/var/log/Xorg.0.log", Time: Tue Jan 11 01:28:48 2011 (==) Using config directory: "/usr/lib/X11/xorg.conf.d" Fatal server error: xf86OpenConsole: Cannot open /dev/tty0 (No such file or directory) Please consult the The X.Org Foundation support at http://wiki.x.org for help. Please also check the log file at "/var/log/Xorg.0.log" for additional information. ddxSigGiveUp: Closing log

    Read the article

  • rsync : Read input from a file and sync accordingly

    - by Dheeraj
    I have a text file which contains the list of files and directories that I want to copy (one on a line). Now I want rsync to take this input from my text file and sync it to the destination that I provide. I've tried playing around with "--include-from=FILE" and "--file-from=FILE" options of rsync but is is just not working I also tried pre-fixing "+" on each line in my file but still it is not working. I have tried coming with various filter PATTERNs as outlined in the rsync man page but it is not working. Could someone provide me correct syntax for this use case. I've tried above things on Fedora 15, RHEL 6.2 and Ubuntu 10.04 and none worked. So i am definitely missing something. Many thanks.

    Read the article

  • does any sync software find files that have moved paths within a folder?

    - by kpierce8
    Say I have a pictures folder which I reorganized on one computer. I'd like to use that directory as the base and compare it with another version on a backup drive. Will any sync/compare program find that a file in one folder has moved locations within a compare folder. For instance say I reorganized my pictures from trips into folders by year with the trips folders inside each year folder. If I use a regular compare utility I wind up with two copies of everything that's moved in different locations.

    Read the article

  • How to audit a specific folder in Windows Server 2003?

    - by saint
    We have had several cases of file deletion in one of our windows file servers. Server is also a domain controller. Unfortunately we kept the auditing disabled completely due it being such a resource hog. So I was wondering If there is a way to audit a specific folder for a specific security group, for just file and folder deletions within the specified folder. Also is there a recommended third party application for auditing or monitoring. Many Thanks.

    Read the article

  • Xcopy batch file behavior

    - by piagetblix
    I have the following backup batch file: xcopy C:\test.txt \server\Data\user\"data test" /e /y /i The issue is why does it not only copy the file listed but proceeds to copy the whole root directory? When i remove the switches it only copies the file. There is nothing in the description of those switches that seems to cause the behavior... They are there because I'm picking up where someone else left and all the batch files use these 3 switches. Thanks

    Read the article

  • Opening a non-currupt PST file

    - by Kyle Brandt
    I have a a PST file that a user is unable to open. The permissions look fine to me, and I had to repair the file, but now the PST file is repaired and I can open it from my Outlook (In other words, seems to be a local issue). Any recommendations on what to do to see why the user can't open it with their outlook?

    Read the article

  • nginx: js file loads indifferently every refresh

    - by poymode
    I have this nginx problem wherein a js file in a rails app loads indifferently. Whenever I try to access the JS file in the browser and refresh the page, the scrollbar changes length meaning sometimes it loads half the js page, sometimes the whole and sometimes just a part of it. the js file size is 71K. my nginx server is on different server,separate from my rails app. when I try to access the js file directly through the app server, lets say 10.48.30.150:3000/javascripts/file.js it works fine and doesnt show any half-loaded page. but when I use the nginx server which upstreams the rails app, it shows the indifferent page loads. here is my nginx http conf error_log /usr/local/nginx/logs/error.log; pid /usr/local/nginx/logs/nginx.pid; events { worker_connections 1024; } http { include mime.types; default_type application/octet-stream; server_names_hash_bucket_size 256; access_log /usr/local/nginx/logs/access.log; sendfile on; #tcp_nopush on; keepalive_timeout 0; tcp_nodelay on; #gzip on; #gzip_min_length 4096; #gzip_buffers 16 8k; #gzip_types application/x-javascript text/css text/plain; large_client_header_buffers 4 8k; client_max_body_size 2G; include /usr/local/nginx/conf.d/*.conf; }

    Read the article

  • Write once, read many (WORM) using Linux file system

    - by phil_ayres
    I have a requirement to write files to a Linux file system that can not be subsequently overwritten, appended to, updated in any way, or deleted. Not by a sudo-er, root, or anybody. I am attempting to meet the requirements of the financial services regulations for recordkeeping, FINRA 17A-4, which basically requires that electronic documents are written to WORM (write once, read many) devices. I would very much like to avoid having to use DVDs or expensive EMC Centera devices. Is there a Linux file system, or can SELinux support the requirement for files to be made complete immutable immediately (or at least soon) after write? Or is anybody aware of a way I could enforce this on an existing file system using Linux permissions, etc? I understand that I can set readonly permissions, and the immutable attribute. But of course I expect that a root user would be able to unset those. I considered storing data to small volumes that are unmounted and then remounted read-only, but then I think that root could still unmount and remount as writable again. I'm looking for any smart ideas, and worst case scenario I'm willing to do a little coding to 'enhance' an existing file system to provide this. Assuming there is a file system that is a good starting point. And put in place a carefully configured Linux server to act as this type of network storage device, doing nothing else. After all of that, encryption on the files would be useful too!

    Read the article

  • pcfg_openfile: unable to check htaccess file, ensure it is readable

    - by rxt
    After moving a website folder on my local development machine to another drive, then moving it back, I got a 403 error. Most of this problem had probably to do with rights that got messed up. After deleting the code and restoring it from SVN, the rights seemed allright. The error stayed however. The setup is a bit complex, as follows: I have Ubuntu 10.4 as development machine, trying to mimic the server as much as possible We use Eclipse + SVN and I create all projects in a local folder under my user account In /var/www-vhosts I create folders for each vhost, like this one: test.localhost test.local/index.php: includes the index file of the project test.local/.htaccess is a dynamic link to the htaccess file in a project subfolder I get the following error in the apache error log: [Thu Jul 08 15:55:56 2010] [crit] [client 127.0.0.1] (13)Permission denied: /var/www-vhosts/test.localhost/.htaccess pcfg_openfile: unable to check htaccess file, ensure it is readable The problem seems to be the .htaccess file, or the link to it. When I empty the htaccess, nothing changes When I remove the link, the index-include produces some output (in the apache error log) When I remove the link and replace it with the actual file, I get another error: [Thu Jul 08 16:47:54 2010] [error] [client 127.0.0.1] Symbolic link not allowed or link target not accessible: /var/www-vhosts/test.localhost/test I'm lost here, don't know what to do next. Do you have any ideas what I can try? This setup has worked before, but I don't know what is different now.

    Read the article

  • Upload large database SQL file

    - by Devy
    I've a database of more than 20Gb of size on my hard disk. What is the best way to upload it with the least (money) load possible on the server? - I'm on Windows 7. - I have FTP and SSH access on the server. I avoid using FTP because my connection cuts off a lot, I can't imagine I re-upload again the file after failing on 99%. I found some tools that split the large .sql file to small .sql files, but they didn't mention how to gather these files again into one file. Another way is to archive the big .sql file to .rar with -v option, upload them through FTP then unpack them. But unpacking will also cost, right? I know it will cost in any cases, but any best practice will be strongly appreciated.

    Read the article

< Previous Page | 228 229 230 231 232 233 234 235 236 237 238 239  | Next Page >