Search Results

Search found 79415 results on 3177 pages for 'log file'.

Page 299/3177 | < Previous Page | 295 296 297 298 299 300 301 302 303 304 305 306  | Next Page >

  • Dir and Findstr commands taking a long time to complete in Batch File

    - by user2405934
    dir %DRIVE_NAME%: /S /C /A-D /Q /T:C | findstr ".zip$ .doc$ .xls$ .xpt$ .cpt$ .cpo$ .xlsx$ .pdf$ .dat$ .txt$ .docx$ .csv$" >> file.info I am using above command to list all information in file, as below: 03/27/2013 01:02 PM 86,280 uusr\fr02 h123_frf67_rk_20140327.txt 03/27/2013 01:02 PM 5,513 usr\fr02 h123_frf67_rk_20140328.txt %DRIVE_NAME%: is mapped drive. Folders will be the same; not more than 100 folders and their sub-folders, and there will only be 2 or 3 files at time in any one of the folders. Now the issues is that for one folder it works perfect, but for 80 to 90 folders it is taking too much time. I think it's because of findstr and the different extensions used. Is there any way to make it faster?

    Read the article

  • Double try_files to solve the nginx's "No input file specified" issue

    - by Howard
    I am following the nginx's wiki (http://wiki.nginx.org/WordPress) to setup my wordpress location / { try_files $uri $uri/ /index.php?$args; } By using the above lines, when a static file which is not found it will redirect to index.php of wordpress, that is okay but.. Problem: When I request an non-existence php script, e.g. http://www.example.com/foo.php, nginx will give me No input file specified I want nginx to return 404 instead of the above message, so in the main fcgi config, I add the 2nd try_files location ~ \.php$ { try_files $uri =404; fastcgi_split_path_info ^(.+\.php)(/.+)$; include /etc/nginx/fastcgi_params; ... } And this worked, but I am looking if there are any better way to handle it?

    Read the article

  • Why are my log in times taking so long in Linux?

    - by Jamie
    In recent weeks, login times on my Ubuntu server have started timing out; both through SSH and the local command line console. Examination of the /var/auth.log yields nothing interesting. How can I diagnose long log in times on my Ubuntu server? I should mention, also, that no updates have been performed since the problem has started, and that the /, /boot/ and /usr/ file systems are mounted as readonly. [Edit] This is a stand alone machine, so it doesn't authenticate with Active Directory, LDAP etc. Also, the login prompt is responsive, as is the password prompt. Upon typing the password then CR, I'll timeout. After four a five tries, I will be able to login, although I'm worried this will start taking longer.

    Read the article

  • Is there a simpler way to log into a local account on a domain workstation?

    - by David
    Occasionally I need to log into a local machine account on a Windows workstation joined to our domain. The syntax for specifying a domain account looks like this: DOMAINNAME\myusername whereas the syntax for logging into a local account is HOSTNAME\myusername The problem is that I often don't know the host name off the top of my head. It is possible to find out by clicking the "How do I log onto another domain" link, but this requires me to memorize or write down an often cryptic hostname. Is there another, simpler way to do this?

    Read the article

  • Context Menu for Browser to download file to specific folder

    - by elcojon
    There is this website which has customized audio files for me. I would like to save them in a special folder. Now I don't want to select the "special folder" each time in the file-chooser dialogue of my browser. I would rather prefer to have a custom entry in the context menu when I right-click the download link. This context-menu-entry should do the trick and download the file to the predefined "special folder". How would I start about that? I am using Safari and Chrome. So a solution in either browser is fine. To get into the context menu of the browser, what kind of programming do I need to do? Is it an extension, plugin, etc.? Thanks

    Read the article

  • User account automatically filling up with dead.letter file

    - by jeroen
    I have one user account on a server with about 400 accounts that is filling up automatically. The dead.letter file in the users home directory automatically grows until the account is full (about 10 - 40 Mb per day). The user is using Microsoft Outlook to send and receive mail. What can be causing this and how can I avoid it from happening? Right now I have an emergency cron-job to delete the file but I would like "real" solution. Edit: The server version is Red Hat Enterprise Linux ES release 4 (Nahant Update 4) Edit 2: It seems mainly spam and I see different mailer headings (from php to Outlook Express) and a frequent appearing header is [email protected] Update: I have asked the hosting provider where I use that dedicated server to look into the problem as well, as it's their Control Panel that could be a cause of the problem.

    Read the article

  • How can I use the proxy settings on Epic privacy browser to log on to Facebook?

    - by EddieN120
    I love the Epic privacy browser because it is built from the ground up to enhance privacy. It's built on Chromium but because it has stripped out all code that tracks users across the Internet, pages load faster and things work snappier. With one click you can enable a proxy to hide your IP address, sort of like Chrome's "Incognito" mode on steroids. But there's a problem: if I load Epic, go to facebook.com, log in, and then click the proxy button, I can use Facebook for a while. But eventually, Facebook would throw up an error screen, saying that it thinks that my account has been hacked, and then it would make me verify my identity, force me to change my password, etc. I've had to change my password four times in as many days, which is very annoying. Now I turn on the proxy for browsing on to every other site but Facebook. Question: how can I use the proxy settings on Epic privacy browser to successfully log onto and use Facebook?

    Read the article

  • Clamdscan scans file in 0 seconds

    - by SupaCoco
    I have to run clamav on large files. I was wondering which command was the fastest between clamscan and clamdscan. But it seems that clamdscan is not working properly: it scans file larger than 1 GB. Could you guys help me find why the heck clamdscan isn't working ? Between clamscan and clamdscan which one is less resource consuming ? I run ClamAV 0.97.8/18037 on Ubuntu 12.04.3 LTS. Please find below the execution result of both commands: clamscan myfile.zip ----------- SCAN SUMMARY ----------- Known viruses: 2864504 Engine version: 0.97.8 Scanned directories: 0 Scanned files: 1 Infected files: 0 Data scanned: 0.00 MB Data read: 1024.16 MB (ratio 0.00:1) Time: 9.145 sec (0 m 9 s) clamdscan myfile.zip /home/ubuntu/workspace/benchmark/myfile.zip: OK ----------- SCAN SUMMARY ----------- Infected files: 0 Time: 0.000 sec (0 m 0 s) And here are the clamav log file: Wed Oct 30 10:26:32 2013 -> Received POLLIN|POLLHUP on fd 4 Wed Oct 30 10:26:32 2013 -> Got new connection, FD 9 Wed Oct 30 10:26:32 2013 -> Received POLLIN|POLLHUP on fd 5 Wed Oct 30 10:26:32 2013 -> fds_poll_recv: timeout after 5 seconds Wed Oct 30 10:26:32 2013 -> Received POLLIN|POLLHUP on fd 9 Wed Oct 30 10:26:32 2013 -> got command CONTSCAN /home/ubuntu/workspace/benchmark/myfile.zip (51, 7), argument: /home/ubuntu/workspace/benchmark/myfile.zip Wed Oct 30 10:26:32 2013 -> mode -> MODE_WAITREPLY Wed Oct 30 10:26:32 2013 -> Breaking command loop, mode is no longer MODE_COMMAND Wed Oct 30 10:26:32 2013 -> Consumed entire command Wed Oct 30 10:26:32 2013 -> Number of file descriptors polled: 1 fds Wed Oct 30 10:26:32 2013 -> fds_poll_recv: timeout after 3600 seconds Wed Oct 30 10:26:32 2013 -> THRMGR: queue (single) crossed low threshold -> signaling Wed Oct 30 10:26:32 2013 -> THRMGR: queue (bulk) crossed low threshold -> signaling Wed Oct 30 10:26:32 2013 -> /home/ubuntu/workspace/benchmark/myfile.zip: OK Wed Oct 30 10:26:32 2013 -> Finished scanthread Wed Oct 30 10:26:32 2013 -> Scanthread: connection shut down (FD 9) Wed Oct 30 10:26:32 2013 -> THRMGR: queue (single) crossed low threshold -> signaling Wed Oct 30 10:26:32 2013 -> THRMGR: queue (bulk) crossed low threshold -> signaling

    Read the article

  • Reclaiming deleted disk space from file vault

    - by cbrulak
    I have my main user account encrypted with file vault. After deleting some data (like 20 GB) my free space on the hard drive hasn't change (yes I emptied the trash, confirmed that the files are actually gone, etc,etc). I also tried "erasing free space" in the disk utility app. I logged off, and rebooted and so far that space hasn't been reclaimed. I'm assuming file vault or disk utility has some method of reclaiming but I can't find it. Any ideas?

    Read the article

  • Findstr not finding file when searching by extension

    - by feiht thief
    I am experiencing some odd behaviour when using findstr to search a collection of .php files. Depending on how I specify the file list, the file is not being found, even though it contains the string in question. These two files contain the string "personemail": Content.php and People.php A) Works as expected (finds personemail in Content.php and People.php) findstr /i /s /c:"personemail" * B) Fails (finds only in Content.php) - expected to find in Content.php and People.php findstr /i /s /c:"personemail" *.php C) Works as expected (finds in People.php): findstr /i /s /c:"personemail" p*.php What is going on?

    Read the article

  • Enable/disable wireless interface in a bat file

    - by Herms
    Is there a way to enable/disable a wireless network interface via the command line (so I can put it in a bat file)? When I'm in the office I use a wired connection, but Windows still occasionally bugs me about my wireless connection not being connected (even though I turned off the option for notifying me when there's no connectivity). I'm guessing the only way to stop it from bugging me is to disable the interface, but I'd rather not have to go into the network settings every time I need to do so. I'd like to set up a bat file or a shortcut that I can use to enable/disable the wireless (preferably a single one that toggles the current state), and then just set up a keyboard shortcut for that. I just have no idea how to do so from the command line.

    Read the article

  • Extracting httpdocs from Plesk Panel 9.5.4 Webserver backup file

    - by Paddington
    Good day, I am having problems manually extracting domains from Plesk 9.5 backup that was FTPed onto my back up server. I have followed the article http://kb.parallels.com/en/1757 using method 2. The problem is here: zcat DUMP_FILE.gz DUMP_FILE My backup file CP_1204131759.tar is a tar archive and zcat does not work with it. So I proceed to run the command: cat CP_1204131759.tar CP_1204131759. But when I try # cat CP_1204131759 | munpack I get an error that munpack did not find anything to read from standard input. I went on to extract the tar backup file using the xvf flags and got a lot of files (20) similar to these ones: CP_sapp-distrib.7686-0_1204131759.tgz CP_sapp-distrib.7686-35_1204131759.tgz CP_sapp-distrib.7686-6_1204131759.tgz How best can I extract the httpdocs of a domain from this server wide Plesk 9.5.4 backup?

    Read the article

  • I get a 403 when requesting a JS file from CloudFront

    - by Roland
    This is new to me so please excuse me if I have no idea what I'm talking about (: I'm trying to set up my own CDN with CloudFront and S3 through a subdomain by adding a CNAME to that subdomain to point to the CloudFront. It seems like I get a 403 when trying to load the file, this is the original s3 link : https://s3.amazonaws.com/chaoscod3r_aws_cdn/libs/polyfills/json3_polyfill.js ; which seems to be working after setting the permission to everyone to open / download. But when trying to use the subdomain to request the file : http://cdn.chaoscod3r.com/libs/polyfills/json3_polyfill.js ; it seems like I get that 403. Could anyone help me out with this one ?

    Read the article

  • Realtime file-level mirroring from local NTFS to network drive

    - by hurfdurf
    We have some data collection machines running WinXP. After a new file is written, we would like to immediately copy the new file to network storage (a NetApp CIFS share) automagically. We need realtime or near realtime copies generated (copy upon filehandle close would be fine -- these are not long-running system logs). Two commercial applications I've found so far are MirrorFile and IBM's Tivoli CDP. Are there any reliable open source programs or simple ways to get Shadow Copy to do something similar? Bonus points if it runs as a service.

    Read the article

  • Not able to output to file in the Windows command line

    - by Sachin
    In the following code, I need to take the path and size of folder and subfolders into a file. But when the loop runs for the second time, path and size are not getting printed to the file. size.txt only contains the path and size of the 1st folder. Please somebody help me. @echo off SETLOCAL EnableDelayedExpansion SET xsummary= SET xsize= for /f "tokens=1,2 delims=C" %%i IN ('"dir /s /-c /a | find "Directory""') do (echo C%%j >> abcd.txt) for /f "tokens=*" %%q IN (abcd.txt) do ( cd "%%q" For /F "tokens=*" %%h IN ('"dir /s /-c /a | find "bytes" | find /v "free""') do Set xsummary=%%h For /f "tokens=1,2 delims=)" %%a in ("!xsummary!") do set xsize=%%b Set xsize=!xsize:bytes=! Set xsize=!xsize: =! echo.%%q >> size.txt echo.!xsize! >> size.txt )

    Read the article

  • File corruption

    - by firesoul
    Recently, almost every file I try to copy into my HDDs gets corrupted. When writing to the disk, some portion of the file is replaced by random bytes (Checked using Diff apps) There's no problem reading existing files. I'm pretty sure it's not the HDD, becasue I've tried another HDD (a brand new 1TB WD) and the problem still remains. I have an AMD Athlon 64 X2 CPU, an ASUS M2NE motherboard, 2 WD SATA II HDDs (320GB + 1TB) and 3GB of Corsair RAM. My OS is Windows 7 (64 bit) Now, what could be the cause of this? If it's the motherboard, can it be fixed by using a PCI SATA card?

    Read the article

  • Running Coldfusion Commands inside an HTML File?

    - by KM01
    Hello, Is it possible to run CF commands inside an HTML file by updating Apache configs/htaccess file? When I searched online, I didn't come across any answers for CF, saw several posting for PHP (perhaps I am not searching for the right terms ... ?). Specifically, we have CF 6 (I know, don't say it) on Solaris, and Apache 2, and was thinking of using this line: AddType application/x-httpd-coldfusion .html so we can use cfinclude inside the HTML files, but this did not work. Your time, ideas and thoughts are much appreciated! KM

    Read the article

  • Windows 7 file explorer preview window and password protected word documents

    - by Carbonara
    When using the Windows 7 Explorer with the preview pane open you get a little preview of a file when you click on it. This includes Word, Excel spreadsheets, etc. My problem is if the Word document is password protected. Clicking on it in Explorer automatically asks for the password to display its preview. It does this if you single or double clicking on it. You then get an empty Word instance running (which allows it to display the preview) and another instance of Word with your actual file and you're asked for the password twice in total. This is annoying and untidy. Is there a way of stopping the preview pane from wanting to display password protected documents and thus not asking for the password to display a preview?

    Read the article

  • How to copy windows explorer file selections and past filenames as txt

    - by Ricky Williams
    Win7 How can I select multiple files in explorer and get a string of text listing the file names like i would get if i selected the files from within "open file" window of an application. for example if i have a Dir that contains 100 jpegs and i select 01.jpg, 05.jpg, 10.jpg in windows explore via ctrl+clicks or shift+click how can i get a text string that reads " "01.jpg" "05.jpg" "10.jpg" with or without the selected files full path " with out dragging the selected files to another window? I've tried copy and pasting into Note Pad and Word Pad and it either past the picts into the the application or a empty clipboard.

    Read the article

  • Cygwin file and directory user and group

    - by dvanaria
    I use Cygwin as my main development environment on both my home and work computers. In order to share files between the two computers, I use Dropbox, which is installed in the following folder on both computers: c:\cygwin\home\dvanaria\dropbox Everything works great, except for one thing. When I'm working on my home computer and do an ls -l on any directory, all the files show up as owned by dvanaria of group Users. But when I work from my work computer, an ls -l shows all files as being owned by Administrators and of group Domain Users. I know Cygwin uses some kind of mapping between Windows users and permissions to the /etc/passwd file. But to be honest I have no idea how this file works or how it maps to Windows under Cygwin. Could anyone help figure this out? The main problem is that I can't edit any files when using my work computer, only read them.

    Read the article

  • SBS 2008 File access restrictions after 5:30PM

    - by Moif Murphy
    Hi there, SBS 2008, I have an issue whereby a user can't access a file after 5:30PM. They get the message: "cannot be accessed. The file may be corrupted, located on a server that is not responding, or read only I've checked the logon hours on the AD User account in question and there's no restrictions there. I have noticed that the account in question does not appear in the SBS Console, only in ADUC which means the account wasn't added using the wizard. It's a long shot but would that make a difference? Other than that I'm pretty stumped, any suggestions welcomed. Thanks

    Read the article

  • Copy file to WebDAV via Command Line on Windows 2003

    - by Boden
    I need to copy a file from a Windows 2003 server to a WebDAV folder (on the same server, if it matters). This operation will be performed via a batch script executed via Scheduled Tasks. I've enabled the WebClient service on the server. So far I've determined that I can do it like this: net use x: http://host/path copy c:\path\myfile.txt x: net use x: /delete 1) Is there a simpler way than creating a temporary mapped drive? Will it work via a batch file when no user is logged in? 2) Is there anything I should know about enabling the WebClient service on my server? Previously it was disabled, which I assume is default.

    Read the article

< Previous Page | 295 296 297 298 299 300 301 302 303 304 305 306  | Next Page >