Search Results

Search found 40479 results on 1620 pages for 'binary files'.

Page 498/1620 | < Previous Page | 494 495 496 497 498 499 500 501 502 503 504 505  | Next Page >

  • How can I do a partial update (i.e., get isolated changesets) from subversion with subclipse?

    - by Ingvald
    If a file is committed several times with various changes, how can I fetch one change at a time, i.e., one changeset at a time? I use eclipse, subversion, and subclipse, and I can't change the former two for the time being (or the MS platform..). In my list/ overview a file seems to be listed only in the latest relevant changeset even if all changesets are listed. So an earlier changeset doesn't necessarily show the full set of files in the original commit, nor the original diff for a file in a commit. Update: I'm thinking about using changesets for simplified peer review, so I'd like the partial update represented for all the files commited in one changeset. It's easy to get diffs and specific revisions for specific files in eclipse, but I'd like to step through all the changes in one specific commit/ changeset in a practical manner.

    Read the article

  • Lots of dropped packages when tcpdumping on busy interface

    - by Frands Hansen
    My challenge I need to do tcpdumping of a lot of data - actually from 2 interfaces left in promiscuous mode that are able to see a lot of traffic. To sum it up Log all traffic in promiscuous mode from 2 interfaces Those interfaces are not assigned an IP address pcap files must be rotated per ~1G When 10 TB of files are stored, start truncating the oldest What I currently do Right now I use tcpdump like this: tcpdump -n -C 1000 -z /data/compress.sh -i any -w /data/livedump/capture.pcap $FILTER The $FILTER contains src/dst filters so that I can use -i any. The reason for this is, that I have two interfaces and I would like to run the dump in a single thread rather than two. compress.sh takes care of assigning tar to another CPU core, compress the data, give it a reasonable filename and move it to an archive location. I cannot specify two interfaces, thus I have chosen to use filters and dump from any interface. Right now, I do not do any housekeeping, but I plan on monitoring disk and when I have 100G left I will start wiping the oldest files - this should be fine. And now; my problem I see dropped packets. This is from a dump that has been running for a few hours and collected roughly 250 gigs of pcap files: 430083369 packets captured 430115470 packets received by filter 32057 packets dropped by kernel <-- This is my concern How can I avoid so many packets being dropped? These things I did already try or look at Changed the value of /proc/sys/net/core/rmem_max and /proc/sys/net/core/rmem_default which did indeed help - actually it took care of just around half of the dropped packets. I have also looked at gulp - the problem with gulp is, that it does not support multiple interfaces in one process and it gets angry if the interface does not have an IP address. Unfortunately, that is a deal breaker in my case. Next problem is, that when the traffic flows though a pipe, I cannot get the automatic rotation going. Getting one huge 10 TB file is not very efficient and I don't have a machine with 10TB+ RAM that I can run wireshark on, so that's out. Do you have any suggestions? Maybe even a better way of doing my traffic dump altogether.

    Read the article

  • Why when trying to restore a SQL DB from backup, it keeps referring to backup's original location an

    - by rm
    I've done a full DB backup to C:\Backups\MyDb.bak Then I've setup a job to incrementally back the DB up every day to that same location. Then I've moved the backup files to X:\Backups\, changing the target directory in backup job Now that I try to restore the database from backup, it keeps trying to refer to C:\Backups\MyDb.bak, but that file is no longer there. How do I fix the issue w/o having to move the backup files back to C:\Backups?

    Read the article

  • I'm really offtopic. But I've got a really good reason.

    - by lost
    Is there anyway Encryption on an unidentified file can be broken(file in question: config file and log files from ardamax keylogger). These files date back all the way to 2008. I searched everywhere, nothing on slashdot, nothing on google. Ardamax Keyviewer? Should I just write to Ardamax? I am at a loss of what to do. I feel comprimised. Anyone managed to decrpyt files with Crypto-analysis?

    Read the article

  • ASP.NET - fast Segmented downoad of file through webservice

    - by H4mm3rHead
    Hi, Im doing this project where i need to download files through a webservice (images, videos). The download MUST go through an existing webservice. The existing webservice was made when there were no need to upload and download files but the project has changed and now we need to do It through a webservice. Right now I have implemented the download as a method that returns a byte[], I open a streamreader and resds the entire file into a byte[] and returns it to my method. This is working file on small files <~1Mb, above it takes too long time. I want to show some progress (e.g. when the user downlaods a 20Mb video) which i cannot do right now. And i want to make it download much faster (is a strategy to use multithreading and several threads that downloads a part of the file?). It is within a WPF application i need to do this. Any ideas on how to approach this?

    Read the article

  • Save batch file path to local harddisk

    - by Brian
    Hi, I have a batch file that is Run when a selfextraction file is executed. The self extracted files must be copied to to specefic place on the harddisc. In the batch file the user is asked where the path is (if it's not located the default place) Part of the batch file: @ECHO OFF IF EXIST "C:\Program Files\program\program.exe". ( set PROGRAMPATH=C:\Program Files\ ) ELSE ( echo Program folder was not found. Please enter the path for Program set /p PROGRAMPATH=Path: ) echo Copying data to "%PROGRAMPATH%"... copy /Y "*.txt" "%PROGRAMPATH%" Now for my question. If a user then enters a new path, is it possible to save that path. So when he executes the self extraction packagde again, it could remember that new path?

    Read the article

  • Looking for Fiddler2 help. connection to gateway refused...just got rid of a virus?

    - by john mackey
    I use Fiddler2 for facebook game items, and it's been a great success. I accessed a website to download some dat files I needed. I think it was eshare, ziddu or megaupload, one of those. Anyway, even before the rar file had downloaded, I got this weird green shield in the bottom right hand corner of my computer. It said a Trojan was trying to access my computer, or something to that extent. It prompted me to click the shield to begin anti-virus scanning. It turns out this rogue program is called Antivirus System Pro and is pretty hard to get rid of. After discovering the rogue program, I tried using Fiddler and got the following error: [Fiddler] Connection to Gateway failed. Exception Text: No connection could be made because the target machine actively refused it 127.0.0.1:5555 I ended up purchasing SpyDoctor + Antivirus, which I'm told is designed specifically for getting rid of these types of programs. Anyway, I did a quick-scan last night with spydoctor and malware bytes. Malware picked up 2 files, and Spydoctor found 4. Most were insignificant, but it did find a worm called Worm.Alcra.F, which was labeled high-priority. I don’t know if that’s the Anti-Virus Pro or not, but SpyDoctor said it got rid of all of those successfully. I tried to run Fiddler again before leaving home, but was still getting the "gateway failed" error. Im using the newest version of firefox. When I initially set up the Fiddler 2.2.8.6, I couldn’t get it to run at first, so I found this faq on the internet that said I needed to go through ToolsOptionsSettings and set up an HTTP Proxy to 127.0.0.1 and my Port to 8888. Once I set that up and downloaded this fiddler helper as a firefox add-on, it worked fine. When I turn on fiddler, it automatically takes my proxy setting from no proxy (default) to the 127.0.0.1 with Port 8888 set up. It worked fine until my computer detected this virus. Anyway, hopefully I've given you sufficient information to offer me your best advice here. Like I said, Spydoctor says the bad stuff is gone, so maybe the rogue program made some type of change in my fiddler that I could just reset or uncheck or something like that? Or will I need to completely remove fiddler and those dat files and rar files I downloaded? Any help would be greatly appreciated. Thanks for your time.

    Read the article

  • Passing arguments to java vm from NSIS script

    - by CodeBuddy
    I'm developing my first java application using Eclipse. I've recently needed to adjust the amount of memory allocated by passing -Xmx256M to the JVM. The application is currently package up as a runnable jar and installed using the NSIS. I'm having a problem passing arguments to the jar file once its installed. What is the common practice for doing this? Here is what I'm currently doing in my nsi file: CreateShortcut "$SMPROGRAMS\$StartMenuGroup\$(^Name).lnk" "$SYSDIR\javaw.exe" "-jar -Xmx256M $INSTDIR\Foo.jar" This results in the following being created as the shortcut Target on windows: C:\WINDOWS\system32\javaw.exe -jar -Xmx256M C:\Program Files\Foo\Foo.jar Unfortunately this does not work due to the space in C:\Program Files, If I change the link created manually to include quotes all is well: C:\WINDOWS\system32\javaw.exe -jar -Xmx256M "C:\Program Files\Foo\Foo.jar"

    Read the article

  • Web App fails when moved to production environment. Which server permissions do I need?

    - by Ashley Ward
    I have developed a small web app. This app allows users to upload images. It also produces text files with the names of those images (the names are stored and retrieved to/from an MySQL Database.) I have developed this app using MAMP. To create the uploaded image files I use the PHP function imagejpeg('my/path/name.jpg') and to delete the files I use the PHP function unlink('folder1/folder2/name.jpg') to write to the text document I am using the function fopen('folder1/folder2/name.txt', 'w') all three of these functions produce errors related to permissions - now the site has been moved to a live hosting environment. Why is this? and what permissions do I need to set the folder's folder1 and folder2 to? I know that permission 777 is generally bad because it opens up your server to the public. However what I have found is that the functions fail to work unless I use 777 on the folders. Can anyone shed any light on my dilemma?

    Read the article

  • Windows Server 2008 Software Raid 5 - Data integrity issues

    - by Fopedush
    I've got a server running Windows Server 2008 R2, with a (windows native) software raid-5 array. The array consists of 7x 1TB Western Digital RE3 and RE4 drives. I have offline backups of this array. The problem is this: I noticed a few days ago after copying a large file to the disk that there was an integrity issue with that file - it was a ~12GB file that I had downloaded via uTorrent. After moving it to the raid array, I used uTorrent to relocate the download location, and performed a re-check so I could seed it from that location. The recheck found that only 6308/6310 chunks of the copied file were intact. My next step was to write a quick powershell script that would copy files to the array, while performing a SHA1 hash of the original and resultant files and comparing them. Smaller files (100-1000MB) copied over just fine. When I started copying larger data (~15GB), I found that the hash check failed about 2/3rds of the time. The corrupt files had very, very small inconsistencies - less than .01%. I further eliminated the possibility of networking or client issues by placing this large file on the C:\ of the server, and copying it repeatedly from there to the array, seeing similar results. Copying the data via explorer, powershell, or the standard windows command prompt yield the same results. None of the copies fail or report any problems. The raid array itself is listed as healthy in disk management. After a few experiments, I shut down the server and ran memtest overnight. No errors were detected. A basic run of chkdsk found no problems, but I did not use the /R flag, as I was unsure how that might affect a software raid-5 volume. I next ran Crystal Disk Info to check the smart data on the drives - but found that CDI only detected 5 out of 7 of the disks in the array. I have no idea why. Nevertheless, CDI shows the following "caution" flags on a single one of the drives: 05 199 199 140 000000000001 Reallocated Sectors Count C5 200 200 __0 000000000001 Current Pending Sector Count Which is a little bit alarming, but I don't really know what to do with the information. I hardly feel like one reallocated sector could be causing this. At this point, I'm looking for some guidance on what to do next. I need to determine the cause of this issue, but I'm hesitant to run chkdsk /R or any bootable disk health checkers because I'm afraid they might break the array. I've considered triggering a re-sync of the array, but I'm not actually sure how to do that without doing something silly like manually dropping a disk and then restoring it. Any advice that could help me ferret out the precise cause of this issue would be greatly appreciated.

    Read the article

  • eclipse gcov plugin

    - by Taani
    I am using gcov plugin in Eclipse CDT + cygwin compiler & linker to find code coverage. I installed gcov plugin from http://svn.sourceforge.jp/svnroot/ginkgo/trunk/org.ginko.gcov.update/ location and follow the below link to get coverage of my .c file. I was able to create .gcno and .gcda files. But double clicking on those files, opens the files in eclipse editor. I could not find the gcov editor. Also I couldn't see any code coverage results in my .c file. What is the wrong with my procedure?

    Read the article

  • Merging 3 apps - 2.msi and 1.exe

    - by Netguy
    EXACT duplicate of Combining three MSI into a single installer Hi Guys , I am having 3 files - Alky for Applications.msi ( which make Vista Apps work on XP) 2. Windows VIsta sidebar.exe ( Which make that VIsta sidebar work on XP) *3.Gadget Extractor.ms*i ( A part of number 2) Now , the problem is that all the 3 applications are installers and I want to merge them to 1 installer . So please tell me what should I do and I also want to remove some content( normal files) from 2. Note: I do NOT want to bind the files , so that 3 installers start at the same time. I want to make them into one The Person who is able to help me gets a VPS with cPanel with RL/TF allowed :D

    Read the article

  • Windows 7-File Copy Issue

    - by Manoj M
    I am using Windows 7 and my application server is Windows Xp. When I copy files from my local to server,it takes long time. Is there any solution for solving this ? When I copy files from Xp to Xp, it is fast. I found a soln in .net http://www.windowsreference.com/windows-7/slow-network-file-copy-issues-in-windows-7-caused-by-remote-differential-compression/ I tried it ,but no change

    Read the article

  • How to convert an CHM file into a single HTML file?

    - by ruslan
    I have tried many different CHM-to-HTML utilities, but I am having a difficult time finding one that is able to produce a single HTML file. I can decompile a CHM file using hh.exe, but I don't know how to easily merge the resulting files into a single HTML file, all while preserving the correct order of pages. Is there a free tool which can do this? If not, how can I merge the HTML files in order?

    Read the article

  • Is adding users to the group www-data safe on Debian?

    - by John
    Many PHP applications do self-configuration and self-updating. This requires apache to have write access to the PHP files. While chgrp'ing them all to www-data appears like a good practice to avoid making them world writable, I also wish to allow users to create new files and edit existing one. Is adding users to the group www-data safe on Debian? For example: 775 root www-data /var/www 644 john www-data /var/www/johns_php_application.php 660 john www-data /var/www/johns_php_applications_configuration_file

    Read the article

  • Compilation error on a user control

    - by MikeP
    I'm stumped! We have a user control for managing account information. We use this particular control on two pages. On one page, everything works perfectly and meets our expectation. On the second page however we receive compilation errors stating that: "C:\WINDOWS\Microsoft.NET\Framework\v2.0.50727\Temporary ASP.NET Files\lrpcentral\0e987bea\6719c8b6\App_Web_PageThatFails.aspx.f3d462c1.oi52bvii.0.cs(172): error CS0433: The type 'xxxx_ascx' exists in both 'c:\WINDOWS\Microsoft.NET\Framework\v2.0.50727\Temporary ASP.NET Files\APPLICATIONNAME\0e987bea\6719c8b6\App_Web_xxxx.ascx.cdcab7d2.xbnvt2za.dll' and 'c:\WINDOWS\Microsoft.NET\Framework\v2.0.50727\Temporary ASP.NET Files\APPLICATIONNAME\0e987bea\6719c8b6\App_Web_eix7xllr.dll' My problem is similar to Cyril's but the "delete everything from Temp" is not an option for me, and Cyril's solution does not apply since the only variable we have is contained in the designer file, which is not deployed to our production environment (we pre-compile). After reading David's answer (here) I examined my directories for circular dependency and was unable to find any. Structure: Top Level Page that works Control Directory A Page that causes the error

    Read the article

  • Excel 2007 - "The Document Information Panel was unable to load"

    - by Ruffles
    In Excel 2007, if I go to Office button - Prepare - Properties instead of showing document properties, I get a message "The Document Information Panel was unable to load." I have come across a number of posts suggesting I copy ipedintl.dll from C:\Program Files\Microsoft Web Designer Tools\Office12\1033 to C:\Program Files\Microsoft Office\Office12\1033, however I have tried this, and the problem is still occurring. Does anyone have any other ideas of what this might be? Thanks.

    Read the article

  • Limit user profile on specific OU using Group Policy

    - by Sergei
    We have a host that will be used for creating VM clones from time to time for testing purposes.It is used actively for testing and users tend to keep a lot of files in their profiles.We would like to limit users profile size to avoid cloning unnecessary files to new VMs. Is there way to impose limit on user profile on OU level without introducing roaming profiles?

    Read the article

  • How do I find a file that begins with a phrase in Windows Search?

    - by plasmuska
    Hi Guys, What is the syntax for searching a file with file name that STARTS with a certain phrase? Example: I have two files: 60933 blahblah.xls PZ 60933 blahblah.xls I would like to search only for the first one but Windows Search always returns two results. I have tried these but none of them seem to work: filename:60933 filename:^60933* filename:60933..xls My setup: Windows XP Pro CZ, Windows Search 4, files are located on indexed network share.

    Read the article

  • Best HTPC solution that supports MKV

    - by wag2639
    So it seems like there are many solutions (Boxee, GoogleTV, AppleTV, XMBC, etc) out there for HTPC setups but I can't seem to find a good solution for playing MKV files and I would say the bulk of my video library is encoded in MKVs. Whats a good solution for an HTPC if I want the following: NAS / Samba support to use files from server or Windows share Preferably Linux based solution Cleanish UX Nice to have but not a requirement: - Android Remote App so I can control with my Android device

    Read the article

  • avoiding console window display when scheduled task runs batch file

    - by cherouvim
    I have a small batch file which xcopies some files from one folder to another which I've scheduled (via windows scheduled tasks) to run every 1 hour: @echo off xcopy c:\foo c:\bar /E /C /F /Y Since this is my workstation, I'm most probably doing work when the task executes, and then the black dos console window is displayed (lasts 2-3 seconds) and steals window focus. I don't wish to see the files copied and of course the batch file does not ask for any user input. Is there a way to avoid displaying the console completely?

    Read the article

  • How to make `rm` faster on ext3/linux?

    - by depesz
    I have ext3 filesystem mounted with default options. On it I have some ~ 100GB files. Removal of any of such files takes long time (8 minutes) and causes a lot of io traffic, which increases load on server. Is there any way to make the rm not as disruptive?

    Read the article

  • emacsclient: create a frame if a frame does not exist

    - by Idlecool
    I start emacs server using emacs --daemon then open files using emacsclient filename.ext but the first file has to be opened using emacsclient -c filename.ext in order to create a new frame which can be later used by subsequent files without using -c command line flag for emacsclient. I want to automate this. "if there is no emacs frame, emacsclient should create a frame else it should use the current frame". How can it be done?

    Read the article

  • Windows 2003 lost file extensions

    - by pho3nix
    I have a server with windows 2003, recently i installed a software in this server. When i restarted server my all files can't open sending to a command prompt saying: No program associated with ".exe" file extension. This occurred to all files types in system. When i go see file types association i see this: C:\WINDOWS\System32\WScript.exe "%1" %*, change result but when restart return to same. Any idea to resolve?

    Read the article

< Previous Page | 494 495 496 497 498 499 500 501 502 503 504 505  | Next Page >