Search Results

Search found 37650 results on 1506 pages for 'files'.

Page 235/1506 | < Previous Page | 231 232 233 234 235 236 237 238 239 240 241 242  | Next Page >

  • Join multiple consecutive SQLite database dump files into 1 common database? Purpose: Search through ENTIRE Chrome Browsing History

    - by porg
    Google Chrome 's default web browsing history search engine only lets you access the records of the recent 100 days. Nevertheless in your application data, Chrome keeps your entire browsing history in SQLite database files, with the file naming scheme of "History Index YYYY-MM". I am looking for a way to search… …through my entire browsing history, …with sophisticated filters (limit search terms to certain fields such as URL, domain, title, body text; wildcard or regex terms, date ranges). … in … …either some ready-made software. eHistory came close, as it can limit terms to fields, but it lacks wildcards/regexes, and has the same limited time horizon as the default search. Beyond that, I could not find any suited Chrome extension or standalone (Mac) app. …or a command line to join multiple SQLite database files into one database, which I can then query (with the full syntax power). In the spirit of the pseudo code below: Preferred this way: sqlite --targetDatabase ChromeHistoryAll --importFiles /path/to/ChromeAppData/History\ Index* --importOnlyYetUnknownFiles Or if my desired feature --importOnlyYetUnknownFiles is not possible (feature could also be called "avoid duplicate imports by checking UIDs"), then by explicitly only importing files, of which I know, that they have yet not been imported into the ChromeHistoryAll database: cd ChromeAppData; sqlite --databaseTarget ChromeHistoryAll --importFiles YetNotImported1 YetNotImported2 YetNotImported3 All my queries I would then perform in the database "ChromeHistoryAll" P.S.: Additional question of general interest: Is there a way to perform a database query in a temporary database which was created on-the-fly from multiple files? Like: sqlite --query="SQL query" --targetDatabase DbAll --DBtemporaryInRAM --importFiles db1 db2 db3 This is surely not applicable for my Chrome question, as these History Index files have a combined file size of 500MB together, thus such a query would be of bad performance. But it could come handy in other situations.

    Read the article

  • How do I configure Gnome 3 so that it doesn't pop up a dialog for 'open with files' when I mount a drive?

    - by michael
    I am running Gnome 3 on Ubuntu 11.10. In the file manager, when I click a drive under 'Devices', Gnome 3 always pops up a dialog with the choices 'open with files' and 'eject' and then I need to click 'open with files' to get rid of that dialog. Is there a way to configure Gnome 3 not to do that? I am in file manager already, clicking a drive should show the content in the right pane. Why does it still ask me to 'open with files'?

    Read the article

  • I installed XAMPP on Ubuntu, and its running without any of its files.

    - by CDeanMartin
    I installed XAMPP on Ubuntu, and its running without any of its files. Well, not really, but it sure seems that way. I followed the directions on the Apache Friends website to the letter. XAMPP works fine, it's sample apps run like a charm. But I have no idea how it is running. The installation should have created an otf folder, or a XAMPP folder, or an htdoc file according to the tutorials. But there is no trace of any of these files or folders anywhere in the file browser. Are there any linux utilities as good as the Windows Explorer that can find my missing files? I particularly need to find the var/www folder to put my .php files in.

    Read the article

  • (How) does deleting open files on Linux and a FAT file system work?

    - by lxgr
    It's clear to me how deleting open files works on filesystems that use inodes - unlink() just decreases the link count to zero, and when the last file handle to the file is closed, the inode will be removed. But how does it work when using a file system that doesn't use inodes, like FAT32, with Linux? Some experiments suggest that deleting open files is still possible (unlike on Windows, where the unlink call wouldn't succeed), but what happens when the file system is uncleanly unmounted? How does Linux mark the files as unlinked, when the file system itself doesn't support such an operation? Is the directory entry just deleted, but retained in memory (that would guarantee deletion after unmounting in any case, but would leave the file system in an inconsistent state), or will the deletion only be marked in memory, and written at the time the last file handle is closed, avoiding possible corruption, but restoring the deleted files after an unclean unmount?

    Read the article

  • Add closure after concatenating files or keep one for each separate file?

    - by sdespolit
    At the moment every javasrcipt file in our project includes closure wrapper. Before deploying all files are concatenated and minimized. And i'm asking myself, do we really need all that closures OR one closure added after concatenation will do the job. I can see only one reason not to do that - the variable we are sharing between files will get in the global namespace. But there is no conflict so is don't see any problem. The real question is what other cons may be? Or more generally what approach is better - one closure or many?

    Read the article

  • Files: Name column basically useless in list view. Feature or bug?

    - by Luksurious
    So the following is happening in Files whenever the width of the window is somewhat smaller than all the content of the list: In list view the name column is cropped unto a point where no name is visible at all! And it is even not possible to change the column size manually. Funnily, in some situations it quickly flickers from a larger column width to the small one back and forth before settling on the small size. Unnecessarily to say, it is extremely annoying. Is there a way around this or is this just "bad design"? Oh yeah, Ubuntu 13.04 & 13.10, Files version 3.8.2.

    Read the article

  • How to copy a folder with many files with integrity check?

    - by RafaelM
    I just got a new hard drive and I want to move many of the folders from the old hard drive to the new one, but I want to make sure everything is copied over correctly. I tried using md5summer to generate md5 sums of the original files, copied files over and then tried to compare md5 sums of both sets of files. This took ages because there are many large video files. Is there any software I can use to make this process as painless as possible? I just need basic file integrity checking. Thanks in advance

    Read the article

  • How can I decrypt encrypted files using a PEM private key?

    - by Phil Cole
    I have files which have either been encrypted with a public key and the Blowfish algorithm, or a public key and the AES-256 algorithm. I'm looking to put together a Perl script that would be able to use the private keys (which I do have) to decrypt the files. The public and private key files are all in PEM format, and while I can find ways of reading the PEM files, and ways of decrypting data with a key, I haven't yet found a way of going from PEM - key. Any suggestions?

    Read the article

  • How do I convert Word files to PDF programmatically?

    - by Shaul
    I have found several open-source/freeware programs that allow you to convert .doc files to .pdf files, but they're all of the application/printer driver variety, with no SDK attached. I have found several programs that do have an SDK allowing you to convert .doc files to .pdf files, but they're all of the proprietary type, $2,000 a license or thereabouts. Does anyone know of any clean, inexpensive (preferably free) programmatic solution to my problem, using C# or VB.NET? Thanks!

    Read the article

  • Index a set of files to search their text quickly?

    - by Ricket
    I have a unique need: I am frequently searching a large set of text files for a keyword. Right now, I open up Notepad++ and use the "Find in files" feature. It works just fine, but with the amount of files, each search takes several minutes to complete. Is there a good program more suited for this purpose, perhaps that indexes a set of files and then lets you search the set repeatedly and very quickly? It would greatly speed up my workflow.

    Read the article

  • What files should be added to SVN in an eclipse Java project?

    - by Jake Petroules
    I have a Java project I'd like to commit to my SVN repository, created with eclipse. Now, what files (aside from the source code, obviously) are necessary? In the workspace root, there is a .settings folder with many files and subfolders, and inside the project folder there are two files - .classpath and .project, and another .settings folder with a single file - org.eclipse.jdt.core.prefs. Which of these files should be committed to SVN and which can be safely excluded?

    Read the article

  • how to copy files from TFS to clear case??

    - by barry
    Can i use powershell script to copy a set of files from a folder to Clear Case..?? i have the task of synchronising files from TFS to Clear Case.. like i need to take a set of files aftr a certain date from tfs server and synchronise these files to Clear case..

    Read the article

  • What the best approach to iterate and "store" files over a directory in C (Linux) ?

    - by Andrei Ciobanu
    I have written a function that checks if to files are duplicates or not. This function signature is: int check_dup_memmap(char *f1_name, char *f2_name) It returns: (-1) - If something went wrong; (0) - If the two files are similar; (+1) - If the two files are different; The next step is to write a function that iterates through all the files in a certain directory,apply the previous function, and gives a report on every existing duplicates. Initially I've thought to write a function that generates a file with all the filenames in a certain directory and then, read that file again and gain and compare every two files. Here is that version of the function, that gets all the filenames in a certain directory. void *build_dir_tree(char *dirname, FILE *f) { DIR *cdir = NULL; struct dirent *ent = NULL; struct stat buf; if(f == NULL){ fprintf(stderr, "NULL file submitted. [build_dir_tree].\n"); exit(-1); } if(dirname == NULL){ fprintf(stderr, "NULL dirname submitted. [build_dir_tree].\n"); exit(-1); } if((cdir = opendir(dirname)) == NULL){ char emsg[MFILE_LEN]; sprintf(emsg, "Cannot open dir: %s [build_dir_tree]\t",dirname); perror(emsg); } chdir(dirname); while ((ent = readdir(cdir)) != NULL) { lstat(ent->d_name, &buf); if (S_ISDIR(buf.st_mode)) { if (strcmp(".", ent->d_name) == 0 || strcmp("..", ent->d_name) == 0) { continue; } build_dir_tree(ent->d_name, f); } else{ fprintf(f, "/%s/%s\n",util_get_cwd(),ent->d_name); } } chdir(".."); closedir(cdir); } Still I consider this approach a little inefficient, as I have to parse the file again and again. In your opinion what are other approaches should I follow: Write a datastructure and hold the files instead of writing them in the file ? I think for a directory with a lot of files, the memory will become very fragmented. Hold all the filenames in auto-expanding array, so that I can easy access every file by their index, because they will in a contiguous memory location. Map this file in memory using mmap() ? But mmap may fail, as the file gets to big. Any opinions on this. I want to choose the most efficient path, and access as few resources as possible. This is the requirement of the program... EDIT: Is there a way to get the numbers of files in a certain directory, without iterating through it ?

    Read the article

  • How to include files from remote server in Eclipse project without copy to local PC?

    - by user209559
    I have to PC, one is server on Linux containing project files ( also build machine ) and another working desktop machine on WinXP. I want to create a project in Eclipse on my desktop machine without coping files to local machine, actually I want to be able to modify remote files and immediately run build, unlike modifying local files and synchronizing with remote project.Is it possible? Thanks

    Read the article

  • Is there a way to efficiently yield every file in a directory containing millions of files?

    - by Josh Smeaton
    I'm aware of os.listdir, but as far as I can gather, that gets all the filenames in a directory into memory, and then returns the list. What I want, is a way to yield a filename, work on it, and then yield the next one, without reading them all into memory. Is there any way to do this? I worry about the case where filenames change, new files are added, and files are deleted using such a method. Some iterators prevent you from modifying the collection during iteration, essentially by taking a snapshot of the state of the collection at the beginning, and comparing that state on each move operation. If there is an iterator capable of yielding filenames from a path, does it raise an error if there are filesystem changes (add, remove, rename files within the iterated directory) which modify the collection? There could potentially be a few cases that could cause the iterator to fail, and it all depends on how the iterator maintains state. Using S.Lotts example: filea.txt fileb.txt filec.txt Iterator yields filea.txt. During processing, filea.txt is renamed to filey.txt and fileb.txt is renamed to filez.txt. When the iterator attempts to get the next file, if it were to use the filename filea.txt to find it's current position in order to find the next file and filea.txt is not there, what would happen? It may not be able to recover it's position in the collection. Similarly, if the iterator were to fetch fileb.txt when yielding filea.txt, it could look up the position of fileb.txt, fail, and produce an error. If the iterator instead was able to somehow maintain an index dir.get_file(0), then maintaining positional state would not be affected, but some files could be missed, as their indexes could be moved to an index 'behind' the iterator. This is all theoretical of course, since there appears to be no built-in (python) way of iterating over the files in a directory. There are some great answers below, however, that solve the problem by using queues and notifications. Edit: The OS of concern is Redhat. My use case is this: Process A is continuously writing files to a storage location. Process B (the one I'm writing), will be iterating over these files, doing some processing based on the filename, and moving the files to another location. Edit: Definition of valid: Adjective 1. Well grounded or justifiable, pertinent. (Sorry S.Lott, I couldn't resist). I've edited the paragraph in question above.

    Read the article

  • Spotlight query to search for all archives and ISO files that their names do not end in bin

    - by Bytecode Ninja
    The following Spotlight query, correctly returns all archive and ISO files on my system: kind:iso OR kind:archive However, Spotlight treats .bin files as archives too and returns .bin files in the results as well. How can I modify the above query and exclude files that end in bin from the result? I tried the following query but it is not working: (kind:iso OR kind:archive) AND NOT name:*bin Thanks in advance.

    Read the article

  • Can Visual Studio manage function prototypes for me in C++ header files?

    - by Gibybo
    In C++, the common practice is to declare functions in header files and define them in cpp files. This leads to always having two copies of every function's prototype. Then whenever I want to change a function's name/return value/parameter, I have to manually change it in both files. This seems unnecessarily tedious and there must be lots of people that share my pain, so is there a way to automate these changes between files in VS? Bonus points for vim solutions as well.

    Read the article

  • ASP C# How to handle large files from database?

    - by Martijn
    In my web application I am working with files. Some files are very large. I use Response.Write() to write the file to the browser. This goes well for the smaller files, but for large files this can take a while and the bandfwith is fully used. My question is, is it possible to split large documents and send it piece by piece to the browser? I hold the document as a property of an object.

    Read the article

  • How to monitor a directory for files in C++?

    - by sand
    I need to monitor a directory which contains many files and a process reads and deletes the .txt files from the directory; once all the .txt files are consumed, the consuming process needs to be killed. How do I check if all the .txt files are consumed using C++? I am developing my application on Visual Studio on windows platform.

    Read the article

  • How to make two different files (with their contents) equal using command in linux shell?

    - by user2617138
    How to make two different files (their contents) equal using terminal in Linux? Suppose i have a file A in which the content is Hello world and i have a file B in which the content is Hello worlds. Now we find the difference between the 2 files using the diff or sdiff utility. Now i want to append the contents of the 2 different files into a single file or make the two different files (their contents) equal using terminal or any other utility.

    Read the article

< Previous Page | 231 232 233 234 235 236 237 238 239 240 241 242  | Next Page >