Search Results

Search found 70517 results on 2821 pages for 'file creation'.

Page 270/2821 | < Previous Page | 266 267 268 269 270 271 272 273 274 275 276 277  | Next Page >

  • Javascript reference external script file - security implications

    - by rkrauter
    Hi, If I have a reference to an external third party JavaScript file on my website, what are the security implications? Can the JavaScript file be used to steal cookies? One example of this is the Google Analytics JavaScript reference file. Could the third party technically steal cookies or any other sensitive information from my logged on users (XSS)? The whole cross domain scripting has me confused sometimes. Thanks!

    Read the article

  • Batch to move files in a folder to another folder?

    - by Mohammad
    Hello People, I want to create a batch file that moves all the files in this directory to its father directory. Like for example if I have a directory " Something " that contains files : ...New Folder/Something And I want to move the files in " Something " to New folder. Is there such a command? Thanks a bunch :)

    Read the article

  • Read binary data from a MDB-file running under LAMP

    - by BusterX
    I need to be able to connect to an MDB-file in a LAMP-environment (running on Linux) and ultimately insert converted data into a Mysql db. The data I need to access is stored as a BLOB (Long Binary Data according to Access) in the MDB file. I have not yet been able to actually have a look at the data but I have been told that the BLOB consists of byte strings. Something along the lines of: 0x1c 0x10 0x27 0x00 0x00 I need to parse the byte strings and convert these to a format that is human readable. I do have access to the documentation that explains the various byte strings. So this is really two questions: How do a get access to the MDB file via PHP* (running under LAMP) and read the BLOB (I do not have access to a Windows-platform)? What would be the best way to parse the binary data (in PHP*) once I am able to connect to the MDB-file? *Or are there other methods/languages that are more appropriate?

    Read the article

  • Client validation of INPUT of type FILE without postback using jQuery

    - by Fixer
    I want to check on the client side that a file has been selected before the form can be submitted. <script src="@Url.Content("~/Scripts/jquery.validate.min.js")" type="text/javascript"></script> <script src="@Url.Content("~/Scripts/jquery.validate.unobtrusive.min.js")" type="text/javascript"></script> @using (Html.BeginForm("Upload", "Files", FormMethod.Post, new { enctype = "multipart/form-data" })) { <input id="File" name="File" type="file" size="80" /> <input type="submit" name="name" value="Upload" /> } Currently this form is doing postbacks for validation. What is going wrong?

    Read the article

  • Jquery Ajaxform with input type=file and multifile

    - by crunchingnumbers
    Im using ajax form with multifile. However ajaxform just seems to quietly die and does not do anything if using multifile. Multifile is just a jquery plugin that builds a list of input type=file so that you can upload multiple files at once, before which you can add/delete your file upload list. I've looked at multifile and made one change to ensure that it generated a unique name as well as id which it didn't do before and i've checked the form elements multifile is building which all appear to be correct. From looking at ajaxform, it seems that it shoulld just generate the iframe as normal and post the multiple input type=file but it does nothing. Has anyone else experienced problems with multiple input type=file uploads

    Read the article

  • How to estimate memory need by XPathDocument for a specific xml file

    - by bill seacham
    Is there any way to estimate the memory requirement for creating an XpathDocument instance based on the file size of the xml? XpathDocument xdoc = new XpathDocument(xmlfile); Is there any way to programmatically stop the process of creating the XpathDocument if memory drops to a very low level? Since it loads the entire xml into memory, it would be nice to know ahead of time if the xml is too big. What I have found is that when I create a new XpathDocument with a big xml file, an outofmemory exception is never fired, but that the process slows to a crawl, only 5 Mb of memory remains a available and the Task Manager reports it is not responding. This happened with a 266 Mb xml file when there was 584 Mb of ram. I was able to load a 150 Mb file with no problems in 18. After loading the xml, I want to do xpath queries using an XpathNavigator and an XpathNodeIterator. I am using .net 2.0, xp sp3.

    Read the article

  • Read Excel 2007 file (.xlsx) in C#

    - by Safiullah
    My code was quite sufficient for reading xls file in C# (using Microsoft.Jet.Oledb.4.0) but when I tried to read Office 2007 (xlsx) file, it didn't work. I searched and found that it could be read using Microsoft.ACE.Oledb.12, I tried and it worked. Can anyone tell me the difference between the two ? and why xlsx file can not be read using Microsoft.Jet.Oledb.4.0 Provider.

    Read the article

  • How can i run code on the client side from a browser?

    - by acidzombie24
    With LLVM and silverlight this may be possible now (or it may be possible with flash). I like the user to select a file and then do the following things 1) Hash it with md5 and sha1 2) If archive check if an exe is in it 3) If archive check if password protected The first to see if the user has uploaded it already (today, yesterday, last month) 2nd to prevent viruses 3rd i should be fine without but if i decide to not allow protected archives i can warn before the user uploads it. How may i do this through the browser?

    Read the article

  • shell script filter du and find by a string inside a file in a subfolder

    - by Jason
    I have the following command that I run on cygwin: find /cygdrive/d/tmp/* -maxdepth 0 -mtime -150 -type d | xargs du --max-depth=0 > foldersizesreport.csv I intended to do the following with this command: for each folder under /d/tmp/ that was modified in last 150 days, check its total size including files within it and report it to file foldersizesreport.csv however that is now not good enough for me, as it turns out inside each /d/tmp/subfolder1/somefile.properties /d/tmp/subfolder2/somefile.properties /d/tmp/subfolder3/somefile.properties /d/tmp/subfolder4/somefile.properties so as you see inside each subfolderX there is a file named somefile.properties inside it there is a property SOMEPROPKEY=3808612800100 (among other properties) this is the time in millisecond, i need to change the command so that instead of -mtime -150 it will include in the whole calculation only subfolderX that has a file inside them somefile.properties where the SOMEPROPKEY=3808612800100 is the time in millisecond in future, if the value SOMEPROPKEY=23948948 is in past then dont at all include the folder in the foldersizesreport.csv because its not relevant to me. so the result report should be looking like: /d/tmp/,subfolder1,<itssizein KB> /d/tmp/,subfolder2,<itssizein KB> and if subfolder3 had a SOMEPROPKEY=34243234 (time in ms in past) then it would not be in that csv file. so basically I'm looking for: find /cygdrive/d/tmp/* -maxdepth 0 -mtime -150 -type d | <only subfolders that have in them property in file SOMEPROPKEY=28374874827 - time in ms in future and not in past | xargs du --max-depth=0 > foldersizesreport.csv

    Read the article

  • File Output in Powershell without Extension

    - by CaptHowdy
    Here is what I have so far: Get-ChildItem "C:\Folder" | Foreach-Object {$_.Name} > C:\Folder\File.txt When you open the output from above, File.txt, you see this: file1.txt file2.mpg file3.avi file4.txt How do I get the output so it drops the extension and only shows this: file1 file2 file3 file4 Thanks in advance! EDIT Figured it out with the help of the fellows below me. I ended up using: Get-ChildItem "C:\Folder" | Foreach-Object {$_.BaseName} > C:\Folder\File.txt

    Read the article

  • Looping over commits for a file with jGit

    - by Andy Jarrett
    I've managed to get to grips with the basics of jGit file in terms of connecting to a repos and adding, commiting, and even looping of the commit messages for the files. File gitDir = new File("/Users/myname/Sites/helloworld/.git"); RepositoryBuilder builder = new RepositoryBuilder(); Repository repository; repository = builder.setGitDir(gitDir).readEnvironment() .findGitDir().build(); Git git = new Git(repository); RevWalk walk = new RevWalk(repository); RevCommit commit = null; // Add all files // AddCommand add = git.add(); // add.addFilepattern(".").call(); // Commit them // CommitCommand commit = git.commit(); // commit.setMessage("Commiting from java").call(); Iterable<RevCommit> logs = git.log().call(); Iterator<RevCommit> i = logs.iterator(); while (i.hasNext()) { commit = walk.parseCommit( i.next() ); System.out.println( commit.getFullMessage() ); } What I want to do next is be able to get all the commit message for a single file and then be able revert the single file back to a specific reference/point in time.

    Read the article

  • Reading web-service information from assembly app.config file

    - by Benjamin Ortuzar
    I have a plugin architecture solution written in .NET C# 3.5, where each plug-in is an assembly loaded by the main project. Each plug-in connects to a different web-service, so I would like to have the configuration of that plugin in its own plugin.dll.config file instead of having it in the app.config of the main project. I have been looking around and I saw that i could load from each class its own config file: // Get the application configuration file. System.Configuration.Configuration config = ConfigurationManager.OpenMappedExeConfiguration(fileMap,userLevel) I see how that would help me get the basic settings from the appSettings section, but I cant see a way to read the web-service information stored in the plugin.dll.config file. Any help on how to approach this situation is very welcome.

    Read the article

  • Save Sql Recordset into the Flat text file?

    - by mahesh kumar
    Hi, i need to dump my sql query result into the text file. i have created the following query, DECLARE @cmd VARCHAR(2048) SET @cmd = 'OSQL -localhost -CRN370 ' + ' -UCRN370 -PCRN370' + ' -Q"SELECT TOP 5 GageId FROM EwQMS370..msgages"' + ' -oc:\authors.txt' EXEC master..xp_cmdshell @cmd, NO_OUTPUT The above query created the text file authors.txt. But the content of the file shows the following error message " Error: Conflicting switches : -U and -E " Any help really appreciated

    Read the article

  • How to create file of files?

    - by TheMachineCharmer
    class Node { FooType Data; // I can save Data to file with extension .foo void Save() { // save Data to .foo file } } Now , class Graph { List<Node> Nodes; void Save() { foreach(Node node in Nodes) { node.Save(); } } } Now when I invoke someGraph.Save(); then it creates Nodes.Count number of files. I would like to appear those files as one file with extension somename.graph and be able to read it again as Nodes. How can I do this? Is there a way to bundle files into a single different file?

    Read the article

  • Computer generated files - how do they work ?

    - by hory.incpp
    Hello, .... ‹BÿЃÀ‰D$Ç„$  ....... that's what happens when you open (notepad) such a file that I'm talking about How do algorithms decode that information and when does a program use/generate it ? Does some notepad-like application exist that open such files and transform them to readable code/data ? Any more information which will clarify about these files will be very helpful. Thank you for your time, P.S I'm not talking strictly about .exe files

    Read the article

  • how to store a file handle in perl class

    - by Haiyuan Zhang
    please look at the following code first. #! /usr/bin/perl package foo; sub new { my $pkg = shift; my $self = {}; my $self->{_fd} = undef; bless $self, $pkg; return $self; } sub Setfd { my $self = shift; my $fd = shift; $self_->{_fd} = $fd; } sub write { my $self = shift; print $self->{_fd} "hello word"; } my $foo = new foo; My intention is to store a file handle within a class using hash. the file handle is undefined at first, but can be initilized afterwards by calling Setfd function. then write can be called to actually write string "hello word" to a file indicated by the file handle, supposed that the file handle is the result of a success "write" open. but, perl compiler just complains that there are syntax error in the "print" line. can anyone of you tells me what's wrong here? thanks in advance.

    Read the article

  • read java.security.key stored as object inside a file which is in jar

    - by Tal
    I saved a PublicKey instance in a file using ObjectOutputStream. This file is then stored inside a jar file which is then loaded by JBoss. I'm trying to read this file but it throws me an exception telling that it's not serializable. Here is the code : InputStream input = KeyLoader.class.getClassLoader().getResourceAsStream(resource); ObjectInputStream objectInputStream = new ObjectInputStream(input); Object obj = objectInputStream.readObject(); Key output = (Key) obj; objectInputStream.close(); return output; which throws me this exception An exception occurred: java.io.NotSerializableException

    Read the article

  • Read / Write XML file from Java application bundle

    - by Peter
    Hi there! I've got an XML file that is parsed and written in my application. From within my IDE (Eclipse) I simply address it like this: Reading: private String xmlFile = "file.xml"; and then I build the document: doc = sax.build(xmlFile); Writing is done like this: writer = new FileWriter("file.xml"); Runs great so far, but after bundling my application, the file is no longer accessible. What exactly do I have to do to make it accessible from within an application bundle? I'm absolutely not familiar with the classpath or whatever I need for that, so please go easy on me! Thank you very much!

    Read the article

  • dirname(__FILE__) VS setting global variable to directories

    - by SAFAD
    what are the pros and cons of using this : $globals['server_url'] = dirname(__FILE__); $globals['mainfiles'] = dirname(__FILE__).'/main'; and the pros and cons of using this : $globals['server_url'] = '/srv/www/htdocs/somwhere/'; $globals['mainfiles'] = '/srv/www/htdocs/somwhere/main'; And what do you suggest. by the way: these are set in config.php file which is called by other files as well, to stop directory conflicts when including files we use it like this : require_once($globals['server_url'].'/test.php');

    Read the article

< Previous Page | 266 267 268 269 270 271 272 273 274 275 276 277  | Next Page >