Search Results

Search found 38781 results on 1552 pages for 'offline files'.

Page 401/1552 | < Previous Page | 397 398 399 400 401 402 403 404 405 406 407 408  | Next Page >

  • Problem in doing Add Web Reference (http://.../service.asmx)

    - by flopdix
    I am trying to Add a Web Reference to http//.../service.asmx file in my project. First it gives me an error: "... Unable to download following files from: http://.../service.asmx?wsdl Do you want to skip these files and continue ..." When i click 'Yes', the proxy gets created with .disco and other reference files, but does not adds .wsdl file. Other option i tried, i used http//.../service.asmx?wsdl to add web reference. In this case, i dont get any error, but the proxy gets added with .wsdl and reference files, but it does not adds .disco file. Can someone help me on why this is happening? I thought, adding web reference to .asmx should add everything under the proxy.

    Read the article

  • Proper QUuid usage in Qt ? (7-Zip DLL usage problems (QLibrary, QUuid GUID conversion, interfaces))

    - by whipsnap
    Hi, I'm trying to write a program that would use 7-Zip DLL for reading files from inside archive files (7z, zip etc). Here's where I'm so far: #include QtCore/QCoreApplication #include QLibrary #include QUuid #include iostream using namespace std; #include "7z910/CPP/7zip/Archive/IArchive.h" #include "7z910/CPP/7zip/IStream.h" #include "MyCom.h" // {23170F69-40C1-278A-1000-000110070000} QUuid CLSID_CFormat7z(0x23170F69, 0x40C1, 0x278A, 0x10, 0x00, 0x00, 0x01, 0x10, 0x07, 0x00, 0x00); typedef int (*CreateObjectFunc)( const GUID *clsID, const GUID *interfaceID, void **outObject); void readFileInArchive() { QLibrary myLib("7z.dll"); CreateObjectFunc myFunction = (CreateObjectFunc)myLib.resolve("CreateObject"); if (myFunction == 0) { cout outArchive; myFunction(&CLSID_CFormat7z, &IID_IOutArchive, (void **)&outArchive); } int main(int argc, char *argv[]) { QCoreApplication a(argc, argv); readFileInArchive(); return a.exec(); } Trying to build that in Qt Creator will lead to following error: cannot convert 'QUuid*' to 'const GUID*' in argument passing How should QUuid be correctly used in this context? Also, being a C++ and Qt newbie I haven't yet quite grasped templates or interfaces, so overall I'm having trouble getting through these first steps. If someone could give tips or even example code on how for example an image file could be extracted from ZIP file (to be shown in Qt GUI later on*), I would highly appreciate that. My main goal at the moment is to write a program with GUI for selecting archive files containing image files (PNG, JPG etc) and displaying those files one at a time in the GUI. A Qt based CDisplayEx in short.

    Read the article

  • Can't detect MIME type when using Zend Zend_File_Transfer_Adapter_Http !

    - by mrblue
    Hi all, I am writing a small script to upload and detect MIME type, using Zend Framework. For the design purpose, I can't use Zend_Form but normal instead. And I simply apply from the manual : $adapter = new Zend_File_Transfer_Adapter_Http(); $files = $adapter-getFileInfo(); $mime = $files-getMimeType(); But the system inform that the funcion getMimeType() does not existed. Then, I tried: $adapter = new Zend_File_Transfer(); $files = $adapter-getFileInfo(); $mime = $files-getMimeType(); This time, it didn't work either. So, how can I get the MIME type ? Thank you so much for your help

    Read the article

  • How to automate response to msgbox

    - by kensy
    I am developing a C# application to automate the running of a legacy VBScript(vbs) file which calls several VB6 .exe files. The .exe files have message box pop-ups that I need to 'respond' to in order to allow the VBScript process to run unattended. The response would need to be the Enter key. I do not have the source for the .exe files and I do not know exactly what they do. I would greatly appreciate any help with this...

    Read the article

  • Account for cancelling an update process

    - by Aerus
    I'm trying to develop the logic of updating my (Java) application but I'm stuck on a particular problem. The user can, at any time, be it on purpose (by clicking the Cancel button) or unwillingly (system/power failure) cancel the update process. If the updater was busy downloading files or unpacking files and thus didn't make any changes to the main application yet, this can be easily recovered by re-downloading the files and continuing the update process like normally. But, how can i account for changes made to the main application (i.e. a file was only partially replaced, some files were corrupted, etc.) that result in a main application that can't be launched anymore? Is it possible to rollback the changes made during the update without having to keep a full backup of my application or without doing a full reinstall? Are there any standards for updating an application that can help me with this?

    Read the article

  • Best directory to store application data with read\write rights for all users?

    - by Wodzu
    Hi guys. Until Windows Vista I was saving my application data into the directory where the program was located. The most common place was "C:\Program Files\MyApplication". As we know, under Vista and later the common user does't have rights to write under "Program Files" folder. So my first idea was to save the application data under "All Useres\Application Data" folder. But it seams that this folder has writing restrictions too! So to sum up, my requirements are: Folder should exist under Windows XP and above Microsoft's systems. All useres of the system should read\write\creation rights to this folder and it subfolders and files. I want to have only one copy of file\files for all useres. Thanks for your time.

    Read the article

  • How to ignore an exception and continue processing a foreach loop?

    - by Barry Dysert
    I have a test program that is supposed to loop over all the files under C:. It dies when it hits the "Documents and Settings" folder. I'd like to ignore the error and keep looping over the remaining files. The problem is that the exception is thrown in the foreach, so putting a try/catch around the foreach will cause the loop to exit. And putting a try/catch after the foreach never fires (because the exception is thrown in the foreach). Is there any way to ignore the exception and continue processing the loop? Here's the code: static void Main(string[] args) { IEnumerable<string> files = Directory.EnumerateFiles(@"C:\", "*.*", SearchOption.AllDirectories); foreach (string file in files) // Exception occurs when evaluating "file" Console.WriteLine(file); }

    Read the article

  • Using svn remove - preserve file on disk instead of having it deleted?

    - by user246114
    Hi, I am creating an eclipse project, which generates a lot of files I don't want to keep under version control. For example, it creates a project folder like: project/ bin/ horse.class cow.class src/ horse.java cow.java so what I do is add the project folder to svn, like: svn add project this puts everything under control, even the bin and .class files. All of those files will have the little 'a' next to them. How can I simply remove the 'add' status from those files? If I try: svn remove bin/horse.class I have to use the --force option and it deletes the file from disk. Is there a way to simply remove it from source but not delete the file from disk? Thanks

    Read the article

  • RubyMine Folder Tree doesn't refresh

    - by user336514
    Hi Using RubyMine 2.0.2 for the first time today, on Mac OSX Leopard. If I create files in the filesystem (with, say, script/generate) those new files do not appear in rubymine. I have had limited success with restarting the program, in that the files in the db/ folder are added and removed, but unfortunately new folders in the views directory are not shown. Pardon my french, but wtf? Thanks

    Read the article

  • While running the JSon application,I got three errors?

    - by Madan Mohan
    I added it from existing files JSON floder and also added the framework and import the "JSON/JSON.h". After that while running apps, It is giving three erros that is no such files, they are #import "SBJSON.h" #import "NSObject+SBJSON.h" #import "NSString+SBJSON.h" When i select framework, Two of the files containing in it Headers and PrivateHeaders is red in colour. Do I need to set the Header and Private path please help me. Thank You.

    Read the article

  • $HOME git repo (selectively) to github?

    - by user428502
    I keep many files in my home directory under git. Important dotfiles, my thesis, etc. I want to push certain files to github, e.g., my emacs configuration, to share. Obviously, I don't want to push the entire repo. Are submodules the way to go? My first thought is to make a directory ~/github/emacs, and rsync selective files here, then add a submodule under that directory, pointing to github, to push. Is this a good idea, or is there a better way? (I don't want my local git repo storing all files to get muddled up with this stuff, though.)

    Read the article

  • #Define's scope throughout library?

    - by Mohit Deshpande
    Say I have a constant: #define PI 3.14 Say I have a static library with multiple header and source files. If I declare this in the header file, will its scope apply to all of the source files? Or do the source files need to include the header with the declaration of PI?

    Read the article

  • R problem with apply + rbind

    - by Carl
    I cannot seem to get the following to work directory <- "./" files.15x16 <- c("15x16-70d.out", "15x16-71d.out") data.15x16<-rbind( lapply( as.array(paste(directory, files.15x16, sep="")), FUN=read.csv, sep=" ", header=F) ) What it should be doing is pretty straightforward - I have a directory name, some file names, and actual files of data. I paste the directory and file names together, read the data from the files in, and then rbind them all together into a single chunk of data. Except the result of the lapply has the data in [[]] - i.e., accessing it occurs via a[[1]], a[[2]], etc which rbind doesn't seem to accept. Suggestions?

    Read the article

  • bash find xargs grep only single occurence

    - by keftebub
    hi. maybe it's a bit strange - and maybe there are other tools to do this but, well.. i am using the following classic bash command to find all files which contain some string: find . -type f | xargs grep "something" i have a great number of files, on multiple depths. first occurence of "something" is enough for me, but find continues searching, and takes a long time to complete the rest of the files. what i would like to do is something like a "feedback" from grep back to find so that find could stop searching for more files. is such a thing possible? thank you

    Read the article

  • Automate downloads from password protected website

    - by Andrew
    I need some help with a work project I have been assigned. At the moment we manually go to the site, logon and then download 2 excel files from a supplier's website every month. The files are then loaded into SQL. We want to automate this process. Now the loading of the files into SQL I can do, but I am not sure how I can automate logging onto the website entering my user details and collecting the files. I mostly deal with SQL and have very little .NET experience, so any code samples would be most appreciated.

    Read the article

  • Document conversion and viewing, what are the cutting edge solutions?

    - by DigitalLawyer
    Goal: building a web application where a user can: Upload a document (doc, docx, pdf, additional office formats a +) View that document in a browser, preferably in html Download the document (in doc, pdf, additional open formats a +) Current solution: Ruby on Rails Application on Rackspace Users can upload doc and pdf files (AWS) Files can be downloaded in the format in which they were uploaded Thumbnail generation ([doc, pdf] - pdf - png) is done through AbiWord. Certain doc files do not convert well. Documents can be viewed in embedded Google docs viewer (https://docs.google.com/viewer). Certain doc files cannot be displayed. Little flexibility. Potential improvements: Document viewing in pdf through pdf.js Viewing in html (+ annotation) through Crocodoc I'd be glad to hear other users' experiences, and will add good recommendations to this list.

    Read the article

  • Programatically importing a file and creating an instance of a class in Python

    - by benofsky
    I have a (python) list of strings which refer to python source files and subsequently classes within those files which I want to import and then create an instance of the classes within the files (everything follows a strict naming convention, making this theoretically possible), in Ruby I would do something like: require "lib/sources/#{source}.rb" s = source.constantize.new How would I do something similar in Python? Thanks, Ben

    Read the article

  • Menu contribution for popupsubmenu ?

    - by Mathan
    In eclipse plugin development i could add files in toolbar, menu and popupmenu by using menu contribution. In my project I want to add a set of files in the popupsubmenu, eg : Like the following Project Explorer - Right Click - New - Annotation, Class, Enum .... I want to add my files abc,def and xyz.. under the new menu item What is the locationuri for popupsubmenu ? Help me on this Thanks in advance Regards Mathan

    Read the article

  • How to force tracker to re-index a folder?

    - by ubuntudroid
    I'm using the tracker indexing tool to search for and in files on my Ubuntu 10.10 amd64 system. Having recently added some of files into a single folder I wondered how to force tracker to re-index this folder so I would be able to perform search-queries on these files. Any ideas? A terminal command would be completely sufficient.

    Read the article

  • Most secure way to access my home Linux server while I am on the road? Specialized solution wanted

    - by Ace Paus
    I think many people may be in my situation. I travel on business with a laptop. And I need secure access to files from the office (which in my case is my home). The short version of my question: How can I make SSH/SFTP really secure when only one person needs to connect to the server from one laptop? In this situation, what special steps would make it almost impossible for anyone else to get online access to the server? A lot more details: I use Ubuntu Linux on both my laptop (KDE) and my home/office server. Connectivity is not a problem. I can tether to my phone's connection if needed. I need access to a large number of files (around 300 GB). I don't need all of them at once, but I don't know in advance which files I might need. These files contain confidential client info and personal info such as credit card numbers, so they must be secure. Given this, I don't want store all these files on Dropbox or Amazon AWS, or similar. I couldn't justify that cost anyway (Dropbox don't even publish prices for plans above 100 GB, and security is a concern). However, I am willing to spend some money on a proper solution. A VPN service, for example, might be part of the solution? Or other commercial services? I've heard about PogoPlug, but I don't know if there is a similar service that might address my security concerns? I could copy all my files to my laptop because it has the space. But then I have to sync between my home computer and my laptop and I found in the past that I'm not very good about doing this. And if my laptop is lost or stolen, my data would be on it. The laptop drive is an SSD and encryption solutions for SSD drives are not good. Therefore, it seems best to keep all my data on my Linux file server (which is safe at home). Is that a reasonable conclusion, or is anything connected to the Internet such a risk that I should just copy the data to the laptop (and maybe replace the SSD with an HDD, which reduces battery life and performance)? I view the risks of losing a laptop to be higher. I am not an obvious hacking target online. My home broadband is cable Internet, and it seems very reliable. So I want to know the best (reasonable) way to securely access my data (from my laptop) while on the road. I only need to access it from this one computer, although I may connect from either my phone's 3G/4G or via WiFi or some client's broadband, etc. So I won't know in advance which IP address I'll have. I am leaning toward a solution based on SSH and SFTP (or similar). SSH/SFTP would provided about all the functionality I anticipate needing. I would like to use SFTP and Dolphin to browse and download files. I'll use SSH and the terminal for anything else. My Linux file server is set up with OpenSSH. I think I have SSH relatively secured. I'm using Denyhosts too. But I want to go several steps further. I want to get the chances that anyone can get into my server as close to zero as possible while still allowing me to get access from the road. I'm not a sysadmin or programmer or real "superuser". I have to spend most of my time doing other things. I've heard about "port knocking" but I have never used it and I don't know how to implement it (although I'm willing to learn). I have already read a number of articles with titles such as: Top 20 OpenSSH Server Best Security Practices 20 Linux Server Hardening Security Tips Debian Linux Stop SSH User Hacking / Cracking Attacks with DenyHosts Software more... I have not implemented every single thing I've read about. I probably can't do that. But maybe there is something even better I can do in my situation because I only need access from a single laptop. I'm just one user. My server does not need to be accessible to the general public. Given all these facts, I'm hoping I can get some suggestions here that are within my capability to implement and that leverage these facts to create a great deal better security than general purpose suggestions in the articles above.

    Read the article

  • Google Gears hangs when capturing (unmanaged resource store)

    - by Gus
    Our code was written based on the example from Google Gears' own docs. We're using an unmanaged resource store. So we declare the files in an array, create the store, and capture all the files. Trouble is, the capturing process hangs. It always hangs on a random file (no discernible pattern has emerged), and when you reload the page, it always successfully captures. We're capturing 48 files. It seems to have nothing to do with the files themselves, as it hangs on every file type. I've seen it hang on the 6th file or the 47th. Windows and Mac. FF, IE, and Safari. We are not using a WorkerPool, and I'm thinking this may be necessary. Any other ideas why it would hang?

    Read the article

  • Capturing output of find . -print0 into a bash array

    - by Idris
    Using find . -print0 seems to be the only safe way of obtaining a list of files in bash due to the possibility of filenames containing spaces, newlines, quotation marks etc. However, I'm having a hard time actually making find's output useful within bash or with other command line utilities. The only way I have managed to make use of the output is by piping it to perl, and changing perl's IFS to null: find . -print0 | perl -e '$/="\0"; @files=<>; print $#files;' This example prints the number of files found, avoiding the danger of newlines in filenames corrupting the count, as would occur with: find . | wc -l As most command line programs do not support null-delimited input, I figure the best thing would be to capture the output of find . -print0 in a bash array, like I have done in the perl snippet above, and then continue with the task, whatever it may be. How can I do this? This doesn't work: find . -print0 | ( IFS=$'\0' ; array=( $( cat ) ) ; echo ${#array[@]} ) A much more general question might be: How can I do useful things with lists of files in bash?

    Read the article

< Previous Page | 397 398 399 400 401 402 403 404 405 406 407 408  | Next Page >