Search Results

Search found 24784 results on 992 pages for 'process integration packs'.

Page 749/992 | < Previous Page | 745 746 747 748 749 750 751 752 753 754 755 756  | Next Page >

  • How to use ccache selectively?

    - by Anonymous
    I have to compile multiple versions of an app written in C++ and I think to use ccache for speeding up the process. ccache howtos have examples which suggest to create symlinks named gcc, g++ etc and make sure they appear in PATH before the original gcc binaries, so ccache is used instead. So far so good, but I'd like to use ccache only when compiling this particular app, not always. Of course, I can write a shell script that will try to create these symlinks every time I want to compile the app and will delete them when the app is compiled. But this looks like filesystem abuse to me. Are there better ways to use ccache selectively, not always? For compilation of a single source code file, I could just manually call ccache instead of gcc and be done, but I have to deal with a complex app that uses an automated build system for multiple source code files.

    Read the article

  • Processing XML file with Huge data

    - by Manish Dhanotiya
    Hi,be m I am working on an application which has below requiements - 1. Download a ZIP file from a server. 2. Uncompress the ZIP file, get the content (which is in XML format) from this file into a String. 3. Pass this content into another method for parsing and further processing. Now, my concerns here is the XML file may be of Huge size say like '100MB', and my JVM has memory of only 512 MB, so how can I get this content into Chunks and pass for Parsing and then insert the data into PL/SQL tables. Since there can be multiple requests running at the same time and considering 512MB of memory what will be the best possible to process this. How I can get the data into Chunks and pass it as Stream for XML parsing. I googled on this, but didnt find any implementation. :( Thanks,

    Read the article

  • Visual C++ 9 Linker file size limitation.

    - by Raindog
    It appears that the visual C++ 9 linker has a file allocation algorithm that doubles the size of the file every allocation, so you get 512mb, 1024mb, 2048mb, 4096mb. The problem is that it is using a library that cannot handle files larger 2048MB, and as such crashes with an error such as "cannot read file at is the disk full or write protected". Is there a way to bypass this limitation or otherwise replace the linker with something else that works? A bit of background, I have a code generator that generates a large number of files, ~15k cpp files, I've managed to reduce the number of files to something about 6k to get something that at least completes the linking process, I would like to be able to include all 15k without having to create multiple libs.

    Read the article

  • What happens to an instance of ServerSocket blocked inside accept(), when I drop all references to i

    - by Hanno Fietz
    In a multithreaded Java application, I just tracked down a strange-looking bug, realizing that what seemed to be happening was this: one of my objects was storing a reference to an instance of ServerSocket on startup, one thread would, in its main loop in run(), call accept() on the socket while the socket was still waiting for a connection, another thread would try to restart the component under some conditions, the restart process missed the cleanup sequence before it reached the initialization sequence as a result, the reference to the socket was overwritten with a new instance, which then wasn't able to bind() anymore the socket which was blocking inside the accept() wasn't accessible anymore, leaving a complete shutdown and restart of the application as the only way to get rid of it. Which leaves me wondering: with no references left to the ServerSocket instance, what would free the socket for a new connection? At what point would the ServerSocket become garbage collected? In general, what are good practices I can follow to avoid this type of bug?

    Read the article

  • using APIs with oauth for single user

    - by Josh
    I'm trying to make use of various APIs including twitter, youtube, etc because we want to embed recent entries (tweets, videos) on our website. However, since I'm just retrieving my own data, I'm wondering how I can do this simpler than the multi-step process required by OAuth. Twitter provides me with my own access token I can use directly, so that kinda works, but I can't find any such token in the YouTube documentation. So how am I supposed to make use of the api if I just want to get a simple list of stuff? how exaclty am I supposed to authenticate my own website to use my own account? I think i might have things all wrong and if so please point me in the right direction. I tried using rss feeds but they don't give me as much control over what I retrieve as using the API directly... any insight or suggestions are appreciated!

    Read the article

  • Sending JSON content type in from JavaScript objects ( jQuery ajax )

    - by smartnut007
    I am having trouble sending JavaScript objects as a http request that only accepts json content-type. i.e "application/json" or "text/json" But, I am not sure why data2 ( stringified json ) works fine But, data1 ( json object ) does throws 400 ( Bad Request ). i.e I am not sure jQuery does not serialize the json object to a valid json string for the server to process. var data1 = ({ rating : 3 }); //does not work var data2 = '{ "rating" : 3}'; //works fine $.ajax({ url : "/rate", data : data1, type : "POST", contentType: "application/json", success: function(json){ console.log("Ajax Return :"+json); } });

    Read the article

  • Backbone.js: How to utilize router.navigate to manipulate browser history?

    - by Xavier_Ex
    I am writing something like a registration process containing several steps, and I want to make it a single-page like system so after some studying Backbone.js is my choice. Every time the user completes the current step they will click on a NEXT button I create and I use the router.navigate method to update the url, as well as loading the content of the next page and doing some fancy transition with javascript. Result is, URL is updated which the page is not refreshed, giving a smooth user experience. However, when the user clicks on the back button of the browser, the URL gets updated to that of a previous step, but the content stays the same. My question is through what way I can capture such an event and currently load the content of the previous step and present that to the user? Or even better, can I rely on browser cache to load that previously loaded page? EDIT: in particular, I'm trying something like mentioned in this article.

    Read the article

  • Dynamic Custom Fields for Data Model

    - by Jerry Deng
    I am in the process of creating a dynamic database where user will be able to create resource type where he/she can add custom fields (multiple texts, strings, and files) Each resource type will have the ability to display, import, export its data; I've been thinking about it and here are my approaches. I would love to hear what do you guys think. Ideas: just hashing all the custom data in a data field (pro: writing is easier, con: reading back out may be harder); children fields (the model will have multiple fields of strings, fields of text, and fields for file path); fixed number of custom fields in the same table with a key mapping data hash stored in the same row; Non-SQL approach, but then the problem would be generating/changing models on the fly to work with different custom fields;

    Read the article

  • YouTube app failes to open up when launched.

    - by user354738
    Below logs gives any clue ? The app just opens up (we could just see the tite of the app as YouTube and it goes back to launcher tray immd.) cat=[android.intent.category.LAUNCHER] flg=0x10200000 cmp=com.google.android.youtube/.HomeActivity } 05-31 14:06:41.655: INFO/ActivityManager(1102): Start proc com.google.android.youtube for activity com.google.android.youtube/.HomeActivity: pid=2977 uid=10041 gids={3003} 05-31 14:06:41.655: DEBUG/KeyguardViewMediator(1102): setHidden false 05-31 14:06:41.796: INFO/foo(1102): ********** resuming: 1144012408 05-31 14:06:41.897: INFO/ActivityThread(2977): Publishing provider com.google.android.youtube.SuggestionProvider: com.google.android.youtube.suggest.SuggestionProvider 05-31 14:06:41.960: INFO/YouTube(2977): Distribution channel:mvapp-android-google 05-31 14:06:42.116: INFO/AndroidRuntime(2977): AndroidRuntime onExit calling exit(-42) 05-31 14:06:42.116: INFO/dalvikvm(2977): In threadExitCheck 0xbdb0 05-31 14:06:42.116: ERROR/dalvikvm(2977): Native thread exited without telling us 05-31 14:06:42.116: ERROR/dalvikvm(2977): Native thread 2977 status: 7 05-31 14:06:42.124: DEBUG/KeyguardViewMediator(1102): setHidden false 05-31 14:06:42.139: INFO/ActivityManager(1102): Process com.google.android.youtube (pid 2977) has died

    Read the article

  • Import orders file on magento enterprise 1.7 (product with customs options)

    - by wil
    Hello, We need to import some orders file on magento enterprise. In our file, products contains customs Options. We tried to make an extension but we have some problems to import Customs options. The import of standard product is successful but not for the product with customs options. For customs option, missing "info_buyRequest" valu in database. The technical support of magento we responded "the import process currently can't handle importing products with custom options". Magento use custom options when ordering a product with customs options on website. What features do magento use to fill in the fields "info_buyRequest" and "Product_options" when ordering? Have you see a extension pack for import file order with product contains customs options? Thanks for your help.

    Read the article

  • File synck tool with network support (FTP or SSH) and decent GUI

    - by Álvaro G. Vicario
    Is there a decent GUI tool for Windows that allows to publish files from/to a remote FTP (or SFTP) server with the same ease of use that WinMerge offers in local discs? I'll basically use it to upload changes to web sites. I've tried like a dozen tools and they're all terrible. They have difficult interfaces, you need to spend ten minutes configuring absurdly complicate project settings, there's no simple way to ignore specific files or they use custom databases that break when someone else uploads files—not to mention deployment frameworks that force you to script everything in their custom language. I don't need or even want to schedule or automate. I want a manual process I can review. I'm tired of picking files one by one in Filezilla while reading the Subversion logs. I'd love an open source tool but...

    Read the article

  • C semaphores: sem_wait throwing inexplicable error

    - by tocapa
    I'm working on a problem which we have to use semaphores to solve. I have an array which contains two semaphores, gsem, and given certain conditions call sem_wait(&(gsem[me])), which is supposed to waiting until that particular process is woken up. However, for some reason it gives me the error Bad file descriptor. I looked up sem_wait and the Open Group spec says this is not an error sem_wait can cause. This is making my whole program crazy and I have no idea why this is failing.

    Read the article

  • How to detect an active x component that does not respond any more?

    - by koschi
    My application is written in C++ and uses the Qt framework. I use the QAxWidget class to access an active x component. Now I need some kind of mechanism that notifies my application each time the active x component has crashed or does not respond any more (due to dead lock or infinite loop). (1) can easily been done by watching the external process of the active x component. But maybe there is a more elegant approach? But how can (2) be implemented?

    Read the article

  • checkbox selected value is lost after coming back to the pagei.e., viewstate value is lost with use

    - by M.Vijay bhasker
    In the left navigation i have a series of check boxes and a grid view in the page body.Grid view results changes based on the check box selection,whenever i click on the links in grid view it goes to the second page and there is a link on the page which takes me back to the first page,but the check box selected value is maintained for the first time only and if we repeat the same process for the second time and i see that the first time selection of the check box is maintained.Here the grid view is placed in update panel and the check box in the left is called using .... AsyncPostBackTrigger ControlID="cphLeftNavigation". I've tried removing the above code referring the cphLeftNavigation and i see the results are coming up correctly and the last selected check box value is being maintained correctly and i want the same functionality to be implemented with/without update panel. Can anyone let me know,what exactly should i change to implement the above mentioned functionality.

    Read the article

  • How can I notify an application of a programmatically set scrollbar value?

    - by Frerich Raabe
    My code involves standard Scroll Bar control and it happens that I need to change its value programmatically in some cases. I do this using SetScrollInfo function, as in this example: void setScrollBarValue( HWND scrollBar, int value ) { SCROLLINFO si = { sizeof( SCROLLINFO ); } si.fMask = SIF_POS; si.nPos = value; ::SetScrollInfo( scrollBar, SB_CTL, &si, true /* redraw */ ); } This appears to work fine (the thumb of the scrollbar moves around) but it fails to notify the rest of the application of the new scrollbar value. For instance, an edit control which uses the scroll bar (much like in the Windows notepad application) fails to scroll around because it doesn't get notified about the new scrollbar value. In case it matters: the scrollbar I'm modifying is not in the same process as the above setScrollBarValue function. Does anybody know how to achieve this?

    Read the article

  • SVN - How to export changed files in dependency project when create updates?

    - by Relax
    I have a project with it depends on another project Core, i use svn:external to maintain the dependency. when i go to create an update of my project, i export those changed files of two tags of my project, but the changed files in Core won't be exported. For example, my project 2.0 depends on Core 1.0, where 3.0 depends on Core 2.0, then i create an update, changed files of my project 3.0 will all be exported, but none of those in Core 2.0 This can be done thru a two-steps process, first export changed files of the project 3.0, then the Core 2.0, it is ok if i can remember it everytime when i create an update, but in case i forget someday, the broken update will most probably cause a crash. I wonder, is there a way to cover this situation in TortoiseSVN, will let me do it in one operation? I use TortoiseSVN + VisualSVN

    Read the article

  • Add params before submit form ROR

    - by Jorge Najera T
    It's possible to add some parameters before submit an form? My problem is that I need to send the ticket id to my payment controller. A possibility is to send it through an hidden input field, but there's any other secure way to achieve this? Any help will be appreciated. Thanks. The process of buying a ticket 0) Select the event 1) User select the kind of ticket he wants to buy. 2) User add his personal information 3) Finally the Checkout (payment controller)

    Read the article

  • JQuery post to php

    - by RussP
    Why is it that I can never get JQuery serialize to work properly. I guess I must be missing something. I can serialize a form data and it shows in an alert: var forminfo = $j('#frmuserinfo').serialize(); alert(forminfo); I then post to my PHP page thus: $j.ajax({ type: "POST", url: "cv-user-process.php", data: "forminfo="+forminfo, cache: false, complete: function(data) { } }); But WHENEVER (not the first time) I try to insert/update the data in the DB I only ever get 1 varaible passed: Here is my PHP script: $testit = mysql_query("UPDATE cv_usersmeta SET inputtest='".$_POST['forminfo']."' WHERE user='X'"); the data passed only ever gets the first variable. why? I think it is more the way I deal with the php but it drives me nuts and always takes me far too long to find where I am going wrong.

    Read the article

  • Network ActiveRecord relation with Rails

    - by Zag zag..
    Hi, I have a has and belongs to many relation between User and Article models, and I would like to link them even if an article if not hosted on the same database then a user. For example, If an article exists at foo.com/articles/3 and a user exists at bar.com/users/1, If would like to be able to do from foo.com web interface or bar.com web interface this kind of query: a_user.articles (or an_article.users). I think this can be possible adding a field like "url" in users and articles tables. But I don't know how to process for ActiveRecord. My Article model looks like this: class Article < ActiveRecord::Base has_and_belongs_to_many :users end Is there yet some example of project using this kind of relation over internet? Many thanks

    Read the article

  • WCF web.config is getting overwritten after every compilation?

    - by AJ
    Hi I have a Silverlight application calling a WCF service. SimplehttpBinding stuff. Every I make changes to silverlight xaml code, the web.config gets refrshed also. Even if make any changes to web.cofig file, they get overwritten too. Its as if, some other process is writing these files. Why is that happening? How can I make sure that it does not get overwritten after every compile? Please advise. THanks AJ.

    Read the article

  • Millisecond-Accurate Scheduling of Future Events in C++/CLI

    - by A Grad Student at a University
    I need to create a C++/CLI mixed assembly that can schedule future calls into a native DLL with millisecond accuracy. This will, of course, mean setting a timer (what kind?) for a millisecond or three beforehand, then spinning until the moment and calling the native DLL function. Based on what I've read, I would guess that the callback that the timer calls will need to be native to make sure there are no thunks or GC to delay handling the timer callback. Will the entire thread or process need to be native and CLR-free, though, or can this be done just as accurately with #pragma unmanaged or setting one file of the assembly to compile as native? If so, how? If there is indeed no way to do this in mixed-mode C++/CLI, what would be the easiest way to set up an app/thread (ie, DLL or exe?) to handle it and to get the data back and forth between the native and managed threads/apps?

    Read the article

  • How to implement low pass filter using java

    - by chai
    Hello Everyone, I am trying to implement a low pass filter in Java. My requirement is very simple,I have to eliminate signals beyond a particular frequency (Single dimension). Looks like Butterworth filter would suit my need. Now the important thing is that CPU time should be as low as possible. There would be close to a million sample the filter would have to process and our users don't like waiting too long. Are there any readymade implementation of Butterworth filters which has optimal algorithms for filtering. Regards, Chaitannya

    Read the article

  • Storing task state between multiple django processes

    - by user366148
    I am building a logging-bridge between rabbitmq messages and Django application to store background task state in the database for further investigation/review, also to make it possible to re-publish tasks via the Django admin interface. I guess it's nothing fancy, just a standard Producer-Consumer pattern. Web application publishes to message queue and inserts initial task state into the database Consumer, which is a separate python process, handles the message and updates the task state depending on task output The problem is, some tasks are missing in the db and therefore never executed. I suspect it's because Consumer receives the message earlier than db commit is performed. So basically, returning from Model.save() doesn't mean the transaction has ended and the whole communication breaks. Is there any way I could fix this? Maybe some kind of post_transaction signal I could use? Thank you in advance.

    Read the article

  • Processing large recordsets in Rails

    - by japancheese
    Hello, I'm trying to perform a daily operation on a larger than normal dataset (2m+ records). However, Rails seems to take a very long time performing operations on such a dataset. Operations like Dataset.all.each do |data| ... end take a very long time to complete (I assume this is because it can't fit all the items into memory at once, right?). Does anyone have any strategies on how I could handle this situation? I know SQL would probably speed up the process, but I'm looking to use the Rails environment as I can do many more complicated things to the data than I can with just SQL statements.

    Read the article

  • Visual Studio 2010 debug minidump

    - by Snake
    Hi all, Consider the following code (written with Visual Studio 2010 and .NET 4.0) using System; namespace DumpTester { class Program { static void Main(string[] args) { int test = new Random().Next(); Console.WriteLine(test + new Random().Next()); Test(); } private static void Test() { throw new Exception(); } } } When running outside of Visual Studio you get this nice window of Microsoft Windows 7 that it is looking for a solution. Obviously, since this is my app, there is none. At that point I create a full dump file of my application with for example Process Explorer. I then open that dmp file from its location and try to debug. But whatever I try, it can't find the location of the source symbols. I tried putting the pdb next to the dump but it just doesn't find it. What am I doing wrong?

    Read the article

< Previous Page | 745 746 747 748 749 750 751 752 753 754 755 756  | Next Page >