Search Results

Search found 16731 results on 670 pages for 'memory limit'.

Page 403/670 | < Previous Page | 399 400 401 402 403 404 405 406 407 408 409 410  | Next Page >

  • AS3: why is this happening?

    - by Bin Chen
    Hi, I just encounter a strange problem: var a:ClassA = new ClassA; var b:ClassA = a; The program keeps running sometime, the a = null, b = null. The program is a complex one, I am sure that no part will touch a, and b. My question is, will the runtime(garbage collector) to collect the memory of "a" and then assign a and b to null? I am confused, thanks!

    Read the article

  • Python templates for huge HTML/XML

    - by newtover
    Hello, Recently I needed to generate a huge HTML page containing a report with several thousand row table. And, obviously, I did not want to build the whole HTML (or the underlying tree) in memory. As result, I built the page with the old good string interpolation, but I do not like the solution. Thus, I wonder whether there are Python templating engines that can yield resulting page content by parts.

    Read the article

  • Destructors in C++

    - by user265260
    does the Destructor deallocate memory assigned to the object which it belongs to or is it just called so that it can perform some last minute housekeeping before the object os deallocated by the compiler?

    Read the article

  • GitHub updating repository?

    - by user1804933
    I am trying to setup GitHub on my server and gotten to the point where I am running the command "git push -u origin master". However, a large file was detected and the following error was received: remote: error: GH001: Large files detected. remote: error: Trace: 5520a70fd2eeaa2eafd7de049a590fb5 remote: error: See http://git.io/iEPt8g for more information. remote: error: File app/logs/dev.log is 2041.59 MB; this exceeds GitHub's file size limit of 100 MB I ended up deleting that file and tried adding the git again but I keep running into that error. Any ideas on how to work around this?

    Read the article

  • Ubutu Server Edition: useful for the home user?

    - by D Connors
    My question is simple. What (if anything) does Ubuntu Server Edition have to offer to home users? This question is mostly out of curiosity really, but I like asking. I've got a home network set up here with some 6 to 7 machines (most Windows, one Linux). And I was wondering how useful would it be to have a dedicated computer in my house running Ubuntu Server full time. We've had an awful experience with file sharing so far, would it simplify file sharing/transferring? Would I be able to limit the internet bandwidth granted to each PC? Would I be able to monitor in/out internet traffic (both real time and monthly statistics)? Last, and most important, if I'm completely off as to what Ubuntu Server actually is, please say so. I am completely new to it.

    Read the article

  • Features to remove from C++

    - by Justin Ethier
    This question was inspired by What features would you like to see added to C++? (also see What features do you miss in C++?). C++ is a great general-purpose language, but perhaps too general and feature-rich: multiple inheritance, operator overloading, manual memory management, templates, smart pointers, virtual destructors, legacy frameworks (think MFC), and I could go on. Is there any one feature or aspect of C++ that you would like removed to make our lives easier as C++ developers? One feature per answer, please.

    Read the article

  • SWI-Prolog: how to load rdf triples using semweb/rdf_db library?

    - by Li Li
    Hi, I have a rdf file (file.trp) in n-triples format, where each line is: "subject predicate object ." I tried to use rdf_load in semweb/rdf_db to load it into memory, but failed. Here is what I tried: ?- rdf_load('file.trp'). ?- rdf_load('file.trp', [format(triples]). the manual says it supports xml and triples. But it only loads rdf xml files. How can I load such rdf triple file? Thanks, Li

    Read the article

  • Concurrent processes do not utilize all available CPU

    - by metdos
    I run some processes on an EC2 cc2.8xlarge instance which has 32 virtual processors. For some type of processes, when I run 16 processes on parallel, all of them use 100% of CPU cycles. But for other type of processes, they are not using 100% CPU and they finish considerably slower than a single thread. There is no time spend on IO and all data is served from memory. Do you have any idea about the reason of this problem?

    Read the article

  • modify class on runtime?

    - by joe doe2
    how can i modify the method in memory so it makes new objects with that modified method? for example i want to modify or add page_init handler on runtime for the next time the page loads? see this also for clarification: http://stackoverflow.com/questions/3005338/page-load-another-handler

    Read the article

  • Efficiently generate numpy array from list comprehension output?

    - by shootingstars
    Is there a more efficient way than using numpy.asarray() to generate an array from output in the form of a list? This appears to be copying everything in memory, which doesn't seem like it would be that efficient with very large arrays. (Updated) Example: import numpy as np a1 = np.array([1,2,3,4,5,6,7,8,9,10]) # pretend this has thousands of elements a2 = np.array([3,7,8]) results = np.asarray([np.amax(np.where(a1 > element)) for element in a2])

    Read the article

  • Processing large (over 1 Gig) files in PHP using stream_filter_*

    - by mike
    $fp_src=fopen('file','r'); $filter = stream_filter_prepend($fp_src, 'convert.iconv.ISO-8859-1/UTF-8'); while(fread($fp_src,4096)){ ++$count; if($count%1000==0) print ftell($fp_src)."\n"; } When I run this the script ends up consuming ~ 200 MB of RAM after going through just 35MB of the file. Running it without the stream_filter zips right through with a constant memory footprint of ~10 MB. What gives?

    Read the article

  • Facebook data collection ethical issues

    - by the_great_monkey
    If I have a Facebook app, and my users agree to allow my app to access their information, photos, friends, etc, is it ethical to grab their information when they log in, and then saving it in memory so that the next time he goes to my app, it can load faster? If so, what about when the user logged off? Is the right thing to do to is to delete all the cached information and photos that the user provided? Has Facebook got any way to detect that we're doing this (saving their information, etc)?

    Read the article

  • Is there any way to force an iPhone application to relaunch?

    - by Jared P
    Basically -[UIApplication openURL:] would be perfect, if it wasn't for the fact that it doesn't work within the same application (at least on simulator -- if this works on device someone please tell me). It's not a super big deal though if it just relaunches w/o URLness since I can obviously just persist some data. The goal of this is to essentially free all memory in the application and call all of the loading procedures.

    Read the article

  • Java execution details in System.out

    - by John
    As I remember there is a magic command line option in Java that turn on writing of operations that are currently executed to console. The output was looked like byte code. I do not mean -verbose, because it prints only class loading, while this option outputs information like memory allocation etc.

    Read the article

< Previous Page | 399 400 401 402 403 404 405 406 407 408 409 410  | Next Page >