Search Results

Search found 27655 results on 1107 pages for 'visual python'.

Page 621/1107 | < Previous Page | 617 618 619 620 621 622 623 624 625 626 627 628  | Next Page >

  • A Tkinter StringVar() Question

    - by Graham
    I would like to create a StringVar() that looks something like this: someText = "The Spanish Inquisition" #Here's a normal variable whose value I will change eventually TkEquivalent = StringVar() #and here's the StringVar() TkEquivalent.set(string(someText)) #and here I set it equal to the normal variable. When someText changes, this variable will too... HOWEVER: TkEquivalent.set("Nobody Expects " + string(someText)) If I do this, the StringVar() will no longer automatically update! How can I include that static text and still have the StringVar() update to reflect changes made to someText? Thanks for your help.

    Read the article

  • Django - count date between

    - by DJPy
    I have many record in my database wich contains datetime field (e.g. 2010-05-23 17:45:57). I want to count all records between e.g. 15:00 and 15:59 (it all can by from other day, month or year). How can I do this?

    Read the article

  • How do I set prifix to django contrib tables names?

    - by Software Enthusiastic
    HI One of my requirement is to have prefix on all the tables of the django based project (Because db is hosted on shared server). I have used db_table Meta option to set the prefix for the tables which I have created. Now my query is how do I set the prefix for tables provided by django.contrib. Instead of auth_group django should create prefix_auth_group.... How do I do that? Thank you very much...

    Read the article

  • For what purposes have YOU used T4?

    - by Chris Melinn
    T4 has existed for several years in Visual Studio, but doesn't get a lot of attention. However, for those that know it, there seems to be some very creative and useful purposes. I am researching some different ways that T4 is used, and I would appreciate to hear how YOU may have used it for real life scenarios. I am primarily interested in non-standard and creative uses. Some interesting examples: Phil Haack uses T4 to create static CSS files from .less To Generate WPF and Silverlight Dependency Properties using T4 Templates Note: I realize this is a discussion-oriented question, but the answers could be helpful to others. I have tagged it as subjective and also marked as "community wiki", so please allow the question to remain open. Thanks!

    Read the article

  • Pagination of Date-Based Generic Views in Django

    - by Apreche
    I have a pretty simple question. I want to make some date-based generic views on a Django site, but I also want to paginate them. According to the documentation the object_list view has page and paginate_by arguments, but the archive_month view does not. What's the "right" way to do it?

    Read the article

  • Live Updating Widget for 100+ concurrent users

    - by flavio87
    Hi there what would you use if you had to have a div box on your website that would have to be updated constantly with new HTML content from the server. simple polling is probably not very resource inefficient - imagine also having 10'000 users and the div has to update. what is the most efficient or elegant solution for such a problem? are there existing widgets which contain this "autoupdate" functionality?

    Read the article

  • Macro to create macros?

    - by JMarsch
    Over the years, I've built up a number of macros that I like to have available in visual studio. It's always a pain to reload them and rebind them to the keyboard when I go to a different machine/rebuild/use a VM/etc. Someone mentioned to me once that there is a way that you can write a macro that will recreate your macros and bind them to keys automatically. Anyone know how to do that? Is there another way to easily export/import macros (nonsensically, VS has an "export macro" function, but no import).

    Read the article

  • What does Postgres do when BEGIN is run on a connection in autocommit mode?

    - by DNS
    I'm trying to better understand the concept of 'autocommit' when working with a Postgres (psycopg) connection. Let's say I have a fresh connection, set its isolation level to ISOLATION_LEVEL_AUTOCOMMIT, then run this SQL directly, without using the cursor begin/rollback methods (as an exercise; not saying I actually want to do this): INSERT A INSERT B BEGIN INSERT C INSERT D ROLLBACK What happens to INSERTs C & D? Is autocommit is purely an internal setting in psycopg that affects how it issues BEGINs? In that case, the above SQL is unafected; INSERTs A & B are committed as soon as they're done, while C & D are run in a transaction and rolled back. What isolation level is that transaction run under? Or is autocommit a real setting on the connection itself? In that case, how does it affect the handling of BEGIN? Is it ignored, or does it override the autocommit setting to actually start a transaction? What isolation level is that transaction run under? Or am I completely off-target?

    Read the article

  • Fastest method in merging of the two: dicts vs lists

    - by tipu
    I'm doing some indexing and memory is sufficient but CPU isn't. So I have one huge dictionary and then a smaller dictionary I'm merging into the bigger one: big_dict = {"the" : {"1" : 1, "2" : 1, "3" : 1, "4" : 1, "5" : 1}} smaller_dict = {"the" : {"6" : 1, "7" : 1}} #after merging resulting_dict = {"the" : {"1" : 1, "2" : 1, "3" : 1, "4" : 1, "5" : 1, "6" : 1, "7" : 1}} My question is for the values in both dicts, should I use a dict (as displayed above) or list (as displayed below) when my priority is to use as much memory as possible to gain the most out of my CPU? For clarification, using a list would look like: big_dict = {"the" : [1, 2, 3, 4, 5]} smaller_dict = {"the" : [6,7]} #after merging resulting_dict = {"the" : [1, 2, 3, 4, 5, 6, 7]} Side note: The reason I'm using a dict nested into a dict rather than a set nested in a dict is because JSON won't let me do json.dumps because a set isn't key/value pairs, it's (as far as the JSON library is concerned) {"a", "series", "of", "keys"} Also, after choosing between using dict to a list, how would I go about implementing the most efficient, in terms of CPU, method of merging them? I appreciate the help.

    Read the article

  • how to trigger a script located on a machine in one domain from a machine on another domain

    - by user326814
    Hi, I am basically from QA. What we testers do each day is 1. Open a web browser. Type in http://11.12.13.27.8080/cruisecontrol (since we are in a particular network, only we can access this) 2. Check if the latest nightly build has been successful. If it is successful, deploy it on a test environment by clicking on 'Deploy this build' link. This deploying takes around 1-1.5 hours. During this time we cannot use our machines to work on anything else. Only after this deploying can we begin to test. Now, i wanted to know if its possible to do the below. When at home in the morning, i use something which will trigger a script (which will be on my machine at workplace). This script will inturn automatically deploy the build. I already have such a similar script. What i want to know is how is it possible to trigger this script from my home machine? Is it even possible? For e.g the external trigger will say "Deploy xxx branch on yyy test environment". So the script on my workplace machine will be invoked and it will automatically deploy it before i actually come to my desk. Please help. I am from QA and have no idea about all this.

    Read the article

  • recursive cumulative sums

    - by user1816377
    I need to write a program that compute cumulative sums from a list of numbers with def but ONLY with recursion. I did it, but now I need to write the same program without using the method sum, but no success so far. Any idea? my code: def rec_cumsum(numbers): ''' Input: numbers - a list of numbers, Output: a list of cumulative sums of the numbers''' if len(numbers)==0: return numbers return rec_cumsum(numbers[:-1])+ [sum(numbers)] input: 1 [1,2,3] 2 [2, 2, 2, 3] output: 1 [1,3,6] 2 [2, 4, 6, 9]

    Read the article

  • add new records using signal in django admin

    - by ganesh
    I've a model called broadcastinfo, It has fields viz.. info,userid...userid is excluded. when i add an new info, my broadcastinfo table should get the records of all userid from user table and the given message. Im trying this via signal.Any idea is highly appreciated. Thanks

    Read the article

  • TypeError: object not callable when making instance

    - by TSM
    I've searched around other threads with similar questions, but I'm not finding the answer. Basically, I have a class: import Android_Class class Android_Revision(object): def __init__(self): # dict for storing the classes in this revision # (format {name : classObject}): self.Classes = {} self.WorkingClass = Android_Class() self.RevisionNumber = '' def __call__(self): print "Called" def make_Class(self, name): newClass = Android_Class(name) self.Classes.update({name : newClass}) self.WorkingClass = newClass def set_Class(self, name): if not(self.Classes.has_key(name)): newClass = Android_Class(name) self.Classes.update({name : newClass}) self.WorkingClass = self.Classes.get(name) I'm trying to make an instance of this class: Revision = Android_Revision() and that's when I'm getting the error. I'm confused because I have another situation where I'm doing almost the exact same thing, and it's working fine. I can't figure out what differences between the two would lead to this error. Thanks.

    Read the article

  • VS2010 patch: why it's take so much time to install it? [closed]

    - by Mendy
    Visual Studio 2010 RC has a few of patches release. For more information about them take a look here. What I'm expect from patch program, is to replace a few dll's of the program to a new fixed version of them. But when I run each of this 3 patches, they take a lot of time (5 minutes each), and you think that the program was frozen because the progress bar stay on the begging. This is question may not be so important, but it really interesting me to know, why this happens? It's really confusing to see that each VS2010 (or Microsoft in general) is frozen to 4-5 minutes.

    Read the article

  • issue with c# xml documentation

    - by galford13x
    I have the following comment. /// <summary> /// MSDN Time Format compatible with <see cref="DateTime.ParseExact(string, string, IFormatProvider)"/> /// </summary> /// <returns>MSDN Time Format compatible with <see cref="DateTime.ParseExact(string, string, IFormatProvider)"/></returns> but I'm not sure why I receive the following warning Warning 7 XML comment on 'MSLab.DateTime.SystemTimeProvider.GetTimeFormat()' has cref attribute 'DateTime.ParseExact(string, string, IFormatProvider)' that could not be resolved F:\My Documents\Visual Studio 2010\Projects\MSLab\trunk\MSLab\MSLab\DateTime\SystemTimeProvider.cs 110 57 MSLab

    Read the article

  • How to verify object creation in Django ?

    - by Martin
    So.. this never crossed my head before but now I just can't figure out how to do that !! I want to verify that the object I created was really created, and return True or False according to that : obj = object(name='plop') try: obj.save() return True except ???: return False Any idea ? Cheers, -M

    Read the article

  • Obtain distances to nearest landmarks (Mall, Hospital, and Airport etc.) using google map

    - by user227290
    I am working on a project in which I have around 100000 addreses in major cities in India(it is a table in a database). I want to know if it is possible to obtain the distances to to nearest landmarks (Mall, Hospital, and Airport etc.). Ideally I want these distances to be mergeed to the parent table. We have Java and Php coders to get it done once we find out how to go about it. Any pointers will be of great help. Thank you.

    Read the article

  • Efficient job progress update in web application

    - by Endru6
    Hi, Creating a web application (Django in my case, but I think the question is more general) that is administrating a cluster of workers doing queued jobs, there is a need to track each jobs progress. When I've done it using database UPDATE (PostgreSQL in this case), it severely hits the database performance, because each UPDATE creates a new row in a table, and in my case only vacuuming DB removes obsolete rows. Having 30 jobs running and reporting progress every 1 minute DB may require vacuuming (and it means huge slow downs on a front end side for all the employees working with the system) every 10 days. Because the progress information isn't critical, ie. it doesn't have to be persistent, how would you do the progress updates from jobs without using an overhead database implies? There are 30 worker servers, each doing 1 or 2 jobs simultaneously, 1 front end server which serves a web application to users, and 1 database server.

    Read the article

  • How do you extend the Site model in django?

    - by John Giotta
    What is the best approach to extending the Site model in django? Creating a new model and ForeignKey the Site or there another approach that allows me to subclass the Site model? I prefer subclassing, because relationally I'm more comfortable, but I'm concerned for the impact it will have with the built-in Admin.

    Read the article

  • Python2.7: How can I speed up this bit of code (loop/lists/tuple optimization)?

    - by user89
    I repeat the following idiom again and again. I read from a large file (sometimes, up to 1.2 million records!) and store the output into an SQLite databse. Putting stuff into the SQLite DB seems to be fairly fast. def readerFunction(recordSize, recordFormat, connection, outputDirectory, outputFile, numObjects): insertString = "insert into NODE_DISP_INFO(node, analysis, timeStep, H1_translation, H2_translation, V_translation, H1_rotation, H2_rotation, V_rotation) values (?, ?, ?, ?, ?, ?, ?, ?, ?)" analysisNumber = int(outputPath[-3:]) outputFileObject = open(os.path.join(outputDirectory, outputFile), "rb") outputFileObject, numberOfRecordsInFileObject = determineNumberOfRecordsInFileObjectGivenRecordSize(recordSize, outputFileObject) numberOfRecordsPerObject = (numberOfRecordsInFileObject//numberOfObjects) loop1StartTime = time.time() for i in range(numberOfRecordsPerObject ): processedRecords = [] loop2StartTime = time.time() for j in range(numberOfObjects): fout = outputFileObject .read(recordSize) processedRecords.append(tuple([j+1, analysisNumber, i] + [x for x in list(struct.unpack(recordFormat, fout))])) loop2EndTime = time.time() print "Time taken to finish loop2: {}".format(loop2EndTime-loop2StartTime) dbInsertStartTime = time.time() connection.executemany(insertString, processedRecords) dbInsertEndTime = time.time() loop1EndTime = time.time() print "Time taken to finish loop1: {}".format(loop1EndTime-loop1StartTime) outputFileObject.close() print "Finished reading output file for analysis {}...".format(analysisNumber) When I run the code, it seems that "loop 2" and "inserting into the database" is where most execution time is spent. Average "loop 2" time is 0.003s, but it is run up to 50,000 times, in some analyses. The time spent putting stuff into the database is about the same: 0.004s. Currently, I am inserting into the database every time after loop2 finishes so that I don't have to deal with running out RAM. What could I do to speed up "loop 2"?

    Read the article

< Previous Page | 617 618 619 620 621 622 623 624 625 626 627 628  | Next Page >