Search Results

Search found 27238 results on 1090 pages for 'local variable'.

Page 805/1090 | < Previous Page | 801 802 803 804 805 806 807 808 809 810 811 812  | Next Page >

  • Combo box items tool tip not displayed

    - by SriniWeb
    Hello, I have written a combo box populated with items in ExtJs framework. I could not able to get the tooltips when the user hovers over the items. Below is the code: $formPanel .= " new Ext.form.ComboBox({ tpl:'<tpl for =\".\"><div ext:qtip=\"{abbr}\" class=\"x-combo-list-item\">{state}</div></tpl>', id:'type_blackout_weekly', store: sstore, displayField:'state', typeAhead: true, mode: 'local', triggerAction: 'all', emptyText:'Select a state...', selectOnFocus:true, forceSelection: true, width:135 }),"; var statedata = [ ['AL', 'Alabama'], ['AK', 'Alaska'], ['AZ', 'Arizona'], ['WV', 'West Virginia'], ['WI', 'Wisconsin'], ['WY', 'Wyoming'] ]; var sstore = new Ext.data.SimpleStore({ fields: ['abbr', 'state'], data : statedata }); Any guidance is highly appreciated.

    Read the article

  • C / JSON Library in popular Linux distros?

    - by Tim Post
    I have a program written in C that has to input and output JSON over a local domain socket. I've found several C / JSON libraries that 'almost work' through searches. Prior to taking one of the libraries that I found .. I want to be sure that I'm not over-looking a library that is commonly found on modern Linux distros. I'd also really appreciate links to libraries that you use. Most likely, I'll just drop it in tree, unless I realize that I've over looked something widely distributed. I am tagging this as subjective because the answer that I select is the one linking to a library that works for me, that does not mean its the 'best' library. I want to take an existing array and easily convert it to a buffer that can be sent, or take a buffer and easily convert it into an allocated array. Thanks in advance!

    Read the article

  • WPF - Get combobox checked property from ListBox

    - by Chris Klepeis
    I have a listbox, which is defined like so: <ListBox ItemsSource="{Binding Source={x:Static local:ResourceCollection.resourceList}}" Height="143" HorizontalAlignment="Left" Margin="6,6,0,0" Name="assignmentLB" VerticalAlignment="Top" Width="287" FontSize="12" FontWeight="Normal" IsEnabled="True" Grid.Column="0"> <ListBox.ItemTemplate> <DataTemplate> <StackPanel Orientation="Horizontal"> <CheckBox /> <TextBlock Text="{Binding Content}" /> </StackPanel> </DataTemplate> </ListBox.ItemTemplate> </ListBox> How can I loop through this listbox and retrieve the TextBlock.Text value for only items whose Checkbox has been checked?

    Read the article

  • MinGW screw up with COLORREF and RGB

    - by kjoppy
    I am trying to build a 3rd party open source project using MinGW. One of the dependencies is wxWidgets. When I try to make the project from MSYS I get a compiler error from /MinGW/msys/1.0/local/include/wx-2.8/wx/msw/private.h In function 'COLORREF wxColourToRGB(const wxColour&)': error: cannot convert 'RGB' to 'COLORREF {aka long unsigned in}' in return This is somewhat odd given that, according to Microsoft the RGB macro returns a COLORREF. In fact, looking in H:\MinGW\include I find wingdi.h with the following code #define RGB(r,g,b) ((COLORREF)((BYTE)(r)|((BYTE)(g) << 8)|((BYTE)(b) << 16))) What sort of thing would cause this error? Is there some way I can check to see if COLORREF and RGB are being included from wingdi.h and not somewhere else? Is that even worth checking? Specifications GCC version 4.7.2 wxWidgets version 2.8.12 (I'm new to C++ and MinGW specifically but generally computer and programming literate)

    Read the article

  • Accept-Encoding headers being sent by browser but not received by server

    - by Daniel Jacobs
    I have been trying to debug this for weeks. All of the browsers on all of the clients on my home network are sending 'Accept-Encoding: gzip,deflate'. However, that header is somehow, somewhere being dropped before the request makes it to a web server. For example, http://www.whatsmyip.org/http_compression/ says 'No, your browser is not requesting compressed content'. I've used Fiddler to make sure that all of my browsers are indeed sending the header. I've swapped out my router. I've turned off all anti-virus software. Brighthouse/Roadrunner (the local cable ISP) says they are not doing any filtering (and I can't see why they would in this case). Any suggestions would be most welcome!

    Read the article

  • How to embed a font in website

    - by Mayur
    Hi All, I am trying to embed my custom font in my web site, got a link folder from http://www.fontsquirrel.com/fontface/generator this site after uploading a font on this site, it gives me a css @font-face { font-family: 'VoltaEFTU-Regular'; src: url('voltaeftu-regular-webfont.eot'); src: local('?'), url('voltaeftu-regular-webfont.woff') format('woff'), url('voltaeftu-regular-webfont.ttf') format('truetype'), url('voltaeftu-regular-webfont.svg#webfonttKmU3jX8') format('svg'); font-weight: normal; font-style: normal; } but its not working in my machines how can i embed a font in my web site.... Thanks Mayur Mate

    Read the article

  • Optimizing python code performance when importing zipped csv to a mongo collection

    - by mark
    I need to import a zipped csv into a mongo collection, but there is a catch - every record contains a timestamp in Pacific Time, which must be converted to the local time corresponding to the (longitude,latitude) pair found in the same record. The code looks like so: def read_csv_zip(path, timezones): with ZipFile(path) as z, z.open(z.namelist()[0]) as input: csv_rows = csv.reader(input) header = csv_rows.next() check,converters = get_aux_stuff(header) for csv_row in csv_rows: if check(csv_row): row = { converter[0]:converter[1](value) for converter, value in zip(converters, csv_row) if allow_field(converter) } ts = row['ts'] lng, lat = row['loc'] found_tz_entry = timezones.find_one(SON({'loc': {'$within': {'$box': [[lng-tz_lookup_radius, lat-tz_lookup_radius],[lng+tz_lookup_radius, lat+tz_lookup_radius]]}}})) if found_tz_entry: tz_name = found_tz_entry['tz'] local_ts = ts.astimezone(timezone(tz_name)).replace(tzinfo=None) row['tz'] = tz_name else: local_ts = (ts.astimezone(utc) + timedelta(hours = int(lng/15))).replace(tzinfo = None) row['local_ts'] = local_ts yield row def insert_documents(collection, source, batch_size): while True: items = list(itertools.islice(source, batch_size)) if len(items) == 0: break; try: collection.insert(items) except: for item in items: try: collection.insert(item) except Exception as exc: print("Failed to insert record {0} - {1}".format(item['_id'], exc)) def main(zip_path): with Connection() as connection: data = connection.mydb.data timezones = connection.timezones.data insert_documents(data, read_csv_zip(zip_path, timezones), 1000) The code proceeds as follows: Every record read from the csv is checked and converted to a dictionary, where some fields may be skipped, some titles be renamed (from those appearing in the csv header), some values may be converted (to datetime, to integers, to floats. etc ...) For each record read from the csv, a lookup is made into the timezones collection to map the record location to the respective time zone. If the mapping is successful - that timezone is used to convert the record timestamp (pacific time) to the respective local timestamp. If no mapping is found - a rough approximation is calculated. The timezones collection is appropriately indexed, of course - calling explain() confirms it. The process is slow. Naturally, having to query the timezones collection for every record kills the performance. I am looking for advises on how to improve it. Thanks. EDIT The timezones collection contains 8176040 records, each containing four values: > db.data.findOne() { "_id" : 3038814, "loc" : [ 1.48333, 42.5 ], "tz" : "Europe/Andorra" } EDIT2 OK, I have compiled a release build of http://toblerity.github.com/rtree/ and configured the rtree package. Then I have created an rtree dat/idx pair of files corresponding to my timezones collection. So, instead of calling collection.find_one I call index.intersection. Surprisingly, not only there is no improvement, but it works even more slowly now! May be rtree could be fine tuned to load the entire dat/idx pair into RAM (704M), but I do not know how to do it. Until then, it is not an alternative. In general, I think the solution should involve parallelization of the task. EDIT3 Profile output when using collection.find_one: >>> p.sort_stats('cumulative').print_stats(10) Tue Apr 10 14:28:39 2012 ImportDataIntoMongo.profile 64549590 function calls (64549180 primitive calls) in 1231.257 seconds Ordered by: cumulative time List reduced from 730 to 10 due to restriction <10> ncalls tottime percall cumtime percall filename:lineno(function) 1 0.012 0.012 1231.257 1231.257 ImportDataIntoMongo.py:1(<module>) 1 0.001 0.001 1230.959 1230.959 ImportDataIntoMongo.py:187(main) 1 853.558 853.558 853.558 853.558 {raw_input} 1 0.598 0.598 370.510 370.510 ImportDataIntoMongo.py:165(insert_documents) 343407 9.965 0.000 359.034 0.001 ImportDataIntoMongo.py:137(read_csv_zip) 343408 2.927 0.000 287.035 0.001 c:\python27\lib\site-packages\pymongo\collection.py:489(find_one) 343408 1.842 0.000 274.803 0.001 c:\python27\lib\site-packages\pymongo\cursor.py:699(next) 343408 2.542 0.000 271.212 0.001 c:\python27\lib\site-packages\pymongo\cursor.py:644(_refresh) 343408 4.512 0.000 253.673 0.001 c:\python27\lib\site-packages\pymongo\cursor.py:605(__send_message) 343408 0.971 0.000 242.078 0.001 c:\python27\lib\site-packages\pymongo\connection.py:871(_send_message_with_response) Profile output when using index.intersection: >>> p.sort_stats('cumulative').print_stats(10) Wed Apr 11 16:21:31 2012 ImportDataIntoMongo.profile 41542960 function calls (41542536 primitive calls) in 2889.164 seconds Ordered by: cumulative time List reduced from 778 to 10 due to restriction <10> ncalls tottime percall cumtime percall filename:lineno(function) 1 0.028 0.028 2889.164 2889.164 ImportDataIntoMongo.py:1(<module>) 1 0.017 0.017 2888.679 2888.679 ImportDataIntoMongo.py:202(main) 1 2365.526 2365.526 2365.526 2365.526 {raw_input} 1 0.766 0.766 502.817 502.817 ImportDataIntoMongo.py:180(insert_documents) 343407 9.147 0.000 491.433 0.001 ImportDataIntoMongo.py:152(read_csv_zip) 343406 0.571 0.000 391.394 0.001 c:\python27\lib\site-packages\rtree-0.7.0-py2.7.egg\rtree\index.py:384(intersection) 343406 379.957 0.001 390.824 0.001 c:\python27\lib\site-packages\rtree-0.7.0-py2.7.egg\rtree\index.py:435(_intersection_obj) 686513 22.616 0.000 38.705 0.000 c:\python27\lib\site-packages\rtree-0.7.0-py2.7.egg\rtree\index.py:451(_get_objects) 343406 6.134 0.000 33.326 0.000 ImportDataIntoMongo.py:162(<dictcomp>) 346 0.396 0.001 30.665 0.089 c:\python27\lib\site-packages\pymongo\collection.py:240(insert) EDIT4 I have parallelized the code, but the results are still not very encouraging. I am convinced it could be done better. See my own answer to this question for details.

    Read the article

  • The system cannot find the file specified.

    - by Sathish
    Till yesterday my webservice was running fine in my local system today when i run the Webservcie or any webproject from VS2005 i get the below error Server Error in '/MyWebService' Application. The system cannot find the file specified. (Exception from HRESULT: 0x80070002) Description: An unhandled exception occurred during the execution of the current web request. Please review the stack trace for more information about the error and where it originated in the code. Exception Details: System.IO.FileNotFoundException: The system cannot find the file specified. (Exception from HRESULT: 0x80070002) Please help me

    Read the article

  • automatic push to CDN deployment strategy

    - by imanc
    Does anyone have ideas for a strategy to push content to a CDN upon deployment? The key issue I'm facing is that we have a site that is available in various contexts: local development, development server, staging, then finally live. The liver version of the site needs to load assets from a domain, which will be pointed to a CDN: assets.domain.com. However, we will have numerous references to the assets pointing to a relative folder, e.g. /images/ in css, possibly in js, and in HTML & source. Our new site will use capistrano for deployment and it may be that we can hook in another build tool (apache ant?) or some custom script to search / replace paths. I am wondering if anyone has had to deal with this issue before and what solutions you put in place to automate managing the CDN in terms of pushing content up to the CDN and managing html & css references to assets in the CDN. Thanks Imanc

    Read the article

  • problem with date returning wrong day although the timestamp is correct!

    - by Spiros
    I have a bizzare problem with php date function. code: $numDays = 8; $date = strtotime('2010-11-06'); for ($i=1; $i<=$numDays; $i++) { $thisDay = date("D, d M Y", $date); print ($thisDay.'<br>'); $date+=86400; // add one day to timestamp } result on my server (local host, windows): Sat, 06 Nov 2010 Sun, 07 Nov 2010 Mon, 08 Nov 2010 Tue, 09 Nov 2010 Wed, 10 Nov 2010 Thu, 11 Nov 2010 Fri, 12 Nov 2010 Sat, 13 Nov 2010 Result on my web server (linux) Sat, 06 Nov 2010 *Sun, 07 Nov 2010 Sun, 07 Nov 2010* Mon, 08 Nov 2010 Tue, 09 Nov 2010 Wed, 10 Nov 2010 Thu, 11 Nov 2010 Fri, 12 Nov 2010 Notice how Sun, 07 Nov 2010 appears twice on the remote server?? Why is this happening? can anyone explain this Behavior?

    Read the article

  • Tomcat myfaces dynamic source files

    - by Gerrie
    Hi, I'm going to try and make as much sense as possible here. We are working on a myfaces 2 app in tomcat. On our local dev machines, whenever we want to change one of our xhtml files, we have to stop the server, deploy and then start the server back up. This is cumbersome and makes making small changes to the view tedious. I tried changing the deployed xhtml file in tomcat, but the server only picks up the new change after a restart. Is there any type of config in myfaces or tomcat I can do to force the rebuilding of these source xhtml's every time?

    Read the article

  • How can I make sure the Sphinx daemon runs?

    - by Ethan
    I'm working on setting up a production server using CentOS 5.3, Apache, and Phusion Passenger (mod_rails). I have an app that uses the Sphinx search engine and the Thinking Sphinx gem. According to the Thinking Sphinx docs... If you actually want to search against the indexed data, then you’ll need Sphinx’s searchd daemon to be running. This can be controlled using the following tasks: rake thinking_sphinx:start rake ts:start rake thinking_sphinx:stop rake ts:stop What would be the best way to ensure that this takes place in production? I can deploy my app, then manually run rake thinking_sphinx:start, but I like to set things up so that if I have to bounce the server, everything will come back up. Should I put a call to that Rake task in an initializer? Or something in rc.local?

    Read the article

  • Is it possible to prevent a locally-running SWF (AS3) from downloading from my website?

    - by Matt
    I've got a crossdomain.xml file which allows SWFs running on only a certain few domains to download resources from my domain. However, one simple way around this is for a user to download the SWF to their local machine, and run it there (i.e. by double-clicking on it within Windows Explorer, not by running through http://localhost). It seems that when this happens, the crossdomain.xml file is ignored. I understand that in my actionscript, I can do this: if (Security.sandboxType.indexOf(Security.REMOTE) == -1) // running locally - don't allow However it is incredibly easy for someone to decompile the SWF and simply remove this line. Is it possible to do something on the server side to stop a locally running SWF to download from my site? I tried checking the referrer but this field often isn't populated. Does anyone have any other ideas? Thanks, Matt

    Read the article

  • Mirror a website with httrack while executing javascript

    - by Martin
    I want do save a mirror of www.youtube.com/tv. I obviously do not want to save the videos. I want the code running the website in a local copy, everything else can stay remote. The code I want is mainly contained in 2 files: live.js and app-prod.js. I tried using httrack. I have issue parsing the javascript to load anything past the first file: live.js. The %P parameter does not help. httrack www.youtube.com/tv +* -r6 --mirror -%P -j It doesn't go further than live.js because some javascript needs to be executed to load the next file. I know I can do this manually with any browser. I want to automate the process. Is httrack able to do this by itself? If yes, how?

    Read the article

  • Working with multiple GIT severs

    - by th3flyboy
    Hello, I have a question. Is it possible to set up a system so that you have a private GIT server that you host, which automatically syncs with a remote one, hosted by a site like Sourceforge, and then you can commit your local to the private GIT server, and then when you have to merge the changes from your private wip branches that are on your private GIT over to the master/branch/tag from the public GIT, and then push the change to the public GIT? I ask this because I have a lot of personal work I would like to get working before putting it up for the public to see, and I'm shifting between several computers/operating systems in the process. If this is not possible in standard GIT, are there any other options that would allow me to do this? Thanks, Peter

    Read the article

  • Coldfusion returning typed objects / AMF remoting

    - by Chin
    Is the same possible in ColdFusion? Currently I am using .Net/Fluorine to return objects to the client. Whilst in testing I like to pass strings representing the select statement and the custom object I wish to have returned from my service. Fluorine has a class ASObject to which you can set the var 'typeName'; which works great. I am hoping that this is possible in Coldfusion. Does anyone know whether you can set the type of the returned object in a similar way. This is especially helpful with large collections as the flash player will convert them to a local object of the same name thus saving interating over the collection to convert the objects to a particular custom object. foreach (DataRow row in ds.Tables[0].Rows) { ASObject obj = new ASObject(); foreach (DataColumn col in ds.Tables[0].Columns) { obj.Add(col.ColumnName, row[col.ColumnName]); } obj.TypeName = pObjType; al.Add(obj); } Many thanks,

    Read the article

  • SQL Server mirroring connection doesnt work

    - by StNickolas
    I have 2 servers srv-erp1 and srv-erp3. I made them mirror on each other. All setup is done by lots of tutorials and examples. But when I call ALTER DATABASE MIRROR_TEST SET PARTNER = 'TCP://srv-erp3:5022' It`s response is: The server network address "TCP://srv-erp3:5022" can not be reached or does not exist. Check the network address name and that the ports for the local and remote endpoints are operational. I go to cmd on srv-erp3 and use netstat -an... this port is listening. I go to cmd on srv-erp1 and use telnet srv-erp3 5022...and its ok to connect. All firewalls are turned off. The only difference in config of srvrs is that srv-erp1 is on Windows Server 2003 R2 x64, and srv-erp3 is on Windows 2008 R2 x64 What can be the reason of this problem? Regards, Dmitry.

    Read the article

  • Git Submodule or fork

    - by Eric
    I have a private repo in github that is the complete source code to my cms. Now I have a few local customers that I want to use the same code base on but with different themes. Is it better to fork the original project out into a repo for each one. Or use a submodule and create a new repo for each customer? After each site is complete I would imagine the theme files wouldn't change much but would need to pull in changes from the main repo when bugs are discovered.

    Read the article

  • PHP/HTML Strange refresh problem

    - by user270797
    After I upload my PHP files to my web host, I view the page by the URL. So what I usually do as I work on the page is make a change, upload, then refresh the browser to view my changes. What I find is that sometimes I refresh and it shows me a previous version of the page, I hit refresh 5 times and it shows me 5 different versions of the page some old changes some new changes, it makes it really difficult to know which is the latest version. I dont think this is a local caching problem, Ive disabled caching in all browsers, the problem occurs in IE, FF, Chrome I have a feeling it might be the web server. I believe the hosting company uses Zeus (?) web server product.

    Read the article

  • Test Jboss JMS externally - outside of the application server

    - by bmatsliah
    Hi All, I'm pretty new to the JMS functionality. I need to test a remote JMS which is on Jboss 4.2.1. I want to develop an external test (e.g from a main class) which sends and receives messages to/from the app. server. (I do have full access to the remote server.) My questions are: 1) How do I send messages to the Jboss JMS? 2) Where is the information needed to send a message located on the Jboss? ip, port, queue, user, pw, ect. to set up the local messaging test. 3) Is there a way to config/ develop a mechanism on the remote JMS Jboss server that sends back acknowledgments to the sender (in this case)? I greatly appreciate your help. Thanks, Ben

    Read the article

  • Getting Postfix to process the mail queue on Mac OS X

    - by paperclip
    I'm in the process of getting Sendmail/Postfix setup so that I can send and test my PHP scripts when using the mail() function. I've got to the point that when I run the mail() function in PHP, the script executes without any errors and the mail is sent to my mailq but it then does not get processed and simply times-out with a message of: Operation timed out. An excerpt from Terminal: -Queue ID- --Size-- ----Arrival Time---- -Sender/Recipient------- 137AA96B6C2 897 Tue Mar 16 22:27:05 nobody@imac.local (connect to alt4.gmail-smtp-in.l.google.com[74.125.93.27]: Operation timed out) [email protected] Any ideas as to how I can fix this timeout issue? Thanks, -P.

    Read the article

  • Is Work Stealing always the most appropriate user-level thread scheduling algorithm?

    - by Il-Bhima
    I've been investigating different scheduling algorithms for a thread pool I am implementing. Due to the nature of the problem I am solving I can assume that the tasks being run in parallel are independent and do not spawn any new tasks. The tasks can be of varying sizes. I went immediately for the most popular scheduling algorithm "work stealing" using lock-free deques for the local job queues, and I am relatively happy with this approach. However I'm wondering whether there are any common cases where work-stealing is not the best approach. For this particular problem I have a good estimate of the size of each individual task. Work-stealing does not make use of this information and I'm wondering if there is any scheduler which will give better load-balancing than work-stealing with this information (obviously with the same efficiency). NB. This question ties up with a previous question.

    Read the article

  • howto: vimrc change part of file path and execute script

    - by posop
    I would like to set up a command to execute launch the php script i am editing. :echo expand('%:p:h') yields: C:\xampp\htdocs\my my localhost path is C:\xampp\htdocs i would like to cut the contents of local host off my current directory and append a file separator so i would have: g:var = \my\ so the end goal would be to have something like this in my .vimrc (need help with the concatenate) map <F5> :w<CR>:silent execute '!"c:\Program Files (x86)\Mozilla Firefox\firefox.exe"' "localhost . $var . %"<CR> is this possible? or is there another way to do this?

    Read the article

  • Script.PostDeployment.sql not executing in CI build VS 2010

    - by Suirtimed
    I have a database project in VS 2010 (SQL 2008). The local Deploy Solution action works and executes all of the SQL in the Script.PostDeployment.sql file. When I check changes in, the Build Definition for the continuous integration server executes. The database is deployed into the CI environment, but the PostDeployment script doesn't get executed. I wasn't able to find anything specific to this particular scenario. I also expect I'll need to provide additional information unless this is a trivial problem that I missed somewhere. Additional Information: The build is executing vsdbcmd.exe to deploy. The deployment manifest references the PostDeployment.sql file and it's present in the path with the rest of the files. Here is a reference to a thread on social.msdn.microsoft.com regarding this problem.

    Read the article

  • Manage DNS Zone in "slave" Mode with MS Windows 2008 R2

    - by kockiren
    Hello @all following issue, I configure a DNS Zone "location.domain.tld" for my internal Network and it works well, but now I want to manage the domain.tld for my internal Network, but the domain.tld is managed by an external DNS Provider. In location.domain.tld there are all Clients and Servers for internal use only (with local IPs) all these clients resolve the global Mailserver (for example) over his external Address but now i want to catch single Domainnames and resolve it on my own way. But i did not find a way to solve this issue. Any Ideas? Regards Rene

    Read the article

< Previous Page | 801 802 803 804 805 806 807 808 809 810 811 812  | Next Page >