Search Results

Search found 63736 results on 2550 pages for 'sample data'.

Page 407/2550 | < Previous Page | 403 404 405 406 407 408 409 410 411 412 413 414  | Next Page >

  • Refreshing <div> and load data from php

    - by forgatn
    I have on my page and there is a tag where is some and values filled from mySQL DB. I need some JavaScript I think. When I select one option, I want to display in this propriate DATAs which are in DB. without refreshing whole page. Can you tell me how to do it, if you know that please?:) <div id="country1" class="tabcontent"> <label>Choose protocol</label> <SELECT name="cisloprot"> <?php $con = mysql_connect("localhost", "root", "123456"); $sql = "SELECT kod FROM prot GROUP BY kod"; $rs = mysql_query($sql,$con); while ($r = mysql_fetch_array($rs)) { echo "<OPTION VALUE=".$r['kod'].">".$r['kod']."</OPTION>"; } ?> </SELECT> </div>

    Read the article

  • Creating a function to grab data from an Oracle database (array by ID)

    - by Nick
    I'm trying to create a function that will simply allow me to pass an SQL statement into it, and it will generate an array based on a unique ID I pass it: function oracleGetGata($query, $id="id") { global $conn; $sql = OCI_Parse($conn, $query); OCI_Execute($sql); OCI_Fetch_All($sql, $results, null, null, OCI_FETCHSTATEMENT_BY_ROW); return $results; }   For example I'd like this query $array = oracleGetData('select * from table') to return something like: [1] => Array ( [Title] => Title 1 [Description] => Description 1 ) [2] => Array ( [Title] => Title 2 [Description] => Description 2 ) [3] => Array ( [Title] => Title 3 [Description] => Description 3 )   Rather than what it's returning at the moment: [0] => Array ( [ID] => 3 [TITLE] => Title 3 [DESCRIPTION] => Description 3 ) [1] => Array ( [ID] => 1 [TITLE] => Title 1 [DESCRIPTION] => Description 1 ) [2] => Array ( [ID] => 2 [TITLE] => Title 2 [DESCRIPTION] => Description 2 )   I'd really appreciate any help with this, as the function would save me lots of time! Thank you.

    Read the article

  • Complexity in using Binary search and Trie

    - by user121196
    given a large list of alphabetically sorted words in a file,I need to write a program that, given a word x, determines if x is in the list. Preprocessing is ok since I will be calling this function many times over different inputs. priorties: 1. speed. 2. memory I already know I can use (n is number of words, m is average length of the words) 1. a trie, time is O(log(n)), space(best case) is O(log(n*m)), space(worst case) is O(n*m). 2. load the complete list into memory, then binary search, time is O(log(n)), space is O(n*m) I'm not sure about the complexity on tri, please correct me if they are wrong. Also are there other good approaches?

    Read the article

  • Is is faster to filter and get data or filter then get data ?

    - by remi bourgarel
    Hi I have this kind of request : SELECT myTable.ID, myTable.Adress, -- 20 more columns of all kind of type FROM myTable WHERE EXISTS(SELECT * FROM myLink WHERE myLink.FID = myTable.ID and myLink.FID2 = 666) myLink has a lot of rows. Do you think it's faster to do like this : SELECT myLink.FID INTO @result FROM myLink WHERE myLink.FID2 = 666 UPDATE @result SET Adress = myTable.Adress, -- 20 more columns of all kind of type FROM myTable WHERE myTable.ID = @result.ID

    Read the article

  • Gridview footer data format string

    - by Germ
    Is there a way to format a gridview column footer value as currency. I'v set the DataFormatString of the BoundField to {0:c} but this doesn't seem to affect the footer. Do I have to do it in the RowDataBound event?

    Read the article

  • Handling large (object) datasets with PHP

    - by Aron Rotteveel
    I am currently working on a project that extensively relies on the EAV model. Both entities as their attributes are individually represented by a model, sometimes extending other models (or at least, base models). This has worked quite well so far since most areas of the application only rely on filtered sets of entities, and not the entire dataset. Now, however, I need to parse the entire dataset (IE: all entities and all their attributes) in order to provide a sorting/filtering algorithm based on the attributes. The application currently consists of aproximately 2200 entities, each with aproximately 100 attributes. Every entity is represented by a single model (for example Client_Model_Entity) and has a protected property called $_attributes, which is an array of Attribute objects. Each entity object is about 500KB, which results in an incredible load on the server. With 2000 entities, this means a single task would take 1GB of RAM (and a lot of CPU time) in order to work, which is unacceptable. Are there any patterns or common approaches to iterating over such large datasets? Paging is not really an option, since everything has to be taken into account in order to provide the sorting algorithm.

    Read the article

  • What exactly is the Nullify delete rule doing?

    - by dontWatchMyProfile
    Does that mean that if I delete an managed object which has references (relationship) to some others, the relationships are removed to those others? Example: objectA references objectB and objectC. objectA gets deleted, it's relationship to objectB and objectC is set to the Nullify rule. What happens in detail?

    Read the article

  • error while storing data in datastore

    - by Rahul99
    for st in Traks : trak = TrakHtml() trak.hawb = st url = 'http://etracking.cevalogistics.com/eTrackResultsMulti.aspx?sv='+st result = urlfetch.fetch(url) trak.htmlData = result.content trak.put() trak.htmlData is a textproparty(). It's giving this error: UnicodeDecodeError: 'ascii' codec can't decode byte 0xc2 in position 29284: ordinal not in range(128)

    Read the article

  • hierarchical data from self referencing table in tree form

    - by Beta033
    It looks like this has been asked and answered in all the simple cases, excluding the one that I'm having trouble with. I've tried using a recursive CTE to generate this; however maybe a cursor would be better? Or maybe a set of recursive functions will do the trick? Can this be done in a cte? consider the following table PrimaryKey ParentKey 1 NULL 2 1 3 6 4 7 5 2 6 1 7 NULL should yield PK 1 -2 --5 -6 --3 7 -4 where the number of - marks equal the depth, my primary difficulty is the ordering.

    Read the article

  • How to store data with N columns

    - by iconiK
    I need a way to store an int for N columns. Basically what I have is this: Armies: ArmyID - UINT UnitCount1 - UINT UnitCount2 - UINT UnitCount3 - UINT UnitCount4 - UINT ... I can't possible add a column for each and every unit, so I need a fast way to store the number of each units in an army (you might have guesses it's for a game by now). Using XML is not an option as it will be dead slow.

    Read the article

  • Vb net textbox cursor position always 0 after databinding

    - by user1166557
    please help me with the following if you can, i have the simple following command me.textbox1.databindings.clear me.textbox1.databindings.add("text",TicketsBindingSource,"TicketSubject") the command exeutes succesfully and i can see in the textbox the title, BUT once i click on the textbox1 the position of the cursor it's always moving to the position 0 and not to the area of the textbox i clicked. For example my textbox has the following text: "Hello World". If i click with my mouse inside the textbox at the letter W or anywhere i click the cursor is moving to the 0 index. eg. at the beggining , in order to move my cursor left or right i have to do that with my keyboard arrows keys, Dose anybody know how i can solve this issue?

    Read the article

  • Cakephp, Retreive Data for HABTM Models using conditional find

    - by ion
    There are 2 Models: Project & Category that are bind with HABTM relationship. I would like to perform a search from projects controller that can do the following: FIND all DISTINCT Project.scedule WHERE Category.slug != 'uncategorised' Apologies for the syntax, I'm no sequel expert. What I have managed to do is to retrieve all projects that do not belong to Category uncategorised into an array however I'm not sure as to how to search again the array result for DISTINCT Project.schedule values (needed to fill out a form drop down) I hope I made myself clear.

    Read the article

  • how to select and group mysql data based on the follwoing table

    - by user1151680
    how can I achieve the desired result in mysql if my table looks like this. result|year 1 |2011 2 |2011 1 |2011 0 |2011 1 |2012 2 |2012 1 = Won, 2 = lost, 0 = draw Every year can have multiple values like this. Not sure how I can get the desired result like below. year won lost draw totalPlayed 2011 2 1 1 3 2012 1 1 0 2 I have tried the following query but does not get the desired result select year, league_types.league_name, sum(if(result = 1,1,0)) as won, sum(if(result = 0,1,0)) as draw, sum(if(result = 4,1,0)) as noResult, sum(if(result = 2,1,0)) as lost, sum(if(result = 3,1,0)) as tied, sum(if(result > 0 and result < 4,1,0)) as played from match_score_card inner join fixtures on match_score_card.match_id = fixtures.match_id inner join league_types on fixtures.league_id = league_types.league_id where team_id = 1 group by year order by year desc

    Read the article

  • Cannot perform an ORDERBY against my EF4 data

    - by Jaxidian
    I have a query hitting EF4 using STEs and I'm having an issue with user-defined sorting. In debugging this, I have removed the dynamic sorting and am hard-coding it and I still have the issue. If I swap/uncomment the var results = xxx lines in GetMyBusinesses(), my results are not sorted any differently - they are always sorting it ascendingly. FYI, Name is a varchar(200) field in SQL 2008 on my Business table. private IQueryable<Business> GetMyBusinesses(MyDBContext CurrentContext) { var myBusinesses = from a in CurrentContext.A join f in CurrentContext.F on a.FID equals f.id join b in CurrentContext.Businesses on f.BID equals b.id where a.PersonID == 52 select b; var results = from r in myBusinesses orderby "Name" ascending select r; //var results = from r in results // orderby "Name" descending // select r; return results; } private PartialEntitiesList<Business> DoStuff() { var myBusinesses = GetMyBusinesses(); var myBusinessesCount = GetMyBusinesses().Count(); Results = new PartialEntitiesList<Business>(myBusinesses.Skip((PageNumber - 1)*PageSize).Take(PageSize).ToList()) {UnpartialTotalCount = myBusinessesCount}; return Results; } public class PartialEntitiesList<T> : List<T> { public PartialEntitiesList() { } public PartialEntitiesList(int capacity) : base(capacity) { } public PartialEntitiesList(IEnumerable<T> collection) : base(collection) { } public int UnpartialTotalCount { get; set; } }

    Read the article

  • MYSQL updating data from other table

    - by atif089
    Hi, I have two tables like of this structure content (content_id, content_type, user_id, time, comment_count) comments (comment_id, content_id, userid, comment, comment_time) What I wold like to do is update the comments_count field with sum of comments i.e COUNT(content_id) from the comments table. I am not able to figure out the right syntax Thanks

    Read the article

  • How to **delete-protect** a file or folder on Windows Server 2003 and onwards using C#/Vb.Net?

    - by Steve Johnson
    Hi all, Is it possible to delete-protect a file/folder using Registry or using a custom written Windows Service in C#? Using Folder Permissions it is possible, but i am looking for a solution that even restricts the admin from deleting specific folders. The requirement is that the administrator must not be easily track the nature of protection and/or may not be able to avert it easily. Obviously all administrators will be able to revert the procedure if the technique is clearly understood. Like folder Permissions/OwnerShip Settings can easily be reset by an administrator. SO that is not an option. Folder protection software can easily be uninstalled and show clear indication that a particular folder is protected by some special kind of software. SO that too is not an option. Most antivirus programs protect folders and files in Program Dir. Windows itself doesnt allow certain files such as registry files in c:\windows\system32\config to not even copied. Such a protection is desired for folders which allowse to read and write to files but not allow deletion. Similar functionality is desired. The protection has to seemless and invisible. I do not want to use any protection features like FolderLock and Invisible secrets/PC Security and Desktop password etc. Moreover, the solution has to be something other than folder encryption. The solution has to be OS-native so ** that it may implemented ** pro grammatically using C#/VB.Net. Please help. Thanks

    Read the article

  • How to convert my mysql data into this?

    - by sky
    <script type="text/javascript"> var cityList = new Array(....etc), signImgList = new Array('http:\/\/pic.sitename.com:80\/file\/11\/34\/01\/22\/default\/SIGN11340122_48x48.jpg?t=1257391453468'....etc), titleList = new Array(....etc), userSex = new Array(....etc), userAge = new Array('19','26'....etc), userMid = new Array('lwowl','kylin0621'....etc); </script>

    Read the article

  • Caching vector addition over changing collections

    - by DRMacIver
    I have the following setup: I have a largish number of uuids (currently about 10k but expected to grow unboundedly - they're user IDs) and a function f : id - sparse vector with 32-bit integer values (no need to worry about precision). The function is reasonably expensive (not outrageously so, but probably on the order of a few 100ms for a given id). The dimension of the sparse vectors should be assumed to be infinite, as new dimensions can appear over time, but in practice is unlikely to ever exceed about 20k (and individual results of f are unlikely to have more than a few hundred non-zero values). I want to support the following operations efficiently: add a new ID to the collection invalidate an existing ID retrieve sum f(id) in O(changes since last retrieval) i.e. I want to cache the sum of the vectors in a way that's reasonable to do incrementally. One option would be to support a remove ID operation and treat invalidation as a remove followed by an add. The problem with this is that it requires us to keep track of all the old values of f, which is expensive in space. I potentially need to use many instances of this sort of cached structure, so I would like to avoid that. The likely usage pattern is that new IDs are added at a fairly continuous rate and are frequently invalidated at first. Ids which have been invalidated recently are much more likely to be invalidated again than ones which have remained valid for a long time, but in principle an old Id can still be invalidated. Ideally I don't want to do this in memory (or at least I want a way that lets me save the result to disk efficiently), so an idea which lets me piggyback off an existing DB implementation of some sort would be especially appreciated.

    Read the article

< Previous Page | 403 404 405 406 407 408 409 410 411 412 413 414  | Next Page >