Search Results

Search found 19375 results on 775 pages for 'codeigniter url'.

Page 341/775 | < Previous Page | 337 338 339 340 341 342 343 344 345 346 347 348  | Next Page >

  • Help submitting an asp.net form with jquery

    - by batman_man
    I'm trying to post the entire asp.net form to a certain url. I have tried: $.post("http://www.someaddress.com", $("form").serialize()); I have also tried: $.ajax({ type:"POST", url:"http://www.someaddress.com", data: $('form').serialize(), success: function(){ alert('yay'); } }); In both cases the submit is fine but no data is passed along with it. When i test the form.serialize() in firebug console, this shows my form serialized just fine. When i view the submit in fiddler, i can see that the data part is not set. Maybe im not understanding the data part, but every single tutorial shows this as the way to go - serialize the form and set that as data. What must i do to get my serialized form as the data in my request? What am i missing? Also - why does the NET tab in firebug show all these requests as method OPTIONS?

    Read the article

  • Is it possible to create a service like Feed My Inbox on my own server?

    - by Mark Bowen
    I was just wondering if it's at all possible to create a service like Feed My Inbox on my own server using PHP? Basically I have a site which has RSS feeds which are dynamic in nature and can search from thousands of posts based on many different criteria. I have the RSS feed working fine and bringing back data dynamically for whatever criteria I want so that bits fine. I am using the ExpressionEngine CMS to handle the site and there will be thousands of users on the site (currently there are around 2,0000) but that number is exponentially growing every single day. What I want to be able to do is allow the users to choose from certain criteria which will then build a dynamic RSS URL which will then be stored in a database table (one row for each user). This bit I will be able to do myself but then I want to be able to send out new RSS feed items via e-mail to each user. This is the part I'm a little stuck on. I'm guessing I would somehow need to run a cron job to hit a page which would check each users RSS feed and then if there are new items to send them to the user via e-mail. That's where I am totally stuck though and I'm just wondering what the best way to go about it would be? That or any software in PHP that already does this sort of thing would be great. I tried out phpList but it has severe problems working with RSS and I only ever got it to work once and now never again and I've read that lots of people have had this same problem so unfortunately it's not just me :-( I know there are services such as Feed My Inbox which I could easily set up so that users click a link and their RSS feed URL is added to go and use that service but I want to keep users from seeing the dynamic nature of the feed or they will easily be able to modify it to get at other items in the feed. I need this so that I can charge for access to the feeds but if people can see the URL of the feed then I will be totally unstuck as they will be able to get at whatever they want very easily. Therefore I'd like to be able to send the items out to them. Would really love to hear if anyone knows if this kind of thing is possible at all and what would be involved?

    Read the article

  • How To Discover RSS Feeds for a given site.

    - by ktolis
    The quest is, given a site url (say http://stackoverflow.com/ ) to return the list of all the feeds available on the site. Methods acceptable: a) use a 3rd party service (google?, yahoo?, ...) programmatically b) using a crawler/spider (and some tips on how to configure the spider to return the rss/xml feeds only) c) programmatically using c/c++/php (any language/library) The task here is not to get the feeds contained on the page returned by the url but ALL the feeds that are available on the server at any depth... in any cases please provide a simple usage example.

    Read the article

  • Getting 403 error when using CSRF filter with tomcat 6.0.32

    - by sps
    This is my filer config in web.xml <filter> <filter-name>CSRFPreventionFilter</filter-name> <filter-class>org.apache.catalina.filters.CsrfPreventionFilter</filter-class> <init-param> <param-name>entryPoints</param-name> <param-value>/login<param-value> </init-param> </filter> <filter-mapping> <filter-name>CSRFPreventionFilter</filter-name> <url-pattern>/*</url-pattern> </filter-mapping> <filter> Am I missing something? Are any code-changes necessary to enable csrf protection in tomcat

    Read the article

  • Simple ASP.NET MVC Routing question

    - by Robert
    Hi there, I have two pages in my simple MVC App with two defined routes: routes.MapRoute( "Results", // Route name "Results/{id}", // URL with parameters new { controller = "Results", action = "Index", id = "" } // Parameter defaults ); routes.MapRoute( "Default", // Route name "{controller}/{action}/{id}", // URL with parameters new { controller = "Main", action = "Index", id = UrlParameter.Optional } // Parameter defaults ); I needed to have the results page load with just a product ID such as this: [MyDomain....]/Results/12345. But also the main page does a POST (using JQuery) to the Results Controller for updates using this route: [MyDomain....]/Main/Update along with a data bag. This works fine when I only have the "Default" route. But when I added the other "Results" route, all the POST calls to update are failing. Any ideas what I'm doing wrong??? Thanks a lot.

    Read the article

  • how to pass javascript function argument within JSF component?

    - by Milan
    Hello everybody! I have the folowing code: <script Language="JavaScript"> function load(url) { var load = window.open(url,'','scrollbars=no,menubar=no,height=600,width=800,resizable=yes,toolbar=no,location=no,status=no'); } </script> <h:commandLink value="aaa" onclick="load('<h:outputText value="http://www.google.com" /> '); /> I want to pass attribute in JS function but probably inside onclick is not the right way. Any solution?

    Read the article

  • Can a Javascript bookmarklet overlay an image on a web page?

    - by songdogtech
    Can a bookmarklet be used to overlay an image on a web page? Not as a pop-up, but as a image positioned by CSS and with a high z-index to display on top of other elements. And without a mask i.e., Shadowbox or similar jQuery effect. Just an image from a URL and positioned in the bottom left hand corner of the browser window. This is what I have so far, but it may be the wrong direction to be going: javascript:(function(){document.write("body {background-image:url(http://mydomain.com/image.png); position: absolute; left:50px; top:300px; z-index:9999;}");})() I have a JS function that works as a bookmarklet to change the case of text on the page, and now I'd like to be able to show an image when the bookmarklet is used.

    Read the article

  • Use alternative browser with Subversion URLs

    - by Mark Howard
    We have subversion setup with http access through apache at http ://svn.local/ This provides access through svn clients and the standard svn repository browser. We also have a separate repository browser installed at http ://svn-browser.local/ (fisheye in this case). Is it possible to change the subversion/apache server so that browser requests are redirected to the subversion browser, but requests from svn clients go to the subversion server? Essentially, I'd like to use a single URL to identify a resource in both svn clients and the svn browser and be able to copy the url from a svn client into a web browser to have the enhanced functionality provided by the browser.

    Read the article

  • How to check a complex condition in Smarty(PHP)

    - by php html
    I need to display a section or another in a smarty template. My condition is simple: if a smarty value starts with a string I should display one section, otherwise the other smarty section should be displayed. I can change only the tpl files. {php} if (substr($url,0,4) != 'http') { {/php} section 1 {php} } else { {/php} section 2 {php} } {/php} The problem is that I can not read the url varible which was previously assigned using $smarty-assign. Basically, I'm looking for the smarty function that can be used to retrieve a value, or if there is a better solution.

    Read the article

  • Using CURL with Google

    - by TheBounder
    I want to CURL to Google to see how many results it returns for a certain search. I've tried this: $url = "http://www.google.com/search?q=".$strSearch."&hl=en&start=0&sa=N"; $ch = curl_init(); curl_setopt($ch, CURLOPT_HEADER, 0); curl_setopt($ch, CURLOPT_VERBOSE, 0); curl_setopt($ch, CURLOPT_RETURNTRANSFER, true); curl_setopt($ch, CURLOPT_USERAGENT, "Mozilla/4.0 (compatible;)"); curl_setopt($ch, CURLOPT_URL, $url); curl_setopt($ch, CURLOPT_POST, true); $response = curl_exec($ch); curl_close($ch); But it just returns a 405 Method Allowed google error. Any ideas? Thanks

    Read the article

  • Python BeautifulSoup Print Info in CSV

    - by Codin
    I can print the information I am pulling from a site with no problem. But when I try to place the street names in one column and the zipcodes into another column into a CSV file that is when I run into problems. All I get in the CSV is the two column names and every thing in its own column across the page. Here is my code. Also I am using Python 2.7.5 and Beautiful soup 4 from bs4 import BeautifulSoup import csv import urllib2 url="http://www.conakat.com/states/ohio/cities/defiance/road_maps/" page=urllib2.urlopen(url) soup = BeautifulSoup(page.read()) f = csv.writer(open("Defiance Steets1.csv", "w")) f.writerow(["Name", "ZipCodes"]) # Write column headers as the first line links = soup.find_all(['i','a']) for link in links: names = link.contents[0] print unicode(names) f.writerow(names)

    Read the article

  • How do I hide the text links over a toggleable horizontal list with background images.

    - by Sivakanesh
    I'm trying to create a UL/LI horizontal list with background images only, with no text link visible. The reason for this is so that when I over over a list item, the background would rollover and when I click on it the current item would toggle. basically it is a horizontal menu with background images that can be toggled; mimicking the job of a radio button. I have done it like this; <div id="options"> <ul id="list"> <li class="active"><a href="#" class="option1 active" id="link1"><span>XXXXX</span></a></li> <li><a href="#" class="option2" id="link2"><span>XXXXX</span></a></li> <li><a href="#" class="option3" id="link3"><span>XXXXX</span></a></li> </ul> </div> The CSS for option1, option2 and option3 simply define the background image. #options LI{list-style-type: none; display : inline} a.option1{ background:url('../images/option1.png') no-repeat;} a.option2{ background:url('../images/option2.png') no-repeat;} a.option3{ background:url('../images/option3.png') no-repeat;} a.option1, a.option2, a.option3{ background-position:top; display:inline; width:230px; height:40px; } And the hover & active css part simply sets the background position like so- a.option1:hover, a.option2:hover, a.option3:hover{ background-position:bottom; } a.active{ background-position:bottom !important; } This works fine, however on top of the background I get the words "XXXXX" as text links and I'm struggling to hide them. They are interfering with the hover action and preventing rollover (even if I replace XXXXX with a period or something short). I can't just remove the text from the link as it would hide the whole LI element. I have tried to use display:none; or text-indent:-999px but then the whole UI element becomes invisible. I can't understand what I'm doing wrong. Are you able to help? Thanks

    Read the article

  • Why my HttpPost can't receive all response data?

    - by Johnny
    I'm on Android 1.5, and my code is like this: HttpPost httpPost = new HttpPost(url); HttpEntity entity = new UrlEncodedFormEntity(params, HTTP.UTF_8); httpPost.setEntity(entity); HttpResponse response = httpClient.execute(httpPost); HttpEntity respEntity = response.getEntity(); String result = EntityUtils.toString(respEntity, DEFAULT_CHARSET); After successfully executed these codes, the result is a stripped string. I've tried using browser to test the url+param, it works fine and got all data. What's wrong with this code? Is there any parameters I need to specified?

    Read the article

  • Parsing XML with jQuery

    - by Jamie
    I've used this: $(document).ready(function () { $.ajax({ type: "GET", url: "http://domain.com/languages.php", dataType: "xml", success: xmlParser }); }); function xmlParser(xml) { $('#load').fadeOut(); $(xml).find("result").each(function () { $(".main").append('' + $(this).find("language").text() + ''); $(".lang").fadeIn(1000); }); } I used a local XML file on my computer, it works fine, but when I change the URL to an website, it just keeps loading all the time... How do I get this to work?

    Read the article

  • How to find why NSMutableData is invalid

    - by 4thSpace
    I access a RESTFUL url and get back results. The results are in JSON. I turn the response into a string via: - (void)connectionDidFinishLoading:(NSURLConnection *)connection { NSString *json = [[NSString alloc] initWithBytes:[self.receivedData mutableBytes] length:[self.receivedData length] encoding:NSUTF8StringEncoding]; The json variable has a value of 0x0. When I mouse over it, I see <Invalid CFStringRef. How can I debug this to tell why it is invalid? I render the JSON given back through the browser in A JSON parser. That checks out fine. Results are given back by entering an ID in the URL. Other IDs return results without issue. The result set is fairly large.

    Read the article

  • PHP Image content type problem

    - by Mirko
    Hi everybody, I have a specific problem, and cant get over it. For my latest project I need a simple PHP script that display an image according to its ID sent through URL. Here's the code: header("Content-type: image/jpeg"); $img = $_GET["img"]; echo file_get_contents("http://www.somesite.hr/images/$img"); The problem is that the image doesn't show although the browser recognizes it (i can see it in the page title), instead I get the image URL printed out. It doesn't work neither on a server with remote access allowed nor with one without. Also, nothing is printed or echoed before the header. I wonder if it is a content type error, or something else. Thanks in advance.

    Read the article

  • jQuery: How do I pass a value into an Ajax call?

    - by Legend
    I am updating some div as follows: for(var i = 0; i < data.length; i++) { var query = base_url + data[i]; $.ajax({ url: query, type: 'GET', dataType: 'jsonp', timeout: 2000, error: function() { self.html("Network Error"); }, success: function(json) { $("#li" + i).html("<img src='" + json.result.list[0].url + "' />") } }); } The value of i does not work inside the ajax call. I am trying to pass the value of i so that it can attach the element to the proper div. Can someone help me out?

    Read the article

  • how to give 2 parameters to remote_function thanks to 'with' ???

    - by bahamut100
    Hi, I try to give a lot of parameters in a remote_function like this : <?php echo remote_function(array( 'update' => 'test', 'url' => 'conges/verifdate', 'with' => "'date_deb=' + $('date_debut').value" )); ?> This code works, in my action.class I can to get the parameter. But when I try to give a second parameter, it doesn't work : <?php echo remote_function(array( 'update' => 'test', 'url' => 'conges/verifdate', 'with' => "'date_deb=' + $('date_debut').value"."'+&date_fin=' + $('date_fin')" )) ?> Why ??

    Read the article

  • How do HTTP proxy caches decide between serving identity- vs. gzip-encoded resources?

    - by mrclay
    An HTTP server uses content-negotiation to serve a single URL identity- or gzip-encoded based on the client's Accept-Encoding header. Now say we have a proxy cache like squid between clients and the httpd. If the proxy has cached both encodings of a URL, how does it determine which to serve? The non-gzip instance (not originally served with Vary) can be served to any client, but the encoded instances (having Vary: Accept-Encoding) can only be sent to a clients with the identical Accept-Encoding header value as was used in the original request. E.g. Opera sends "deflate, gzip, x-gzip, identity, *;q=0" but IE8 sends "gzip, deflate". According to the spec, then, caches shouldn't share content-encoded caches between the two browsers. Is this true?

    Read the article

  • Convert UTC date to actual date and time

    - by evann
    I have a chart of bitcoin prices. I have the correct prices on the Y-axis, but I cannot get the correct time on the X-axis. The time shows up in a UTC format in my console. I am adding price and date to the series each iteration. I need to get the date of that particular result and find the YEAR, MONTH and DAY of that so I can put it in the right format. Any help is appreciated thanks. $.ajax({ url: "/chart/ajax_get_chart", // the URL of the controller action method dataType: "json", type: "GET", success: function (result) { var result = JSON.parse(result); series = []; for (var i = 0; i < result.length; i++) { tempArray = [parseFloat(result[i]['price'])]; tempDate = Date.parse(result[i]['date']); series.push(tempDate); series.push(tempArray); }

    Read the article

  • Deny http access to a directory, allow access from WordPress plugin

    - by luke
    Hey. I need to prevent direct access to http://www.site.com/wp-content/uploads/folder/something.pdf through the browser. However the Download Monitor plugin I am using, which allows logged in users to download the file, needs to be able to work. Trying Order Allow,Deny Deny from all Allow from all but the download links do not now work... even though (I think) they are links produced by the script e.g. http://www.site.com/wp-content/plugins/download-monitor/download.php?id=something.pdf Enter that in the address bar and you correctly get a WordPress message, 'You must be logged in to download this file.' However, if someone knows the URL where the file was uploaded http://www.site.com/wp-content/uploads/folder/something.pdf they can still access it directly. I don't know how (guesswork?) they would find the direct URL anyway, but the client wants it stopped! Thanks for any help.

    Read the article

  • Reload Document into Google Docs Viewer (Clear Cache)

    - by Adam
    Google Docs Viewer (http://docs.google.com/viewer) creates a cache of a document after the first viewing. To see what I mean, try the following: Upload file.pdf to your server (i.e., http://example.com). Visit http://docs.google.com/viewer?url=http://example.com/file.pdf Upload a new file to replace file.pdf (but use the same name). Revisit http://docs.google.com/viewer?url=http://example.com/file.pdf. Google Docs Viewer still shows the old file.pdf. Anyone know how to correct this? (I have already tried clearing browser cache, switching browsers, and logging in with a different google account to view the link.)

    Read the article

  • Updating div after full page refresh following AJAX request

    - by TonE
    Hi, I have an AJAX controller action which returns JSON. The returned JSON is handled via jQuery and contains a message displayed in an update div. The JSON also contains a flag which indicates a change in state. If this is true the entire page needs to be reloaded from a different url. This is currently performed in the jQuery using window.location.replace(url). In the situation where a full page refresh occurs, I still want to display the message from the JSON in the update div, but after the page has fully reloaded. How can I ensure the div is updated with the message after the window.location.replace call and the page has been completely reloaded? Thanks. (I realise AJAX is not intended for full page refreshes but I can't think of a simple way of avoiding it in this scenario. Normal behaviour here is updating the div with a status message.)

    Read the article

  • question about System.Net

    - by backdoor
    hi all. when i use some code like this: System.Net.WebClient objClient = new WebClient(); string url = "http://google.com"; objClient.DownloadString(url); it takes some seconds to connection stablished and downloading starts... i reinstall my windows yesterday. before yesterday i havent this problem. some times when i reinstall my windows this problem occures. does anyboddy know why this problem occures?? thanks all and sorry for my bad english....

    Read the article

< Previous Page | 337 338 339 340 341 342 343 344 345 346 347 348  | Next Page >