Search Results

Search found 2201 results on 89 pages for 'webpage'.

Page 8/89 | < Previous Page | 4 5 6 7 8 9 10 11 12 13 14 15  | Next Page >

  • How to select all text inside text area of code field on web pages

    - by Moorage
    In all the forums web pages the download links are given inside the [code] braces that looks like text area with white background. When the download links are in large numbers then there are scroll bars. I find it difficult to scroll up and then select and then drag down to select all links. Is there any way to select all , when I click on select all on right click then it usually selects the text from whole page not from that code segment

    Read the article

  • How to decomment an html/php webpage?

    - by Sam
    A crazy question: Imagine a webpage file called somepage.php And it contains some html php contents in my editor I see: <html><head></head><body> <?=$welcome . $essay . $thatsAllForNowFolks . $footer ?> <!-- Blue Ball Bell Blow Bows Bats Beef Bark Bill Boss --> </body></html> When I browse my site I see those comments in the final result, while I only want that comment to be only in my editor for my secretive inspirations and don't want the whole world to know what I'm thinking when I'm developing, as well as I see those comments for any and all my website visitors as wasted bandwitch of internet speed. How do I decomment my entire html/php files at the moment the html is served? Ideas, code and suggestions are much appreciated. My thanks in advance...

    Read the article

  • Getting all the cookies including those from google analytics from one webpage

    - by DixieFlatline
    Hello! I use apache 4 java libs for http get connections. I get all cookies from 2 other pages. But one other webpage contains google analytics cookies(utma,utmz...) and classic session id cookie (checked with firebug). When i try to get the cookies i only get sesson id cookie. Is there any trick to get all the cookies from java in connection with google analytics? If i want to make a http post, it seems that i have to send google analytics cookies too. I can copy my code if somebody wants, but it works on other pages(for printing cookies). Thank you

    Read the article

  • Animate rotating SVG element on webpage

    - by Zarkonnen
    So I have an SVG file created in Inkscape embedded in a webpage, and I'd like it to rotate slowly. I've tried using Javascript and inserting animation commands directly into the SVG, but nothing works. I don't want to load in an entire JS library for this one task. This is what I have so far: <html> <body bgcolor="#333333"> <embed src="gear.svg" id="gear" width="1000" height="1000" style="position: absolute; top: -500px; left: -500px;" /> <script type="text/javascript"> var gear = document.getElementById("gear"); window.setInterval(function() { // Somehow animate the gear. }, 10); </script> </body> </html>

    Read the article

  • Accessing a webpage in C++

    - by wyatt
    Is there a good, simple library which allows C++ to load a webpage? I just want to grab the source as text. I'm not using any IDE or significant library, just straight command line. Tangentially, is there something fundamental I'm missing about programming in C++? I would think any language in common use today would have droves of web-based functionality, being so central to computer usage, but I can find next to no discussion on how to accomplish it. I realise C++ significantly predates the modern internet, so it lacking any core ability in the regard is reasonable, but the fact that relevant libraries seem so sparse is baffling. Thanks for your help.

    Read the article

  • WatiN downloading a file from webpage

    - by user541597
    I have the following code which does a couple of webpage redirections and then clicks an tag which has a javascript fuction as the href. When it is called a file is download. My problem is I want to be able to download the file without being prompted to cancel, save or open. Please help I am using IE9 using (var browser = new IE("http:url.aspx")) { browser.TextField(Find.ByName("ctl00$ContentPlaceHolder1$Login1$UserName")).TypeText("cpereyra"); browser.TextField(Find.ByName("ctl00$ContentPlaceHolder1$Login1$Password")).TypeText("Maxipereyra15"); browser.Button(Find.ByName("ctl00$ContentPlaceHolder1$Login1$LoginButton")).Click(); browser.GoTo("http://it-motivity-cmc/Movation/MyPage/MyDashboard.aspx?dynamicdashboardid=ab000000-7dea-11c9-b596-d01e04bebb94"); while (browser.Eval("document.readyState") != "complete") { Thread.Sleep(1000); } Div div = browser.Div("ctl00_ContentPlaceHolder1_wrapper_vis_zone1_1"); div.Link(link => link.Text == "Export to CSV").Click(); }

    Read the article

  • Stop moving page title in Firefox?

    - by lilydjwg
    Some websites I'm using always make the web title move or blink when, for example, a new mail has just arrived. It is really disturbing as it's not an emergency and usually I'm busy browsing other pages. I can't stand it any longer, but I have to use those services. I want to stop this kind of behavior either by installing a new extension or writing a Greasemonkey script. But I can't find one and have no idea about how to detect and stop the JavaScript code. What is the solution?

    Read the article

  • How do I Reload Ajax Call Parameters without Reloading the webpage

    - by Snowright
    I'm working with Extjs 2.2.1 with Alfresco 3.2 enterprise. I would like to update the ticket that handles authentication to the alfresco server on components that have been loaded during login. This ticket expires after a set time and this is why I will need to update the ticket. Options that do not seem viable for me(but please let me know if I'm wrong): Reload the components to reload the call parameters - I can't do this because it resets whatever the user was previously working on (ie. Tree panel gets reloaded, grid filters reset, etc). The actual webpage never reloads as everything uses ajax calls to update things on the page. Create a global variable that stores the ticket and attach it as a call parameter with any ajax calls - Any components that were loaded during login will still use the original ticket to make calls to the server.

    Read the article

  • Dynamically changing the component of webpage

    - by peril brain
    Hi, I am working for a project on disaster management sponsered by World Bank, under this project i have assigned work on CMS. and my senior has asked me to make a ASP.NET webpage where those logged with administrator privilege will be able to edit the page. like:- <div id="mydiv"><pre>+++++++MY CONTENT+++++++++++++++++++</pre></div> now if if the admin clicks on mydiv he should be able to edit the pre contents. for this i got many options such as WYSIWYG editors (but they work on textboxs only and further more they cant save the data back to server replacing the previous content) then I came to know about Webparts , but its editing capabilities was so low that my plan for implementing a WYSIWYG editor went in vain. so I m seeking help from u all is their any way to achieve the motive, it would be better if u can suggest something like incorporating WYSIWYG into webpartmanager editor... THANKS..

    Read the article

  • Wait until a webpage finished loading to load the next one in a list

    - by envy
    Hi! I'm using PyQT webView to visit some webpages I have stored on a dictionary the code is something like this: def loadLink(self, url): manager = QNetworkAccessManager() request = QNetworkRequest(QUrl(url)) self.ui.webView.load(QUrl(visitar)) def readUnreadLinks(self): print "Links to read: " + str(len(self.unreadLinks)) for link in self.unreadLinks: print "link-> " + str(link) self.loadLink(link) the problem is it doesn't wait until finished loading the web page and starts loading the next one. I want to load a webpage, wait until it finished loading and then load the next one. Thanks, this is driving me crazy :)

    Read the article

  • How to suspend and resume browser tab groups?

    - by Robin Green
    Modern browsers such as Firefox and Chrome are able to shutdown and restore the same group of tabs they had before the next time you open them, and even sometimes remember additional tab state such as scroll position within the tab. I would like ideally to be able to select a group of tabs (about some coherent theme, perhaps) and save them away into some kind of "folder" that I could later open independently, without opening any other tabs. Obviously I can use bookmarks for this, but bookmarks are kind of old-fashioned. It would be excellent if it could also record forms that I had filled in but not submitted, scroll position, tab position within the tab group, and any other "dynamic" aspects of the current tab state.

    Read the article

  • All browsers hang on windows 7 in tab says 'Waiting for...' need restart

    - by Lakshmila
    I'm having a frustrating problem with browsers on OS windows 7 home premium. IE 10: My web browsing works ok for a while and then pages freeze and tab at top says "Waiting for ...' and never loads. Strangely I can still get to Google search results but if I click anything it freezes. Chrome freezes too and says 'The page has become unresponsive ...kill page' Safari stops also. I have tried clearing cookies, cache, turning off Bitdefender firewall and nothing apart from a restart fixes it. consequently I am forced to restart my computer several times a day to stop this browser hang. I have no idea what could be causing this ..any help greatly appreciated.

    Read the article

  • Is it possible to open several web pages when browser is started?

    - by gotqn
    I am searching for browser option or plugin (it will be best if it is available in Web Kit browsers,Opera or Firefox - not IE) that allows me to open several web pages when it is initially started. For example, let's say that I have some file with settings in which I have pointed the following websites: Google + gmail StackOverflow.com SuperUser.com dba.stackexchange.com linkedin etc... and when I firstly started the Chrome browser, all this sites will be opened in new tabs and because the browser has saved my passwords I will be logged in. I will find this very helpful because: It will saves me time I will not miss anything when I turn off my computer (for example to forget to check my mail)

    Read the article

  • How to save webpage to string with cookies support (httpWebRequest)

    - by Maciej
    I need to read webpage and store its content in string for further processing. Sounds simply but I have problem with cookies support. Opened page says I need browser supporting cookies (or turned on). I've made method trying do that via httpWebRequest - which normally works to me but I've come to a standstill with those unfortunate cookies... Any idea how to make it working? Here is my method: string ReadHtml (string address, string encoding) { Uri url = new Uri(address); CookieContainer cookieContainer = new CookieContainer(); HttpWebRequest httpWebRequest = (HttpWebRequest)HttpWebRequest.Create(url); httpWebRequest.AllowAutoRedirect = true; httpWebRequest.KeepAlive = true; httpWebRequest.CookieContainer = cookieContainer; httpWebRequest.UserAgent = "Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1; SV1; .NET CLR 1.1.4322; .NET CLR 2.0.50727)"; httpWebRequest.Method = "GET"; HttpWebResponse webResponse = (HttpWebResponse)httpWebRequest.GetResponse(); // Code Page Encoding enc = Encoding.GetEncoding(encoding); // Read content StreamReader loResponseStream = new StreamReader(webResponse.GetResponseStream(),enc); string lcHtml = loResponseStream.ReadToEnd(); webResponse.Close(); loResponseStream.Close(); return lcHtml; }

    Read the article

  • Trying to grab just absolute links from a webpage using BeautifulSoup

    - by Kevin
    I am reading the contents of a webpage using BeautifulSoup. What I want is to just grab the <a href> that start with http://. I know in beautifulsoup you can search by the attributes. I guess I am just having a syntax issue. I would imagine it would go something like. page = urllib2.urlopen("http://www.linkpages.com") soup = BeautifulSoup(page) for link in soup.findAll('a'): if link['href'].startswith('http://'): print links But that returns: Traceback (most recent call last): File "<stdin>", line 2, in <module> File "C:\Python26\lib\BeautifulSoup.py", line 598, in __getitem__ return self._getAttrMap()[key] KeyError: 'href' Any ideas? Thanks in advance. EDIT This isn't for any site in particular. The script gets the url from the user. So internal link targets would be an issue, that's also why I only want the <'a'> from the pages. If I turn it towards www.reddit.com, it parses the beginning links and it gets to this: <a href="http://www.reddit.com/top/">top</a> <a href="http://www.reddit.com/saved/">saved</a> Traceback (most recent call last): File "<stdin>", line 2, in <module> File "C:\Python26\lib\BeautifulSoup.py", line 598, in __getitem__ return self._getAttrMap()[key] KeyError: 'href'

    Read the article

  • Webpage data scraping using Java

    - by Gemma
    I am now trying to implement a simple HTML webpage scraper using Java.Now I have a small problem. Suppose I have the following HTML fragment. <div id="sr-h-left" class="sr-comp"> <a class="link-gray-underline" id="compare_header" rel="nofollow" href="javascript:i18nCompareProd('/serv/main/buyer/ProductCompare.jsp?nxtg=41980a1c051f-0942A6ADCF43B802');"> <span style="cursor: pointer;" class="sr-h-o">Compare</span> </a> </div> <div id="sr-h-right" class="sr-summary"> <div id="sr-num-results"> <div class="sr-h-o-r">Showing 1 - 30 of 1,439 matches, The data I am interested is the integer 1.439 shown at the bottom.I am just wondering how can I get that integer out of the HTML. I am now considering using a regular expression,and then use the java.util.Pattern to help get the data out,but still not very clear about the process. I would be grateful if you guys could give me some hint or idea on this data scraping. Thanks a lot.

    Read the article

  • background image shows up on right-click show image, but not on webpage

    - by William
    okay, so I'm trying to set up a webpage with a div wrapping two other divs, and the wrapper div has a background, and the other two are transparent. How come this isn't working? here is the CSS: .posttext{ float: left; width: 70%; text-align: left; padding: 5px; background-color: transparent !important; } .postavi{ float: left; width: 100px; height: 100%; text-align: left; background-color: transparent !important; padding: 5px; } .postwrapper{ background-image:url('images/post_bg.png'); background-position:left top; background-repeat:repeat-y; } and here is the HTML: <div class="postwrapper"> <div class="postavi"><img src="http://prime.programming-designs.com/test_forum/images/avatars/hacker.png" alt="hacker"/></div><div class="posttext"><p style="color: #ff0066">You will have bad luck today.</p>lol</div> </div> Edit: at request, here is a link to the site: http://prime.programming-designs.com/test_forum/viewthread.php?thread=33 Edit: edited css to be correct, still suffering same problem

    Read the article

  • how to save sitecore webpage in html file on local disk or server

    - by Sam
    WebRequest mywebReq ; WebResponse mywebResp ; StreamReader sr ; string strHTML ; StreamWriter sw; mywebReq = WebRequest.Create("http://domain/sitecore/content/test/page10.aspx"); mywebResp = mywebReq.GetResponse(); sr = new StreamReader(mywebResp.GetResponseStream()); strHTML = sr.ReadToEnd(); sw = File.CreateText(Server.MapPath("hello.html")); sw.WriteLine(strHTML); sw.Close(); Hi , I want to save sitecore .aspx page into html file on local disk , but i am getting exception. But if i use any other webpage example(google.com) it works fine. The exception : System.Net.WebException was caught HResult=-2146233079 Message=The remote server returned an error: (404) Not Found. Source=System StackTrace: at System.Net.WebClient.DownloadFile(Uri address, String fileName) at System.Net.WebClient.DownloadFile(String address, String fileName) at BlueDiamond.addmodule.btnSubmit(Object sender, EventArgs e) in c:\inetpub\wwwroot\STGLocal\Website\addmodule.aspx.cs:line 97 InnerException: Any help. Thanks in advance

    Read the article

  • Kindle Fire Cant Fit My Webpage inside a Webview of specific size

    - by Madhavan Rangarao
    This is baffling to me. Please help, I could not figure it out ... In my sample html file I have set the meta tag to be <meta name="viewport" content="target-densitydpi=device-dpi, user-scalable=no"> to fit the webpage inside my webview of custom size say 800x600. In Android, I had to specify "target-densitydpi=device-dpi" and it did the job nicely. I tested my custom web page with nexus 7 tablet and the web page fits inside my web view correctly. The same code does not work in Kindle Fire. Only a part of my web page is shown and even if I set the 'initial-scale=1.0' did not help. I tried various settings programmatically but it did not help either. webview.getSettings().setBuiltInZoomControls(true); webview.getSettings().setSupportZoom(true); webview.getSettings().setLoadWithOverviewMode(true); webview.setInitialScale(1); webview.getSettings().setUseWideViewPort(true); Any pointers?

    Read the article

  • Read quicktime movie from servlet in a webpage?

    - by khue
    Hi, I have a servlet that construct response to a media file request by reading the file from server: File uploadFile = new File("C:\\TEMP\\movie.mov"); FileInputStream in = new FileInputStream(uploadFile); Then write that stream to the response stream. My question is how do I play the media file in the webpage using embed or object tag to read the media stream from the response? Here is my code in the servlet: public void doPost(HttpServletRequest request, HttpServletResponse response) throws ServletException, IOException { request.getParameter("location"); uploadFile(response); } private void uploadFile(HttpServletResponse response) { File transferFile = new File("C:/TEMP/captured.mov"); FileInputStream in = null; try { in = new FileInputStream(transferFile); } catch (FileNotFoundException e) { System.out.println("File not found"); } try { System.out.println("in byes i s" + in.available()); } catch (IOException e) { } DataOutputStream responseStream = null; try { responseStream = new DataOutputStream(response.getOutputStream()); } catch (IOException e) { System.out.println("Io exception"); } try { Util.copyStream(in, responseStream); } catch (CopyStreamException e) { System.out.println("copy Stream exception"); } try { responseStream.flush(); } catch (IOException e) { } try { responseStream.close(); } catch (IOException e) { } } And here is html page as Ryan suggested: <embed SRC="http://localhost:7101/movies/transferservlet" WIDTH=100 HEIGHT=196 AUTOPLAY=true CONTROLLER=true LOOP=false PLUGINSPAGE="http://www.apple.com/quicktime/"> Any ideas?

    Read the article

  • Loading/Displaying large amount of data on webpage.

    - by jb
    I have a webpage which contains a table for displaying a large amount of data (on average from 2,000 to 10,000 rows). This page takes a long time to load/render. Which is understandable. The problem is, while the page is loading the PCs memory usage skyrockets (500mb on my test system is in use by iexplorer) and the whole PC grinds to a halt until it has finished, which can take a minute or two. IE hangs until it is complete, switching to another running program is the same. I need to fix this - and ideally i want to accomplish 2 things: 1) Load individual parts of the page seperately. So the page can render initially without the large data table. A loading div will be placed there until it is ready. 2) Dont use up so much memory or local resources while rendering - so at least they can use a different tab/application at the same time. How would I go about doing both or either of these? I'm an applications programmer by trade so i am still a little fizzy on the things I can do in a web environment. Cheers all.

    Read the article

< Previous Page | 4 5 6 7 8 9 10 11 12 13 14 15  | Next Page >