Search Results

Search found 9960 results on 399 pages for 'iwork pages'.

Page 180/399 | < Previous Page | 176 177 178 179 180 181 182 183 184 185 186 187  | Next Page >

  • Serverless Web Application

    - by Andrea Di Persio
    In my company we work on a software that produce reports in html format. My bosses love the fact that static html pages can be moved across computer simply by moving/copying a folder and no web server is involved, so the customer only need a browser. The problem is that they asking me to implement a lot of feature which is very hard to implement properly and in a clean way without an application server. Frames cross domain problem, the impossibility to work with GET and POST data, no URLs routing...is very hard to work with this limitations. Anyone had similiar experience and wants to share their tricks/suggestion ? Do I need to tell my boss 'there is no future without a web server'? Regards.

    Read the article

  • Sense of "stop on..." stanza when job is a task

    - by Binarus
    Hi, an upstart question (I think I have read all relevant man pages but could not find the answer there): What is the sense of using a "stop on ..." stanza in the definition of a job which is a task? The manuals tell us that such a job, after being started, just waits until its script (or exec stanza) is executed completely, and then stops automatically. Given that, what is the point in using "stop on ..." stanzas in such job definitions? For example, this is the job definition for Upstart's (very important) rc job in Natty 11.04 (leaving out comments and empty lines): start on runlevel [0123456] stop on runlevel [!$RUNLEVEL] export RUNLEVEL export PREVLEVEL console output env INIT_VERBOSE task exec /etc/init.d/rc $RUNLEVEL IMHO, the job, after being started by a runlevel event, will be stopped automatically as soon as /etc/init.d/rc $RUNLEVEL has finished. Thank you very much for any explanation!

    Read the article

  • How good is it for SEO if you have a widget that lives on other sites?

    - by Genadinik
    I made a widget that would ideally live on other sites. Here is an example: http://www.comehike.com/outdoors/widget.php?type=hike&hike_id=108&width=500&height=500 I guess since the widget would link back to me, it would be an SEO boost for my site. Is that correct? Or would it be just an SEO boost for that particular URL? If it is just an SEO boost for the particular URL, it does me little good since that page does not link to any of my other pages. Am I thinking about this correctly? How are these things typically handled so that there is a benefit to my site's SEO?

    Read the article

  • Can we 301 redirect to a new page, but still publish the old content somewhere else?

    - by KBS
    We have a page on the site which ranks well for an SEO term (top 5) but contains old information. We have added a new page but Google doesn't rank it that well. Information on these pages is time sensitive. Old: example.com/2013-related-information.html New: example.com/2014-related-information.html Obvious solution is to delete old page and do a 301 redirect to the new page. Now, can we still keep the old page by giving it a new URL. example.com/2013-related-information.html is redirected to example.com/2014-related-information.html example.com/2014-related-information.html is recreated with a new address such as example.com/new-2013-related-information.html What we are trying to do is to send the user to the fresh page but still not destroying the record copy if someone wants to go and dig up the old information.

    Read the article

  • alx driver not working properly

    - by user292768
    My alx driver seems to be working fine for everything except apt-get. I can navigate to web pages, the ping command works just fine, but whenever I try to do apt-get update or apt-get install it hangs on trying to connect to whatever archive it's trying to pull info from. There's no error message, nothing to imply that it can't connect, it just hangs there and does nothing at all. I've tried re installing alx, upgrading/downgrading my linux kernel, downloading source and building alx from scratch, etc etc but nothing works. Does anybody else have this problem/know how to fix it?

    Read the article

  • XDIME for Mobile Applications

    - by Carlos Gavidia
    I'm involved in a project that requires to mobile-enable some previously developed Portlets. The Portlets are deployed in WebSphere Portal, and the container offers a technology called IBM Mobile Portal Accelerator that uses XDIME to render mobile pages according to the device. I'm trying to document myself in the technology and I'm having a bad time: Google only shows some outdated sites from IBM and even older posts from Volantis, another company involved in the technology (Amazon shows no related books). So... what's the current status of that technology actually? Is has some decent level of adoption?

    Read the article

  • Compiling old versions of R (2.11 or older) on Ubuntu 13.10 (or newer)?

    - by user283293
    I am trying to compile an old version of R (2.11.1, specifically). "Why?", you might ask. Well, I have a program I am trying to run (that was not written by me, but I am hoping to use in my research) that was written in "old" R, and the program writer has specifically mentioned that R-2.11.1 or older must be used. I have run ./configure with no obvious errors (the entire output being far too long to post here, but I have been warned that I cannot make PDF, DVI, or HTML versions of the help pages, which I'm not too concerned about). Next, when I pass the make command, I am met with the following error (after a long stream of successful commands): Warning in solve.default(rgb): unable to load shared library '/home/sbowerma/Programs/R-2.11.1/modules//lapack.so': /home/sbowerma/Programs/R-2.11.1/lib/libRlapack.so: undefined symbol: _gfortran_compare_string Error in solve.default(rgb) : lapack routines cannot be loaded Error: unable to load R code in package 'grDevices' Execution halted However, I have confirmed that the library lapack.so is, in fact, in the exact position it is supposed to be.

    Read the article

  • Find out when a new domain appears in search results

    - by TerryB
    Does anyone know a way to perform the following: I want to know whenever a new domain starts appearing in the google search results for a particular query. For a given google search query, I'd like to receive an alert whenever a new domain pops up and starts appearing in the search results for that query. Alternatively, it would be great if you could just sort google search results by the age of the domain, making it easy to find new sites. As far as I can tell you can only sort by when the page was "last updated". Is something like this possible? EDIT: Following John's suggestion of Google Alerts. The problem with Google Alerts is that it sends you any new PAGES appearing in the search results, not just new DOMAINS.

    Read the article

  • SEO: Is promoting your backlinks a good strategy for improving search results for my site's name?

    - by user4394
    I run a website that's been around for about three years in the sports space. I am successfully ranking well for targeted keywords, but searching for the name of my site itself returns very poor results - it shows my site, its FB/Twitter, and then 15 pages of unrelated spam that happen to contain two words that, when combined, form my website's name. After that, my backlinks begin to show up spordically. As far as I can tell, I simply don't have enough backlinks and the backlinks I do have are ranked worse than the spam. (Site Explorer lists 200 external links to any page on our domain and 20 external links directly to the front page). To counter this, my strategy is to promote my backlinks so they get a better page rank than the spam. Does that make sense? Am I going in the right direction or should I just focus on getting more backlinks pointing directly to my site? Thanks in advance and I'd be happy to answer any questions I can (without giving away my site of course).

    Read the article

  • iGoogle : Google met fin à son Netvibes maison et à plusieurs autres services, allez-vous être impacté ?

    iGoogle : Google met fin à son Netvibes maison et à plusieurs autres services, allez-vous être impacté ? La désormais traditionnelle opération d'abandon de certains de ses projets par Google afin de rationaliser sa gamme de produits continue. La société vient d'annoncer la fermeture de plusieurs projets qui n'ont pas obtenu le succès escompté ou dont des alternatives ont été développées, dans le cadre de son nettoyage de printemps pendant l'été. C'est donc un ensemble de cinq produits qui seront mis aux oubliettes parmi lesquels figure la solution iGoogle. Lancé en 2005, le portail d'accueil iGoogle permet aux utilisateurs de créer des pages personnalisées, contenant...

    Read the article

  • Google Analytics for subdomains

    - by Sebastian
    I have two WordPress multisites under one domain - city-x.domain.com and city-y.domain.com. domain.com is a landing page where you select your city, and a cookie will redirect the user to that city on subsequent visits. I'd like to be able to track the number of hits on all pages on domain.com, city-x.domain.com and city-y.domain.com separately and combined. How is this On a side note, I've heard that GA underestimates hits. As this is important for advertising purposes, is there a better free service?

    Read the article

  • Using GDI+ vs HTML page

    - by Umesha MS
    I am working on an application which is UI intensive (i.e. we need to customize all the control to look different). For this is I planned to use GDI+ in win32. But one of the suggestion form our team member is, Instead of using GDI+ he asked me to have a web browser control, in that he asked me to display local html.That html page contains customized controls. Form his suggestion I created a web control and used some UI intensive local html pages. That looks good. But my question is, is it is a good idea to use web control to solve this?

    Read the article

  • Javascript Only Search Method [on hold]

    - by user2118228
    I need to put a search function on a website that is going to be on a CD-ROM with no access to the internet. It has 80 pages, and about 500 'items', so I'd prefer to not have to hard code 100's of 'if statements if possible. I've found a few programs you can buy that will index and generate results (Zoom Search, JSS Index, The German Guys') but there are odd quirks with each one. Plus I would rather code it myself to get complete control over it, and to really understand what it's doing. Basically searching for a few words would display the product image and description; clicking on that would take you the related URL. This is kind of complicated, I can't find an easy solution not dealing with hundreds of if Statements. Has anyone ever created anything like this or know a better method? I'm not really sure a better way to go about this. I've used PHP/MYSQL for search results before, but this cannot run any php.

    Read the article

  • How to handle CNAME host redirect to virtual directory?

    - by esac
    I have an internal website and virtual directory http://server2012/logs. I created a CNAME on my DNS server as LOGS - server2012. I would like to set it up so that http://LOGS redirects to http://server2012/logs. Ideally, I would still want it so that all pages appear in the browser as being off from the LOGS URL. So http://LOGS/network.html?site=32 is what is displayed in the browser, but it is really being served from http://server2012/logs/network.html?site=32. I've looked at URL rewrite, but can't seem to get to work.

    Read the article

  • More complex learning source for C# .NET [closed]

    - by Leron
    By complex I don't mean more difficult but including a larger area of possibilities cover. I've started a few years ago with PHP and the transition from learning the syntax of the language and the basic logical structures to working with databases, including JavaScript and so on was very short. And now I'm more interested in studying working with languages like Java/C#. Recently I spend a lot of time reading and writing some simple console applications, I've read almost 2K pages for C# programming and still don't know how to connect to database for example. Just for info I'm interested in web development, socket programing and live streaming, don't know if I'm exceeding myself too much writing that but despite that I want to find some books/internet sources where I can extend my current knowledge of C#/.NET, start using database queries, maybe try something more complicated webwise.

    Read the article

  • SQL Live Monitor

    - by TiborKaraszi
    I just found this one out there and wanted to share it. It connects to an instance and show you a bunch of figures. Nothing you can't extract yourself with SQL queries, but sometimes it is just nice to have one tool which is very easy to use. Here's what it looks like when connecting to an instance with no load on it: As you can see, there are some hyperlinked pages as well, and there are also some interesting options (like logging to CSV or for PAL analysis) under the "Option" button. One more thing...(read more)

    Read the article

  • The life saver HttpContext.Current.Items["ParameterName"]

    - by MoezMousavi
    I got stocked passing parameter to one master page for some reasons, seems the page lifecycle and dynamic loading of the master pages has got some issues with defining public properties in the masterpage within my project. It did not set my values and as a result, properties became useless. A collegue just mentioned using HttpContext. have a look what  MSDN saying "Encapsulates all HTTP-specific information about an individual HTTP request" http://msdn.microsoft.com/en-us/library/system.web.httpcontext.aspx HttpContext.Current.Items["ParameterName"] Also, Page.Items could do the same thing. Page.Items, "Gets a list of objects stored in the page context" http://msdn.microsoft.com/en-us/library/system.web.ui.page.items.aspx as your master page and content page are rendered as a single document anyway.

    Read the article

  • wayfinder, paginator, or breadcrumb?

    - by LOTUSMS
    My question is fairly simple. I've heard one or two people call it a "wayfinder" and I personally call it "breadcrumb", then again "paginator" can also fit the bill. Bottom line is I can't seem to find what I need because I am not coming up with the right term. What I want is the following: I have a site with a multiple-pages form. I want my users to see the steps through the form. I searched for all three terms and nothing I can use has shown. I could build one from scratch but I am interested in what is out there. Someone may have done a better job than I will be doing. Any idea on what to call this and if you have a link I can check out, that will be bonus Thanks

    Read the article

  • A drop in SERP after following webmaster guidelines [on hold]

    - by digiwig
    So here's a puzzle for all you SEO gurus out there. I recently launched my own site. I had target keywords which were ranking very well for about 1 month, within the top five and even appearing in first place. In an attempt to maintain good positioning, I followed guidelines by adding robots.txt, an xml sitemap redirecting non-www to www redirecting index.php to root domain adding htaccess 301 redirect for old pages I added rich snippets created a google+ account, verified my picture to appear, I went through each of the webmaster issues with duplicate titles and meta descriptions and improved header tag document outlines i even created a few more blog posts to keep the content freshing and moving. So now my website appears on page 2 with my target keywords - and all because I followed the guidelines. What is happening? I see competitors with stagnant content superglued to position 1.

    Read the article

  • Why don't my Google Analytics custom segmentation visit numbers match up?

    - by Hates_
    I have three main areas of my site and want to track total usage as well as breakdowns of the three parts. I am trying to segment the "type" of use on each page using a custom variable as such: ['_setCustomVar',1,'Visitor Type','Unknown',1] Visitor type can be one of three values: "Unknown", "Reader" or "publisher". Every page has this value set. Now when I look at my analytics chart and chose all three segments, the individual values do not match the sum. I've double checked the pages to make sure the custom var is there.

    Read the article

  • How to specify ingredients in Schema.org (other than recipes or drugs)?

    - by guillem
    I'd like to use Schema.org for a series of product specs pages. The products are cosmetics (shampoos, creams, etc.) and ideally an INCI declaration will be included for every product. I'd like to use Schema.org in the declaration, but I cannot find a suitable category in Schema.org. Ingredients is meant for cooking recipes, and ActiveIngredient for (medical) drugs and dietary supplements. Did I overlook some part of the tree? Any ideas for a workaround? Or am I trying to do something out of scope?

    Read the article

  • How to fix: "The connection to the server was reset while the page was loading."?

    - by stupidity
    I have a 64-bit Ubuntu 12.04 installed. I used to login to a website (chess.com) but since a few days ago I get this message: The connection was reset The connection to the server was reset while the page was loading. The site could be temporarily unavailable or too busy. Try again in a few moments. If you are unable to load any pages, check your computer's network connection. If your computer or network is protected by a firewall or proxy, make sure that Firefox is permitted to access the Web. This is the case in Firefox, and chrome. A friend of mine confirmed to me that the website is working fine and he could access it. And I can access many other websites and facing no issues except for this website. I tried switching the use TLS 1.0 option as suggested by someone but this didn't fix it. Can someone suggest a way to fix this issue?

    Read the article

  • Weird URLs being access by Googlebot

    - by Avishai
    Lately I've been seing all sorts of strange URLs show up as errors in my Webmaster tools account, but they're URLs that don't actually exist on my site, nor are linked from the pages that Google claims they're linked from. URL Response Code Detected yR3kna/5RfA4+ndtn/X4zcevudMlXbqbIrnPbH9irw= 404 9/16/12 OK4iaOVdr6Ocjmz+u1kuR5Q486mhDo/e45nwjl2+y8= 404 9/9/12 pxGz/oHEA0BS8U3VFBzJcZnnIHMsFXb3/rIxMxh2ws= 404 9/16/12 Af8tbvQ0HniIpf53I8Txz1hM1/JxxrFQxgqPuErWII= 404 9/9/12 7Bk7c0LDmm4PHyTjml017EGwNNPCn/p/0xMSWWPDic= 404 9/16/12 umCwnDvTE8ybpUB19MIb+VRj5xRJncyYGGfAQ2Mxn0= 404 9/1/12 # etc... Do you know how to make these stop? It's not at all clear to me why it would be going to these URLs in the first place.

    Read the article

  • mysql questions for beginners

    - by ankhseeker
    ok, I have a few questions regarding mysql. I am currently running ubuntu 12.04.4 LTS command line version. I am looking for a database that I can use. I am confused at this point because I am uninformed. mysql is just one database that is on the server? or can it contain several or many databases What programs do I use to access it on the server or is it a vt-100 type access? I understand that mysql comes with lamp? or ubuntu. I am thinking that it is already installed but not sure how to access it, but that is another question for later. Outside of the man pages and the ubuntu manual, is there a site for its setup and use? Thanks!

    Read the article

  • Release Notes for 4/6/2012

    Here are the notes for this week’s release: Fixed an issue where authenticating against Mercurial using capital letters would fail. Fixed an issue where comments in pull requests were not word wrapping correctly. Fixed an issue where usernames would overflow their allotted space in discussions. Fixed an issue that would cause unexpected errors on the source code pages if you created a Git and Mercurial fork with the same name. Have ideas on how to improve CodePlex? Visit our ideas page! Vote for your favorite ideas or submit a new one. Got Twitter? Follow us and keep apprised of the latest releases and service status at @codeplex.

    Read the article

< Previous Page | 176 177 178 179 180 181 182 183 184 185 186 187  | Next Page >