Search Results

Search found 56562 results on 2263 pages for 'gerald fauteux@oracle com'.

Page 815/2263 | < Previous Page | 811 812 813 814 815 816 817 818 819 820 821 822  | Next Page >

  • Should the English website use href="x-default" when it doesn't auto-redirect to the user's language or country?

    - by Noam
    For each URL on my site, I'm auto-redirecting according to header accept language. The site arch is English version: http://mydomain.com/page Spanish version http://es.mydomaina.com/page etc.. The english version is displayed unless I'm seeing a specific language other than en and that I support in the header, and then a redirect occurs. Google says this: For language/country selectors or auto-redirecting homepages, you should add an annotation for the hreflang value "x-default" as well: My pages aren't language selectors, nor are they the homepage. But I am auto-redirecting. My question is, should my english version be hreflang="x-default" or/and hrefland="en"?

    Read the article

  • Search the internet for a term selected in a text editor/viewer

    - by cipricus
    It seems possible (as I learned from a comment to this question made by user55822) to use Artha in order to search a term on the Internet by using "keybinding". What the aforementioned user says is that "you could search mouse selection anywhere by using an app like Artha, or by keybinding sh -c 'firefox "http://translate.google.com/#en/fr/$(xclip -o)"' (example given translate english into french)" A lot of things are obscure for me here. I could use Artha or keybinding (as alternative), or both? How to do that in each case? and "The command I gave you will open google translate in Firefox with your mouse selection (text highlighted) as a query. Use sh -c 'firefox "https://www.google.com/search?q=$(xclip -o)"' if you want regular Google over Google translate. And obviously, that would work everywhere" I do not know what to do with that formula. What should I do exactly? How to use it? In what application?

    Read the article

  • using curl command to download file in parts from different interfaces and to run the commands simultaneously from a script

    - by jsjain007
    i wanted to download a file using curl command simultaneously in different parts using ip aliasing(virtual Ethernet ports) so what i did was pasted the commands in a text file and run but the problem as obvious since it is in a file the commands will be executed one by one so is there a way to run all those commands simultaneously. here is the command curl --interface eth0:0 --range 0,38010880 http://wdl.cache.ijinshan.com/wps/download/Linux/unstable/kingsoft-office_9.1.0.4244~a12p3_i386.deb -o kinsoft-office.part1 curl --interface eth0:1 --range 38010880 ,- http://wdl.cache.ijinshan.com/wps/download/Linux/unstable/kingsoft-office_9.1.0.4244~a12p3_i386.deb -o kinsoft-office.part2 cat kinsoft-office.part*>kinsoft-office can anyone help me to run these above 2 commands simultaneously from the script so as to increase download speed

    Read the article

  • Redirect from the main page on multilanguage site

    - by Pluto
    I've decided to start developing multilingual web site. Now I'm thinking about the structure. The idea is to use a simple algorithm to find out what language is most interesting to the visitor when he makes his first visit to the root page, and then forward him to the corresponding address, for example, http://example.com/en/ for English-speaking audience and http://example.com/ru/ for Russians. The question is, how search engines react on such redirect? If this approach does not take away site's reputation, which redirects should I choose: 301 or 302? I would be very grateful for the reasoned response.

    Read the article

  • Webmaster tools showing 404 for non existent folder pages

    - by Jody
    Google webmaster tools is reporting some/many 404 urls that don't exist on my site. The links are things such as domain.com/xyz/ However that doesn't exist, but domain.com/xyz/index.html does exist. The "linked from" pages all show proper links to the "/xyz/index.html". The page without index.html DOES 404, but why is google even trying these urls if they are not linked to? My real question, is there a way to have google stop attempting to load these pages, and ultimately remove these from the crawl errors report. Thanks.

    Read the article

  • GA tracking utm query params after hashbang

    - by hybrid9
    We currently use a hashbang for the portion of our site that generates dynamic content which can also be deep linked. Our analytics team wants to use utm params to track the referral traffic from social networks. We are using Universal Analytics (analytics.js) as well as GTM. Will GA pick up the query parameters after the hashbang or does it always have to go before? For example: example.com/#!/some/content?utm_source=foo&utm_campaign=bar example.com?utm_source=foo&utm_campaign=bar/#!/some/content In #1, I'm concerned that the utm params won't be recorded and in #2 the page will break or the url could be incorrectly written. How does GA pull in those parameters - location.search? regex? Can I get away with using either?

    Read the article

  • error: you need to load kernel first

    - by Angelos318
    I made a clean install on my Sony Vaio laptop, of Ubuntu 11.10 and when the installation was ready, it prompted to remove the usb I was installing the distro from, and press enter to reboot. After this reboot the first thing I got was the following error: error: couldn't read file error: you need to load the kernel first Press any key to continue.. After that it throws me back to the Grub select screen: Ubuntu, with linux 3.0.0-14-generic-pae recovery mode previous linux versions (none since I made a clean install) memory test If i choose the first option it shows only a black screen and never loads anything. If i reboot the same thing happens. Could I repair this using boot-repair? Is there any other way? Note: I know nothing about linux code so i am a total noob on this one Update: boot-repair did not help Grub.cfg here: http://pastebin.com/GKLuDuhM Boot Info Script: http://pastebin.com/indARkKJ

    Read the article

  • My local ubuntu server is not updated

    - by Rakesh Manandhar
    I am from Nepal and it's faster to download apps from server for nepal using through ubuntu software center. But when i use np.archive.ubuntu.com server. There is no updated softwares and few softwares and nvidia driver are outdated. for e.g. i cannot even install openshot when i user server for nepal. It provides error which cannot be solved. I want to know how can the server for nepal be updated frequently with us.archive.ubuntu.com. i use ubuntu 12.10. and i want to use server for nepal since the download speed faster than 2x than from us server.

    Read the article

  • Is IE9 a modern browser?

    - by TATWORTH
    At http://people.mozilla.com/~prouget/ie9/ there is a very provocative article entitled "Is IE9 a modern browser?". There is a rebuttal by Tim Sneath at http://blogs.msdn.com/b/tims/archive/2011/02/15/a-modern-browser.aspx that is well worth a look. Certainly IE9 is already superior to its predecessors. My comment on the matter is that those that consider IE9 to be non-standards compliant, should submit tests to the W3C to demonstrate the non-compliance. Upon acceptance by the W3C, all the competing browsers can then be re-tested. I prefer objective tests to subjective opinion. I have used IE9 and on some sites such as Hotmail, it is noticeably faster. I have so far been unable to apply the promised IE9 lockout of spyware cookies. With Firefox, I just instal NoScript and never enable spyware sites.

    Read the article

  • Publishing Websites From VS.NET 2013 to Azure

    - by D'Arcy Lussier
    Originally posted on: http://geekswithblogs.net/dlussier/archive/2013/10/24/154459.aspxScott Gu recently announced the release of the Azure SDK 2.2. This includes a tonne of new features for accessing Azure resources from within Visual Studio 2013. You can read Scott’s blog post here. One feature is integrated Windows Azure sign in from within Visual Studio. I put a short video together showing how easy it is to publish a web application to Azure Websites. You can check it out here: http://youtu.be/eiuhJbwhZsQ (I couldn’t get it to display properly by embedding, definitely needs the wide-screen 720p view for best viewing) And yes, you can access James Chambers Karaoke Fansite live on Azure! Resources: Azure SDK 2.2 https://www.windowsazure.com/en-us/develop/net/ VS.NET 2013 Express http://www.microsoft.com/visualstudio/eng/downloads#d-2013-express

    Read the article

  • Link to pages on site without .html extension appearing in browser?

    - by Anime163
    I've modified my .htaccess file to allow access to html files without having to include the extension on the end, for example: www.mysite.com/document directs to www.mysite.com/document.html However, when I want to link to pages within my site using something like <a href="page.html"></a> I still get the .html appearing in the URL. So am I allowed to exclude the extension and leave a link as <a href="page"></a> so that the extension doesn't appear in the browser? Or is there a better way to do it?

    Read the article

  • Best way to redirect users back to the pretty URL who land on the _escaped_fragment_ one?

    - by Ryan
    I am working on an AJAX site and have successfully implemented Google's AJAX recommendation by creating _escape_fragment_ versions of each page for it to index. Thus each page has 2 URLs: pretty: example.com#!blog ugly: example.com?_escaped_fragment_=blog However, I have noticed in my analytics that some users are arriving on the site via the "ugly" URL and am looking for a clean way to redirect them to the pretty URL without impacting Google's ability to index the site. I have considered using a 301 redirect in the head but fear that Googlebot might try to follow it and end up in an endless loop. I have also considered using a JavaScript redirect that Googlebot wouldn't execute but fear that Google may interpret this as cloaking and penalize the website. Is there a good, clean, acceptable way to redirect real users away from the ugly URL if for some reason or another they end up arriving at the site that way?

    Read the article

  • Scrolling Box2D DebugDraw

    - by onedayitwillmake
    I'm developing a game using Box2D (javascript implementation - Box2DWeb), and I would like to know how I can pan the debug draw. I know the usual answer is - don't use debug draw, it's just for debugging. I'm not, however not all my objects are on the same screen, and i'd like to see where they are in the physics representation. How can I pan the debug drawing? As you can see the debug draw stuff, is show on the top left, but it only shows a small part of the world. Here is an example of what I mean: http://onedayitwillmake.com/ChuClone/ The game is open source, If you'd like to poke through and note something that perhaps i'm doing something that is obviously wrong: https://github.com/onedayitwillmake/ChuClone

    Read the article

  • Newly compiled (patched with -ck) 3.3 kernel won't boot up

    - by Shiki
    Screen is garbled. Reaaaally garbled. Like some sort of scrambling image from a movie. I can upload some pictures if needed. To be short with the description. Here is my 3.3 + ck .config file: http://pastebin.com/aP2F9RcH The dmesg log of the failed kernel: http://pastebin.com/esc2Rtka The computer is a ThinkPad T500 laptop. I have selected everything needed for the computer (as far as I can tell, but check the .config). The normal generic Ubuntu kernel boots and works perfect!

    Read the article

  • O'Reilly 50& off offer on CSS3 books to 05:00 PT on Oct/28

    - by TATWORTH
    Originally posted on: http://geekswithblogs.net/TATWORTH/archive/2013/10/21/oreilly-50-off-offer-on-css3-books-to-0500-pt.aspxAt  http://shop.oreilly.com/category/deals/css3.do?code=WKCSS&imm_mid=0b155e&cmp=em-prog-books-videos-lp-owo_css3_direct_wkcss, O'Reilly are offering 50% off a number of e-books on mastering CSS3 to 05:00 PT on Oct 28 "CSS3—the technology behind most of the eye-catching visuals on the Web today—is loaded with capabilities that once would have required JavaScript or third-party plugins, such as animation, pseudo-classes, and media queries. Use CSS3 to transform markup into stunning, richly detailed web pages that look great in any browser. For one week only, SAVE 50% on CSS3 ebooks from shop.oreilly.com and take your sites from ordinary to incredible."

    Read the article

  • PEX - are you licenced for it?

    - by TATWORTH
    There is an interesting artcile about PEX at http://blogs.msdn.com/b/ukmsdn/archive/2011/02/22/featured-article-pex-and-visual-studio.aspx PEX can be downloaded from http://research.microsoft.com/en-us/projects/pex/downloads.aspx The licence conditions are: "Pex and Moles are Power Tools available for commercial use for MSDN subscribers. Moles is also available separately for commercial use without requiring an MSDN subscription. Pex and Moles are also available for academic and non-commercial use." I note with interest that it is now available to MSDN subscribers. If I recall correctly it used to be only available to VS Team versions.

    Read the article

  • Removing surrounding noises from voice recording

    - by Peak Reconstruction Wavelength
    I have a wave file whose frequency spectrum looks like this. http://i.stack.imgur.com/2rRaS.png It contains audio, which I want to keep while removing the rest. The problem is that the surround noise changes, just those distinct voice patterns remain. I marked the voice patterns for clarity: http://i.stack.imgur.com/eLkBl.png What could an algorithm look like / a workflow in adobe audition look like that removes everything but the voice patterns? I think that the main characteristic is the line-shaped form over time. Loudness alone is not enough as the noise is loud aswell.

    Read the article

  • I can not download anything

    - by Jason Machen
    I am very new to ubuntu but decided to wipe my windows 7 and install it. I can not download anything from the software center. This is the error message I get. I can use the web in all other ways including this site. What can I do? Thanks, Jason W:Failed to fetch http://security.ubuntu.com/ubuntu/dists/raring-security/main/source/Sources 404 Not Found [IP: 91.189.91.13 80] W:Failed to fetch http://security.ubuntu.com/ubuntu/dists/raring-security/restricted Plus about 20 other lines.

    Read the article

  • Responsive Menu Navigation [closed]

    - by Aaron Brewer
    I am sure you all have heard of Responsive/Adaptive Website Design and Development, but for the sake of beginners to the technique and skill, what are ways to create a Responsive Menu Navigation? I know there are a few standard ways, hence: JavaScript/jQuery Menu that changes functionality at different breakpoints. CSS3 Menu that changes functionality at different breakpoints. If you have had the opportunity to create a Responsive Menu, what technique did you use? How did you do it? Do you have an example? Did your Menu change functionality at different break points? To read up on Responsive and Adaptive Design: http://johnpolacek.github.com/scrolldeck.js/decks/responsive/ To read up on Responsive and Adaptive Design Menus: http://blog.usabilla.com/10-tips-how-to-handle-responsive-navigation-menus-successfully/ I hope this will save Pro Webmasters plenty of duplicate questions.

    Read the article

  • Is adding in the header the license type enough to say: "my code is licensed"?

    - by silverfox
    I read on various sites about licenses. I did just put the license type in the header file (in my case a javascript file, open-source): /* * "codeName" "version" * http://officialsite.com/ * * Copyright 2012 "codeName" * Released under the "LICENSE NAME" license * http://officialsite.com/LICENSE NAME */ javascript code ... In the same folder I leave a copy of the license. The listing of the folder looks like this: * codeName.js * LICENSE In the file LICENSE is the full text of the license my code uses. What I cannot find anywhere that says is this is enough to say my code is licensed (the case of open-source). Is something more required?

    Read the article

  • Do search engines rank internal redirects negatively?

    - by siverd
    A client is in the late stages (code complete) of a website redesign and unfortunately hasn't implemented 301 redirects to point high traffic pages to the new URL's. As I understand it our only option at this point is to create redirects within the CMS. Our CMS allows us to do this: www.mysite.com/category/current-page.html will redirect to www.mysite.com/new-category-name/new-page.html The site now uses custom logic on our 404 page to check this list of redirects and if one exists forwards the user to the new-page.html I understand that using 301 redirects would be the correct way to maintain our page rank but I think that would require a code change which isn't possible. Question How will search engines respond to this? Will they wait until the redirect happens and allow us to keep our page rank (authority, trust, etc) or will they see the 404 page and down-rank us? Worst case...will they make our new-page.html start from a rank of "0"? Thanks for your help.

    Read the article

  • How can i distinguish revenue from different folders in Adsense?

    - by tarrasch
    I am currently running three websites using adsense. Since i do this more to be useful than to earn something, i have a subdomain of a free hoster and my websites live in subfolders e.q. mydomain.hoster.com/website1, mydomain.hoster.com/website2, to which i link with .tk aliases. Is there a way to distinguish adsense income from website1 to website 2? I tried using custom channels, but somehow this part of adsense eludes me. Can someone shed some light on this?

    Read the article

  • Best stats tool for cross-domain traacking

    - by kidbrax
    We build a webapp that allows users to run the app under their own subdomain. So we run the app under search.domainX.com, search.domainY.com and so on. They each have their own Google Analytics to track individual stats. But we want to know what general traffic for all clients of our app. So we want to know stuff like "among all our clients we had x number of views." What is the best way tool to track that sort of thing.

    Read the article

< Previous Page | 811 812 813 814 815 816 817 818 819 820 821 822  | Next Page >