Search Results

Search found 9935 results on 398 pages for 'pages'.

Page 170/398 | < Previous Page | 166 167 168 169 170 171 172 173 174 175 176 177  | Next Page >

  • Update drivers for TL-WN851ND

    - by Tony_GPR
    Today I bought a new PCI wireless card, TP-Link WN851ND, with Atheros AR9227 chipset. It has 2 antennas and is compatible with Wifi N so I thought it would improve the quality of the signal. But after install it on my computer the result is the opposite to expected. It doesn't connect to my network, while my old Wifi BG card connect without problems, I created an access point from my smartphone to try the card, and work, but is very slow loading pages. In Windows 7 works perfectly, so I think the problem is the driver. I have Ubuntu 12.04 LTS with kernel 3.2.0.31, is there a way to update the driver or I can apply a patch to improve the performance of the card? Otherwise, anyone know if there is a work in progress to improve compatibility with this chipset, or is it better to change the card and buy one with better driver compatibility. And finally, which wireless N compatible chipsets have good support under Linux/Ubuntu?

    Read the article

  • convert video file to .ogg

    - by Levan
    I've been having trouble with this because I'm new to Linux: I would like to convert different video formats to ogv. I found some terminal commands like this: ffmpeg -i input.avi -acodec libvorbis -ac 1 -b 768k output.ogg The problem with these type of commands is that they are intended to change bit rate, fps, or even resolution. I would like to just change the file format without changing anything else about the video. I looked at the man pages for ffmpeg and found some useful info but I don't know how to space command-line options. Are there any easy ways to do this? In addition, is there a command to change the bit rate so that it doesn't go over a certain rate?

    Read the article

  • Interactive Reporting Translation Workbench utility is available

    - by THE
    As you may have seen in our  Newsletter, Oracle has released the "Oracle Hyperion Interactive Reporting Translation Workbench" for Hyperion Interactive Reporting (IR) customers who are moving to Oracle Business Intelligence Enterprise Edition (OBIEE). A summary for this utility can be found  here. To get the Utility along with documentation and training material we suggest that you visit the Oracle Technology Network ( OTN ) "Oracle Hyperion Interactive Reporting Downloads" page. Friendly enough, instead of hundreds of pages of "getting started Docs", Oracle has packed some training videos into the downloads, so that getting started is made as easy as possible. But of course the documentation comes with it as well.

    Read the article

  • Unleash Oracle Premier Support's Advanced Proactive Capabilities

    - by swalker
    Where do you go to solve technical problems? Better yet, where do you find out how to prevent them? Oracle Premier Support's proactive capability portfolio can help you prevent, resolve, and upgrade. Join thousands of Oracle customers and partners who are already taking advantage of proactive support. Are You Ready To Get Proactive? Bookmark the proactive capabilities portfolio and start exploring Oracle Premier Support's proactive support capabilities. Search "Get Proactive" in My Oracle Support to view the knowledge, tools and communities available through product specific pages. Act now to get started! Questions? Contact Oracle’s "Get Proactive" team today.

    Read the article

  • Doubts regarding the behaviour of 'autoremove' command and '--auto-remove' flag

    - by Jasper Loy
    After reading several man pages and forums, I thought that running 'apt-get autoremove' without any following argument removes all unused dependencies left on the system, while running 'apt-get autoremove xxx' removes xxx together with its unused dependencies. However I found this to be not true. Running 'apt-get autoremove xxx' not only removes xxx together with its unused dependencies, it also removes all other unused dependencies. So I tried to run 'apt-get remove --auto-remove xxx', thinking that this would remove only xxx and its unused dependencies. To my surprise, this also removed xxx, its unused dependencies and all other unused dependencies. Is this the intended behaviour of the commands or a bug? Is there any quick way to remove xxx and its unused dependencies without removing other unused dependencies?

    Read the article

  • Pagination and duplicate content

    - by jazz090
    I have an archive page that displays the number of articles published. Because there were so many, I ran a pagination script: for 127.0.0.1/archive/2/?p=x&pp=y where p is the page number and pp is number of articles to display per page. The pagination looks like this: Prev 1 2 3 4 ... 12 NEXT with each item linking to p like <a href="?p=x">x</a>. I also have the items per page setter: 25 | 50 | 100 (<a href="?pp=y">y</a>). Now I have a PHP script that fixes pp into a session variable. But I am worried about duplicate content (since incrementing pp values will be inclusive) and also content not getting indexed because its not in the pagination link. so in the example above, pages 5-11 will not be indexed. Any ideas on how to fix this?

    Read the article

  • Information about links disappeared from Webmaster Tools

    - by Bobrovsky
    I discovered that all information about links to my site disappeared from Google Webmaster Tools. Last time I checked the "Links to your site" page in GWT there was nice list of linking domains and all. But now there is only "No data available." There were no changes to the site contents. Why could it be? And what can I do to fix this? About a month earlier I found that PR of all my pages dropped by 2 points. May these changes be related?

    Read the article

  • Un nouvel Android Market Webstore est en ligne, il permet de télécharger une application sur son terminal directement depuis le Web

    Un nouvel Android Market Webstore est en ligne, il permet de télécharger une application sur son terminal directement depuis le Web Google vient de dévoiler son Android 3.0, Honeycomb. Mais ce n'est pas tout, une autre nouveauté de taille a été présentée : l'Android Market Webstore. Celui-ci permet de trouver et d'installer des applications, sans être obligé de passer par le Market de son téléphone, comme c'était le cas jusqu'à maintenant. Seul le navigateur est désormais nécessaire, "tout est connecté", pour une simplicité maximale. Ère du social oblige, une fonction permet d'envoyer par e-mail des liens vers les pages du Market de son choix, et aussi de partager des liens vers des apps via Twitter. ...

    Read the article

  • Which prediction model for web page recommendation?

    - by Nilesh
    I am trying to implement a web page recommendation wherein registered users will be given a recommendation of which page to visit depending upon the previous data.So with initial study I decided to go on with clustering the data with rough sets and then will move forward to find out the sequential patters with the use of prefix span algorithm.So now I want to have a better prediction model in place which can predict the access frequency of pages.I have figured out with Markov model but still some more suggestions will be valuable.Also please help me with some references of the models too.Is it possible to directly predict the next page access with the result of PrefixSpan.If so how?

    Read the article

  • Aliasing resellers domain to primary domain

    - by Ashkan Mobayen Khiabani
    I have designed a website that accepts re-sellers and actually the concept of this website is having local re-sellers for each province (or should we say branches). I have designed this website in a way that anybody who has a domain, can point to our website (a record or cname). well most of the website content are the same, the only difference is that re-sellers website doesn't have some items on the main menu and may have some small descriptions of their own branch in some pages. I read that Google may ban websites with duplicate content (or which are significantly similar). I want to know will this be a problem for me? If yes, what else can I do? we have had considered asking our reseller to use iframe that loads our website but wanted that each reseller can have its own SEO and try harder but what I read about this duplicate thing worries me.

    Read the article

  • I want to learn how to help or contribute to Unity

    - by user17953
    I am a college student studying computer science and would one day like to work on operating systems. The part of ubuntu that really interests me is Unity and I would like to study it and possibly contribute to it. I was reading the wiki pages about it and it was saying to get a copy of all the required components and then start hacking. Would it be wise to do this in a virtual machine with ubuntu on it? Do you have any advice on a good place to start? Do you know of any common pitfalls? Should I also post in the irc for this? Thanks

    Read the article

  • Strange behavior of flash in Google Chrome

    - by Bakhtiyor
    I have the last version of Chrome 5.0.xx. and my Ubuntu is also of the last version 10.04. What I usually do is to open two instances of Chrome and divide the screen into two parts. In one part my son watches cartoons in Youtube and in the other part I just read some news. So, sometimes, when I close some pages of the news I have being read the video which was being played in youtube crashes and immediately stops. I need to refresh youtube page and see the video again. What is the problem? How to solve it

    Read the article

  • Constructor should generally not call methods

    - by Stefano Borini
    I described to a colleague why a constructor calling a method can be an antipattern. example (in my rusty C++) class C { public : C(int foo); void setFoo(int foo); private: int foo; } C::C(int foo) { setFoo(foo); } void C::setFoo(int foo) { this->foo = foo } I would like to motivate better this fact through your additional contribute. If you have examples, book references, blog pages, or names of principles, they would be very welcome. Edit: I'm talking in general, but we are coding in python.

    Read the article

  • URL Rewrite http to https EXCEPT files in a specific subfolder

    - by BrettRobi
    I am trying to force all traffic on my web site to use HTTPS, using the URL Rewrite 2.0 module added to IIS 7.5. I got that working and now have a need to exclude a couple of pages from using SSL. So I need a rule to rewrite all URL except those referencing this folder to HTTPS. I've been banging my head against the wall on this and am hoping someone can help. I tried creating a rule to match all URL except those in a nossl subfolder as in this example: <rule name="HTTP to HTTPS redirect" enabled="true" stopProcessing="true"> <match url="(/nossl/.*)" negate="true" /> <conditions logicalGrouping="MatchAll" trackAllCaptures="false"> <add input="{HTTPS}" pattern="off" /> </conditions> <action type="Redirect" url="https://{HTTP_HOST}/{R:1}" redirectType="Found" /> </rule> But this doesn't work. Can anyone help?

    Read the article

  • Firefox va intégrer Weave en natif pour proposer la synchronisation cryptée et anonyme par défaut :

    Mise à jour du 25/05/10 Firefox va intégrer Weave en natif Pour proposer la synchronisation cryptée et anonyme par défaut Weave est le premier « service Mozilla ». Il s'agit en fait d'une extension qui permet de synchroniser les données de Firefox ? comme les marques-pages, les mots de passe, etc (lire ci-avant) - entre plusieurs machines. La Fondation Mozilla vient de décider d'inclure ce service en natif dans son navigateur. Rien de nouveau sous le soleil par rapport à la concurrence ? Oui et non. Non, parce qu'il est vrai que certaines des concurrents de Firefox le proposent déjà. ...

    Read the article

  • Printing in booklet fomat

    - by To Do
    In the past I had a printer that had the booklet format option that would print the document two pages at a time ordered in a way that folding the whole printout would produce a booklet which could be stapled in the middle. My current printer does not support this feature. I would like to use some utility or script to convert an A4 pdf to an A5 booklet one. I found this page but I'm getting an error : pdfjam ERROR: can't find pdflatex! How do I get past this issue? Does anybody know of any other solution? I'm on Ubuntu 12.10 AMD64

    Read the article

  • 301 Redirects for regional variants of a homepage

    - by Adam Jenkin
    I am planning on implementing a website which has regional homepage variants. For Example: mycompany.com/europe mycompany.com/us The rest of the site is region agnostic and content will continue such as: mycompany.com/news mycompany.com/about-us etc For homepage (.com) requests, I plan on redirecting users to the correct homepage variant (via 301). If I cannot determine the correct one, I will fallback to redirecting them to the US homepage (/us). From an SEO point of view, firstly is this ok? or should I be doing anything additional to this for making search engines aware of the regional differences? As crawlers are region agnostic, I plan on directing them to the US page with a 301, or should I have something on the .com page which they use? Being that the regional homepage's will likely be the most visited pages, they should show up in result sitelinks when searching for mycompany (which I think is a good thing). Apologies for the slightly open question - I know anything SEO related is more opinion/best practice than fact but am purely looking for advice.

    Read the article

  • Ubuntu 12.04 LTS Desktop 64 bit user permissions or apache2 rewrite problem

    - by mtm
    have installed 12.04 Desktop 64 bit, manually installed LAMP, phpmyadmin, php5-dev,PEAR, PECL, apc, ssh, created user to own /var/www/ and transferred 3 sites to /www/. sites are in subfolders, sites - available all configured, and enabled. One site is pure html, two athers - php. Enabled curl, but phpmyadmin started at first, also php sites, than stopped working /show blank pages/ sites said Clean urls cannot be enabled. Html site still working. Where is the problem, and why the php sites stop working? In all apache .conf files Allow Override is set to ALL. php sites have .htaccess files. And this configuration worked with Ubuntu 10.04.

    Read the article

  • Google Sites page never shows up in Google Search organic results?

    - by gus
    I use Google Sites (i.e.: https://sites.google.com/site/EXAMPLE/ ) as a convenient way to maintain up-to-date info on several residential properties, info that's often requested by my property agents, its been around for about 1 year, but I still can never get it to appear in organic Google search results or Bing, even if I search the specific keywords such as the street names. I submitted the URL manually to search engines, knowing that my Sites page probably has very few incoming links. Is this expected behavior? The content of my page has simple formatted text, and outgoing links to Picasa/G+/imgur photo albums. Am I doing something wrong or do all GoogleSites pages have poor organic search rank? Thank you very much.

    Read the article

  • Authorship-verified website not included in "Author Stats" of Google Webmaster Tools?

    - by Yosi Mor
    In Google Webmaster Tools, is it normal for a website for which the Structured Data Testing Tool shows that "Authorship is working for this webpage" -- to not be listed in the "Author Stats" section (under "Labs")? I already understand that successful verification using the Structured Data Testing Tool does not guarantee that Google will actually display authorship in the SERPs, and that Google decides this at its own discretion. However, shouldn't such successful verification at least guarantee that the website is included in the "Author Stats" section (which purportedly covers "pages for which you are the verified author")? I would have assumed that, if Google is not yet displaying authorship for that site, it would show both its Impressions and Clicks as being "<10". Are my assumptions incorrect?

    Read the article

  • navigation menus and SEO

    - by Rodolfo
    I've always have my doubts about navigation menus effect on SEO. You know, the vertical menus on the top that show in every page in the site linking to main sections and subsections. My issue is that if not done dynamically (i.e. after page is loaded or something), from a search engine's point of view it probably looks like a whole bunch of links in the beginning of the page, and links that probably have nothing to do with the page being analyzed, so it's probably not only confusing it, but also giving link 'juice' to the wrong pages or reducing its value. When I've asked SEO people about this, I usually get a "Google is smart, they'll recognize it as a menu and ignore it" response, but I'm not convinced (and the 'Google is smart' argument sounds almost like religion discussion to me). So does it affect SEO negatively or not? Are there any official posts on this topic?

    Read the article

  • Ensure Future Browser Compatibility - SharePoint Branding

    - by KunaalKapoor
    HTML and Future Internet Explorer Compatibility with SharePointAs new versions of Internet Explorer are released, the way HTML is rendered by the browser could change over time. To address the possibility of changes, Microsoft uses the X-UA-Compatible META tag that targets HTML markup to a specific version of Internet Explorer. The default SharePoint 2010 master pages are set to force current and future versions of Internet Explorer to render HTML in Internet Explorer 8 mode like the following markup:<meta http-equiv="X-UA-Compatibile" content="IE=IE8" />The Adventure Works Travel HTML includes the META tag to help ensure future Internet Explorer versions will display the SharePoint HTML properly.For more information about the Internet Explorer Standards Mode, see META Tags and Locking in Future Compatibility.

    Read the article

  • Mouse wheel scrolls down, but not up?

    - by HDave
    I've set up a new machine with Ubuntu 12.04 (x64) and have notice that the mouse wheel scrolls down fine, but when I scroll up, it pages down! I've run xev and have verified that scroll up is registering as button 4 and scoll down is button 5. So everything there is fine. I think that button 4 is somehow mapped to page down instead of scroll up. Any idea how to fix? NOTE: I've read the Ubuntu instructions on how to use imwheel to change mouse button mappings, but it doesn't work on 12.04. See here: https://bugs.launchpad.net/ubuntu/+source/imwheel/+bug/1004812

    Read the article

  • How can I clone or mirror a site without SEO penalties for duplicate content?

    - by Amanda
    I am a web developer and I want to create clones of the sites I've developed for clients, so that I have an "original copy" on a subdomain of my own website, so that I can showcase my work to new clients. What is the best way to not get my clients original websites penalised for duplicate content? I am planning to have a robots.txt file that disallows all robots, as well as using <link href="http://www.client-canonical-site.com/" rel="canonical" /> in the <head> of the pages. Is that sufficient? Should I use rel=nofollow on all the links as well?

    Read the article

  • What should a page's minimum word count be in order to be effectively indexed?

    - by ZakGottlieb
    I'm seeding a new site with hundreds of (high quality) posts, but since I am paying per word written, I'm wondering if anybody in the community has any anecdotal evidence as to how many words of content there should now be for a page to be counted just the same as a 700 word+ post, for example? I know there are always examples of pages ranking well with, for instance, 50 words or less of content, but does anyone have any strong evidence on what the minimum count should be, or has anyone read anything very informative in regards to this issue? Thanks a lot in advance!

    Read the article

< Previous Page | 166 167 168 169 170 171 172 173 174 175 176 177  | Next Page >