Search Results

Search found 50980 results on 2040 pages for 'http compression'.

Page 432/2040 | < Previous Page | 428 429 430 431 432 433 434 435 436 437 438 439  | Next Page >

  • I can not download anything

    - by Jason Machen
    I am very new to ubuntu but decided to wipe my windows 7 and install it. I can not download anything from the software center. This is the error message I get. I can use the web in all other ways including this site. What can I do? Thanks, Jason W:Failed to fetch http://security.ubuntu.com/ubuntu/dists/raring-security/main/source/Sources 404 Not Found [IP: 91.189.91.13 80] W:Failed to fetch http://security.ubuntu.com/ubuntu/dists/raring-security/restricted Plus about 20 other lines.

    Read the article

  • How to boot into Live USB/CD with UEFI on latest Vaio S series?

    - by user76528
    I am unable to boot from USB/CD into a live Ubuntu session to proceed with install. Esc or F12 didn't work and I can't seem to be able to enter the bios settings. Sony Vaio S w/ Ivy Bridge (just one month old, Insyde H2O EFI?) I have read about: creating /EFI/boot on a fat16 stick http://ubuntuforums.org/showpost.php?p=12071096&postcount=2 installing rEFInd using the Alternate install CD (supposedly in order not to mess with W7 UEFI boot partition) It's pretty clear what I have to do afterwards http://askubuntu.com/a/157062 but I am at loss on how to simply boot into a Live Session then install. Thank you

    Read the article

  • What did I do wrong when installed this theme?

    - by Qmal
    I have a problem installing a theme for my Ubuntu, note that I am very new to it and probably messed something up. Still it seems to me that I did everything like it was stated in the INSTALL file. Theme: http://opendesktop.org/content/show.php/?content=140562 So I downloaded the *.zip package, unzipped it in home/.themes folder and changed Controls and Borders inside my theme preference. Still my result is pretty poor. You can look at the image, it looks nothing like what in the screenshot of the author. I also installed fonts running sudo apt-get install ttf-mscorefonts-installer and tried to install GTK2 engine as sudo apt-get install gtk2-engines-murrine gtk2-engines-pixbuf, but it showed that 0 files was modified since I already have it. Please tell me what did I did wrong so I can fix that :) (It seems that I can't upload images yet, no rep. I can give you link tho) http://postimage.org/image/1zhibxx5w/full/

    Read the article

  • wrong SEO result for staging website

    - by OCD
    Hi, We have a domain with URL, lets say: http://www.xxxxstaging.com/abc.php?id=10 which is having a pretty good ranking for some keywords in google. However, this is our staging website and due to some historical reason, it was exposed to public. Now we have a production domain (the one we want to be publicly available), say: http://www.xxxxproduction.com/abc.php?id=10 which have exactly the same html output as xxxxstaging.com our question is, for the exactly same keyword search, it never appear in any page of google. Is it possible for us to "switch" the ranking in google, either tell the robot that they are indeed the same website, or tell google to change it for us? Thank you.

    Read the article

  • O'Reilly Deal to Oct/14/2013 05:00 PT - 50% off E-Books on mastering CSS3

    - by TATWORTH
    Originally posted on: http://geekswithblogs.net/TATWORTH/archive/2013/10/11/oreilly-deal-to-oct142013-0500-pt---50-off-e-books.aspxAt  http://shop.oreilly.com/category/deals/css3.do?code=WKCSS&imm_mid=0b155e&cmp=em-prog-books-videos-lp-owo_css3_direct_wkcss, O'Reilly are offering 50% off a number of e-books on mastering CSS3 "CSS3—the technology behind most of the eye-catching visuals on the Web today—is loaded with capabilities that once would have required JavaScript or third-party plugins, such as animation, pseudo-classes, and media queries. Use CSS3 to transform markup into stunning, richly detailed web pages that look great in any browser. For one week only, SAVE 50% on CSS3 ebooks from shop.oreilly.com and take your sites from ordinary to incredible."

    Read the article

  • error: you need to load kernel first

    - by Angelos318
    I made a clean install on my Sony Vaio laptop, of Ubuntu 11.10 and when the installation was ready, it prompted to remove the usb I was installing the distro from, and press enter to reboot. After this reboot the first thing I got was the following error: error: couldn't read file error: you need to load the kernel first Press any key to continue.. After that it throws me back to the Grub select screen: Ubuntu, with linux 3.0.0-14-generic-pae recovery mode previous linux versions (none since I made a clean install) memory test If i choose the first option it shows only a black screen and never loads anything. If i reboot the same thing happens. Could I repair this using boot-repair? Is there any other way? Note: I know nothing about linux code so i am a total noob on this one Update: boot-repair did not help Grub.cfg here: http://pastebin.com/GKLuDuhM Boot Info Script: http://pastebin.com/indARkKJ

    Read the article

  • Problem connecting to Webmin

    - by railguage48
    I have installed Webmin with the view to try an understand what is all running. Yesterday I had it setup with login and password, but today when I try to get access to the server with: https://ubuntu:10000/ in order to login, I get unable to connect, the page does not load. Seems like there is no connection. I tried http://localhost:10000 and https://localhost:10000 They both returned the same unable to connect response. I am not sure of what it means to turn https on. ... the https and http is not showing. Am I going about this the wrong way?

    Read the article

  • Can't ping external websites

    - by Frantumn
    I can't ping google.com with my virtual ubuntu 12.04 server. I have set up a proxy URL in my /etc/apt/apt.conf file and it says Aquire::http::proxy http://urlname.com:9999; Now, I don't know a lot about how the proxy works, but I do know when we use it on windows VMachines it's a pac script that we place in internet explorer LAN settings and it automatically detects the script and gives internet access. I tried including the 9999/proxy.pac in the apt.conf URL and it didn't seem to work any better. Would ubuntu know how to handle a proxy.pac assuming it was created for windows? Should my URL include the .pac or just end after the port numbers? I've tried both without sucess, but I would like to know. A quick test to ping a fellow co-workers' PC was sucessful. So I can see network computers, but not google. or other internet sources.

    Read the article

  • Setting proxy from terminal

    - by baltusaj
    I have tried changing my proxy settings in a terminal as: export HTTP_PROXY=http://10.1.3.1:8080 and export http_proxy=http://10.1.3.1:8080 but when I try to install a new package or update apt-get, apt-get starts displaying messages from which it seems it is trying to connect to a previously set proxy: sudo apt-get update 0% [Connecting to 10.1.2.2 (10.1.2.2)] [Connecting to 10.1.2.2 (10.1.2.2) I have tried setting the proxy via bashrc file but that din work either. As far as I remember 10.1.2.2 was set using GNOME GUI but I don't have access to the GUI right now so I am trying to set it from terminal.

    Read the article

  • Add in the header of the license type is enough to say: "my code is licensed"? (Open-source)

    - by silverfox
    I do not know if this is the correct place to ask this stackexchange. Note: If a moderator can move to the correct place (if I am in the inappropriate site SE) I read on various sites about licenses. I did just put the license type in the header file (in my case the javascript file - open-source). /* * "codeName" "version" * http://officialsite.com/ * * Copyright 2012 "codeName" * Released under the "LICENSE NAME" license * http://officialsite.com/LICENSE NAME */ javascript code ... In the same folder I leave a copy of the license. The listing of the folder looks like this: * codeName.js * LICENSE In the file LICENSE would leave my code uses. What nobody says is if it is enough to say my code is licensed (the case of an open-source). Or is something more required? Sorry for the bad English. Thanks.

    Read the article

  • What percent of visitors should click on the next page before you enable prefetching?

    - by Kevin Burke
    Mozilla Firefox and Google Chrome support prefetching via an HTML tag: <!-- in chrome --> <link rel="prerender" href="http://example.org/index.html"> I suppose it is always worthwhile to include this tag if 100% of users on a page click on the "Next Page" button or similar, and never worthwhile to include it if only 2% or 3% of users visit the following page. At what percent of clicks should you turn on prefetching of the next page? 65%? Also, does the calculus change if the current page is HTTP and the next page is HTTPS?

    Read the article

  • Compiz window decorator randomly disappears

    - by adjfac
    I know people have posted this problem before such as http://askubuntu.com/users/authenticate/?s=9b515fe8-5da2-4bb9-ba01-dcea538b9b9c&dnoa.userSuppliedIdentifier=https%3A%2F%2Fwww.google.com%2Faccounts%2Fo8%2Fid&openid.mode=cancel&openid.ns=http%3A%2F%2Fspecs.openid.net%2Fauth%2F2.0 but i am looking for a more permanent solution right now. It happens to me at least once a day and when that happens, I usually go to compiz, uncheck and check window decorator and it will bring it back. But it's sort annoying and just wonder if there is a permanent fix out yet? I am using gnome (not unity), on a bran-new thinkpad t520.

    Read the article

  • client website compromised, found a strange .php file. any ideas?

    - by Kevin Strong
    I do support work for a web development company and I found a suspicious file today on the website of one of our clients called "hope.php" which contained several eval(gzuncompress(base64_decode('....'))) commands (which on a site like this, usually indicates that they've been hacked). Searching for the compromised site on google, we got a bunch of results which link to hope.php with various query strings that seem to generate different groups of seo terms like so: (the second result from the top is legitimate, all the rest are not) Here is the source of "hope.php": http://pastebin.com/7Ss4NjfA And here is the decoded version I got by replacing the eval()s with echo(): http://pastebin.com/m31Ys7q5 Any ideas where this came from or what it is doing? I've of course already removed the file from the server, but I've never seen code like this so I'm rather curious as to its origin. Where could I go to find more info about something like this?

    Read the article

  • Avoid penalties for duplicate (multilanguage) shared hosting

    - by Dave
    My concern is about SEO. Now let me explain the scenario. I am making a 3 languages website. The development is alright, but I was targeting local customers with one domain, and international (english version) with another. Eg: Local http://www.minhalojadesapatos.com.br (this is not the real website, just example!) Other http://www.myshoesstore.com.br Both domain point to exactly the same hosting and content, but when user comes through local domain, default language is set to portuguese, otherwise, default is english. Language handling on backend uses PHP Sessions and cookies, so with just a click users can change content language. How to avoid being SEO-penalised in this context? (yeah, I was hungry when focusing market for choosing two domains but the activity really needs that, it is a travel agency).

    Read the article

  • Apress Books - 3 - Pro ASP.NET 4 CMS (ISBN 987-1-4302-2712-0) - Final comments

    - by TATWORTH
    This book is more than just  a book about an ASP.NET CMS system -  it has much practical advice and examples for the Dot Net web developer. I liked the use of JQuery to detect that JavaScript was not enabled. One chapter was about MemCached - this one chapter could justify the price of the book if you run a server farm and need to improve performance. Some links to get you started are: Windows Memcache at http://code.jellycan.com/memcached/ Dot Net Access Library at http://sourceforge.net/projects/memcacheddotnet/ The chapters on scripting, performance analysis and search engne optimisation all provide excellent examples. This certainly is a book that should be part of every Dot Net Web Development team library. Congratulations to the author and to Apress for publishing this book!

    Read the article

  • RewriteRule working local but not on remote server

    - by m0tv
    I have a .htaccess file with one simple RewriteRule: RewriteEngine on RewriteRule ^([A-Za-z0-9-]+)$ ?site=$1 I want to have an url like http://www.example.com/imprint and forward it to http://www.example.com/?site=imprint I checked this rule with an RewriteRule tester which gave me the results I want to achieve. On my local development system it works well too. But on a remote server the URLs just give me a 404 error. Other more simple rewrite rules are working with no problems, so everything must be set up correctly (I think..). The problem is that I don't have access to any error logs or the server configs. So the only thing I can do is to guess... Can anyone tell me if theres something wrong with this rule? Or anything else I can do or test to solve this? Or has someone an idea what could be wrong on the server?

    Read the article

  • Redirect from the main page on multilanguage site

    - by Pluto
    I've decided to start developing multilingual web site. Now I'm thinking about the structure. The idea is to use a simple algorithm to find out what language is most interesting to the visitor when he makes his first visit to the root page, and then forward him to the corresponding address, for example, http://example.com/en/ for English-speaking audience and http://example.com/ru/ for Russians. The question is, how search engines react on such redirect? If this approach does not take away site's reputation, which redirects should I choose: 301 or 302? I would be very grateful for the reasoned response.

    Read the article

  • Writing a basic shader for large input files

    - by Zoltan Varadi
    I started writing a shader for my iOS app and instead of starting from scratch i used this tutorial here: http://www.raywenderlich.com/3664/opengl-es-2-0-for-iphone-tutorial I wrote an import function, first to import wavefront .obj models. My problem is that with I can't handle larger inputs (with a simple cube it was working). I realized that the indices array is an array of GLubyte values, which is unsigned char, so as a result i cant have more than 256 indexes. I modified it to GLuint, but then only get a blank screen. What else needs to me modified? p.s.: the source can be downloaded from here: http://d1xzuxjlafny7l.cloudfront.net/downloads/HelloOpenGL.zip

    Read the article

  • Using todolist (abstract spoon) on ubuntu 11.10?

    - by Tal Galili
    I wish to run todolist 6.3.8 (http://www.codeproject.com/KB/applications/todolist2.aspx) on ubuntu 11.10. I have installed the latest wine and winetricks. I have tried running: http://www.codeproject.com/KB/applications/todolist2.aspx winetricks vcrun2005 winetricks vcrun2008 winetricks vcrun6 Which installed all of the components. When I went to run todolist.exe, it started fine with the "first time wizard" (asking me where to save definitions and what file to open first), and then it stopped saying "the program todolist.exe has encountered a serious problem and needs to close. we are sorry for the inconvenience". What can I do to make this (great) software work on ubuntu? Thanks.

    Read the article

  • My talk at the Launch of the Vancouver Silverlight User Group

    Last week I did a presentation for the Launch of the new Vancouver Silverlight User Group (http://vanslug.net). I had a touch screen laptop so I started my presentation with the Multi Touch Silverlight Puzzle demo, wow effect guaranteed!   So 80 people were Rickrolled this night! You can download and try this demo from this link: http://www.silverlight.net/community/samples/silverlight-4/html-puzzle/   My presentation was mostly inspired by this awesome MIX10 session: Microsoft...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Figuring our complex REST queries for SharePoint

    - by Sahil Malik
    SharePoint 2010 Training: more information A little while ago, I showed the REST query for a relatively complex query. Some readers have emailed me about how to figure out further queries, and especially for complex insert/delete/update scenarios. Well it is quite easy to figure out almost any query for SharePoint REST API. Okay, this is not just about SharePoint – you can apply what you read here for any REST API interface supported by Microsoft, like WCF data services. But, sometimes when you have many columns, or complex update operations, or are working with weird providers, it is tough to figure out the specific HTTP request you need to craft, error free, using REST. Well fear not, there is hope. As an example, what I did is, I created a SharePoint site at http://sp2010.winsmarts.internal/sampledata with 3 lists in it - 1. Artists (with one Column, Title) Read full article ....

    Read the article

  • best/simplest way to inform search engine of sitemap location

    - by Don
    AFAIK, there are 2 ways to make search engines aware of a sitemap's location: Include an absolute link to it in robots.txt Submit it to them directly. The relevant URLs are: http://www.google.com/webmasters/tools/ping?sitemap=SITEMAP_URL http://www.bing.com/webmaster/ping.aspx?sitemap=SITEMAP_URL Where SITEMAP_URL is the absolute URL of the sitemap. Currently, I do both. Regarding (2), I have a job that runs automatically every day which submits the sitemap to Bing and Google. I don't think there's any reason to do (1) and (2), but I'm paranoid, so I do. I imagine you can avoid both (1) and (2) if you just make your sitemap accessible at a conventional URL (like robots.txt). What's the simplest and most reliable way to ensure that search engines can find your sitemap?

    Read the article

  • O'Reilly 50& off offer on CSS3 books to 05:00 PT on Oct/28

    - by TATWORTH
    Originally posted on: http://geekswithblogs.net/TATWORTH/archive/2013/10/21/oreilly-50-off-offer-on-css3-books-to-0500-pt.aspxAt  http://shop.oreilly.com/category/deals/css3.do?code=WKCSS&imm_mid=0b155e&cmp=em-prog-books-videos-lp-owo_css3_direct_wkcss, O'Reilly are offering 50% off a number of e-books on mastering CSS3 to 05:00 PT on Oct 28 "CSS3—the technology behind most of the eye-catching visuals on the Web today—is loaded with capabilities that once would have required JavaScript or third-party plugins, such as animation, pseudo-classes, and media queries. Use CSS3 to transform markup into stunning, richly detailed web pages that look great in any browser. For one week only, SAVE 50% on CSS3 ebooks from shop.oreilly.com and take your sites from ordinary to incredible."

    Read the article

  • Recommend an open source CMS for single page web site

    - by RedMan
    Hi I want to create a single page web site like http://kiskolabs.com/ or http://www.carat.se to display my portfolio. I want to add new products after launching the site without having to edit the entire site. I've looked at opencart (too much for single page site), Magneto (more for ecommerce), Wordpress (couldn't find open source / free templates which i can start from). Can you suggest a CMS which will support the creation of a single page site and allow insertion of new products without having to edit the entire page? I would prefer a CMS which also has open source / free templates which I can tweak for my use. I can do php and mysql, xml. If it is an easier option I can do PSD to site (but don't know much about this at all).

    Read the article

  • VirtualBox Port Forward

    - by john.graves(at)oracle.com
    A great new feature in VirtualBox 4.0 is the ability to use NAT networking and forward ports without needing to use ssh -L/-R tricks.  This is great for booting multiple VM domains simultaneously.  It is possible to have several instances which map back to the host machine and different ports on localhost:* automatically forward to the correct VM.  This avoids the hassle of setting up dns entries or static IP addresses.In this example, I'm mapping the host ports 3xxxx to the VM's well known server ports.Note: It is important to setup the Frontend HTTP host/port to avoid incorrect URL rewriting.You may also need to setup an http channel to deal with local traffic which uses the network address 10.0.2.15Happy VMing.

    Read the article

< Previous Page | 428 429 430 431 432 433 434 435 436 437 438 439  | Next Page >