Search Results

Search found 61530 results on 2462 pages for 'http header fields'.

Page 590/2462 | < Previous Page | 586 587 588 589 590 591 592 593 594 595 596 597  | Next Page >

  • Can't ping external websites

    - by Frantumn
    I can't ping google.com with my virtual ubuntu 12.04 server. I have set up a proxy URL in my /etc/apt/apt.conf file and it says Aquire::http::proxy http://urlname.com:9999; Now, I don't know a lot about how the proxy works, but I do know when we use it on windows VMachines it's a pac script that we place in internet explorer LAN settings and it automatically detects the script and gives internet access. I tried including the 9999/proxy.pac in the apt.conf URL and it didn't seem to work any better. Would ubuntu know how to handle a proxy.pac assuming it was created for windows? Should my URL include the .pac or just end after the port numbers? I've tried both without sucess, but I would like to know. A quick test to ping a fellow co-workers' PC was sucessful. So I can see network computers, but not google. or other internet sources.

    Read the article

  • Cloudcel: Excel Meets the Cloud

    - by kaleidoscope
    Cloudscale  is launching Cloudcel Cloudcel is the first product that demonstrates the full power of integrated "Client-plus-Cloud" computing. You use desktop Excel in the normal way, but can also now seamlessly tap into the scalability and massive parallelism of the cloud, entirely from within Excel, to handle your Big Data. Building an app in Cloudcel is really easy – no databases, no programming. Simply drop building blocks onto the spreadsheet (in any order, in any location) and launch the app to the cloud with a single click. Parallelism, scalability and fault tolerance are automatic. With Cloudcel, you can process realtime data streams continuously, and get alerts pushed to you as soon as important events or patterns are detected ("Set it and forget it"). Cloudcel is offered as a pay-per-use cloud service – so no hardware, no software licenses, and no IT department required to set it up. Private cloud deployments are also available. Please find below link for more detail : http://billmccoll.sys-con.com/node/1326645 http://cloudcel.com/ Technorati Tags: Tanu

    Read the article

  • {12.04} cannot open software center or use apt-get install

    - by d6braun
    i have been trying to get it working for over an hour and a half now, the code below is what i keep getting. the example is from me trying to install synaptic, but i get the errors on most apt-get commands and update manager. sudo apt-get install synaptic Reading package lists... Error! E: Encountered a section with no Package: header E: Problem with MergeList /var/lib/apt/lists/wine.budgetdedicated.com_apt_dists_edgy_main_binary-i386_Packages E: The package lists or status file could not be parsed or opened.

    Read the article

  • error: you need to load kernel first

    - by Angelos318
    I made a clean install on my Sony Vaio laptop, of Ubuntu 11.10 and when the installation was ready, it prompted to remove the usb I was installing the distro from, and press enter to reboot. After this reboot the first thing I got was the following error: error: couldn't read file error: you need to load the kernel first Press any key to continue.. After that it throws me back to the Grub select screen: Ubuntu, with linux 3.0.0-14-generic-pae recovery mode previous linux versions (none since I made a clean install) memory test If i choose the first option it shows only a black screen and never loads anything. If i reboot the same thing happens. Could I repair this using boot-repair? Is there any other way? Note: I know nothing about linux code so i am a total noob on this one Update: boot-repair did not help Grub.cfg here: http://pastebin.com/GKLuDuhM Boot Info Script: http://pastebin.com/indARkKJ

    Read the article

  • O'Reilly Deal to Oct/14/2013 05:00 PT - 50% off E-Books on mastering CSS3

    - by TATWORTH
    Originally posted on: http://geekswithblogs.net/TATWORTH/archive/2013/10/11/oreilly-deal-to-oct142013-0500-pt---50-off-e-books.aspxAt  http://shop.oreilly.com/category/deals/css3.do?code=WKCSS&imm_mid=0b155e&cmp=em-prog-books-videos-lp-owo_css3_direct_wkcss, O'Reilly are offering 50% off a number of e-books on mastering CSS3 "CSS3—the technology behind most of the eye-catching visuals on the Web today—is loaded with capabilities that once would have required JavaScript or third-party plugins, such as animation, pseudo-classes, and media queries. Use CSS3 to transform markup into stunning, richly detailed web pages that look great in any browser. For one week only, SAVE 50% on CSS3 ebooks from shop.oreilly.com and take your sites from ordinary to incredible."

    Read the article

  • Avoid penalties for duplicate (multilanguage) shared hosting

    - by Dave
    My concern is about SEO. Now let me explain the scenario. I am making a 3 languages website. The development is alright, but I was targeting local customers with one domain, and international (english version) with another. Eg: Local http://www.minhalojadesapatos.com.br (this is not the real website, just example!) Other http://www.myshoesstore.com.br Both domain point to exactly the same hosting and content, but when user comes through local domain, default language is set to portuguese, otherwise, default is english. Language handling on backend uses PHP Sessions and cookies, so with just a click users can change content language. How to avoid being SEO-penalised in this context? (yeah, I was hungry when focusing market for choosing two domains but the activity really needs that, it is a travel agency).

    Read the article

  • What percent of visitors should click on the next page before you enable prefetching?

    - by Kevin Burke
    Mozilla Firefox and Google Chrome support prefetching via an HTML tag: <!-- in chrome --> <link rel="prerender" href="http://example.org/index.html"> I suppose it is always worthwhile to include this tag if 100% of users on a page click on the "Next Page" button or similar, and never worthwhile to include it if only 2% or 3% of users visit the following page. At what percent of clicks should you turn on prefetching of the next page? 65%? Also, does the calculus change if the current page is HTTP and the next page is HTTPS?

    Read the article

  • wrong SEO result for staging website

    - by OCD
    Hi, We have a domain with URL, lets say: http://www.xxxxstaging.com/abc.php?id=10 which is having a pretty good ranking for some keywords in google. However, this is our staging website and due to some historical reason, it was exposed to public. Now we have a production domain (the one we want to be publicly available), say: http://www.xxxxproduction.com/abc.php?id=10 which have exactly the same html output as xxxxstaging.com our question is, for the exactly same keyword search, it never appear in any page of google. Is it possible for us to "switch" the ranking in google, either tell the robot that they are indeed the same website, or tell google to change it for us? Thank you.

    Read the article

  • What did I do wrong when installed this theme?

    - by Qmal
    I have a problem installing a theme for my Ubuntu, note that I am very new to it and probably messed something up. Still it seems to me that I did everything like it was stated in the INSTALL file. Theme: http://opendesktop.org/content/show.php/?content=140562 So I downloaded the *.zip package, unzipped it in home/.themes folder and changed Controls and Borders inside my theme preference. Still my result is pretty poor. You can look at the image, it looks nothing like what in the screenshot of the author. I also installed fonts running sudo apt-get install ttf-mscorefonts-installer and tried to install GTK2 engine as sudo apt-get install gtk2-engines-murrine gtk2-engines-pixbuf, but it showed that 0 files was modified since I already have it. Please tell me what did I did wrong so I can fix that :) (It seems that I can't upload images yet, no rep. I can give you link tho) http://postimage.org/image/1zhibxx5w/full/

    Read the article

  • RewriteRule working local but not on remote server

    - by m0tv
    I have a .htaccess file with one simple RewriteRule: RewriteEngine on RewriteRule ^([A-Za-z0-9-]+)$ ?site=$1 I want to have an url like http://www.example.com/imprint and forward it to http://www.example.com/?site=imprint I checked this rule with an RewriteRule tester which gave me the results I want to achieve. On my local development system it works well too. But on a remote server the URLs just give me a 404 error. Other more simple rewrite rules are working with no problems, so everything must be set up correctly (I think..). The problem is that I don't have access to any error logs or the server configs. So the only thing I can do is to guess... Can anyone tell me if theres something wrong with this rule? Or anything else I can do or test to solve this? Or has someone an idea what could be wrong on the server?

    Read the article

  • mod_rewrite for clean URL doesn't work

    - by deathlock
    Basically what I want to do is to convert this: http://localhost/jariungu/user_caleg.php?idCaleg2014=3 into this: http://localhost/jariungu/caleg/3 I have managed to make /jariungu/caleg/3 to direct to the original URL (as in, if I open that URL, it directs me to the appropriate page). The problem is, once opened, the URL returns to the original, ugly one in the address bar. This is what I tried. Could someone provide a help? <IfModule mod_rewrite.c> Options +FollowSymlinks RewriteEngine On RewriteBase /jariungu/ RewriteRule ^caleg\/([0-9]+)\/([a-zA-Z]+\s*[0-9]*)/?$ caleg.php?idCaleg2014=$1&namaCaleg=$2 [NC,L] RewriteRule ^caleg\/([0-9]+)/?$ caleg.php?idCaleg2014=$1 [NC,L] </IfModule>

    Read the article

  • Stop Google Analytics from appending hostname?

    - by Nick Q.
    I've come across an Analytics profile that is appending the rest of a URL to the end of a page's path. For example when looking at the page that exists at http://example.com/page I would expect to see /page but instead it shows me /page/http://example.com/. The profile has no filters applied to it, and until July was reporting as expected (/page), in July the site in question switched hosts (and absolutely nothing else, so I'm not sure that's the problem). The analytics code on the site is the standard Google Async code with a domain set. All other profiles for the site show /page as expected. Any ideas as to how I can get the profile to function as expected?

    Read the article

  • Compiz window decorator randomly disappears

    - by adjfac
    I know people have posted this problem before such as http://askubuntu.com/users/authenticate/?s=9b515fe8-5da2-4bb9-ba01-dcea538b9b9c&dnoa.userSuppliedIdentifier=https%3A%2F%2Fwww.google.com%2Faccounts%2Fo8%2Fid&openid.mode=cancel&openid.ns=http%3A%2F%2Fspecs.openid.net%2Fauth%2F2.0 but i am looking for a more permanent solution right now. It happens to me at least once a day and when that happens, I usually go to compiz, uncheck and check window decorator and it will bring it back. But it's sort annoying and just wonder if there is a permanent fix out yet? I am using gnome (not unity), on a bran-new thinkpad t520.

    Read the article

  • VirtualBox Port Forward

    - by john.graves(at)oracle.com
    A great new feature in VirtualBox 4.0 is the ability to use NAT networking and forward ports without needing to use ssh -L/-R tricks.  This is great for booting multiple VM domains simultaneously.  It is possible to have several instances which map back to the host machine and different ports on localhost:* automatically forward to the correct VM.  This avoids the hassle of setting up dns entries or static IP addresses.In this example, I'm mapping the host ports 3xxxx to the VM's well known server ports.Note: It is important to setup the Frontend HTTP host/port to avoid incorrect URL rewriting.You may also need to setup an http channel to deal with local traffic which uses the network address 10.0.2.15Happy VMing.

    Read the article

  • client website compromised, found a strange .php file. any ideas?

    - by Kevin Strong
    I do support work for a web development company and I found a suspicious file today on the website of one of our clients called "hope.php" which contained several eval(gzuncompress(base64_decode('....'))) commands (which on a site like this, usually indicates that they've been hacked). Searching for the compromised site on google, we got a bunch of results which link to hope.php with various query strings that seem to generate different groups of seo terms like so: (the second result from the top is legitimate, all the rest are not) Here is the source of "hope.php": http://pastebin.com/7Ss4NjfA And here is the decoded version I got by replacing the eval()s with echo(): http://pastebin.com/m31Ys7q5 Any ideas where this came from or what it is doing? I've of course already removed the file from the server, but I've never seen code like this so I'm rather curious as to its origin. Where could I go to find more info about something like this?

    Read the article

  • Redirect from the main page on multilanguage site

    - by Pluto
    I've decided to start developing multilingual web site. Now I'm thinking about the structure. The idea is to use a simple algorithm to find out what language is most interesting to the visitor when he makes his first visit to the root page, and then forward him to the corresponding address, for example, http://example.com/en/ for English-speaking audience and http://example.com/ru/ for Russians. The question is, how search engines react on such redirect? If this approach does not take away site's reputation, which redirects should I choose: 301 or 302? I would be very grateful for the reasoned response.

    Read the article

  • best/simplest way to inform search engine of sitemap location

    - by Don
    AFAIK, there are 2 ways to make search engines aware of a sitemap's location: Include an absolute link to it in robots.txt Submit it to them directly. The relevant URLs are: http://www.google.com/webmasters/tools/ping?sitemap=SITEMAP_URL http://www.bing.com/webmaster/ping.aspx?sitemap=SITEMAP_URL Where SITEMAP_URL is the absolute URL of the sitemap. Currently, I do both. Regarding (2), I have a job that runs automatically every day which submits the sitemap to Bing and Google. I don't think there's any reason to do (1) and (2), but I'm paranoid, so I do. I imagine you can avoid both (1) and (2) if you just make your sitemap accessible at a conventional URL (like robots.txt). What's the simplest and most reliable way to ensure that search engines can find your sitemap?

    Read the article

  • Apress Books - 3 - Pro ASP.NET 4 CMS (ISBN 987-1-4302-2712-0) - Final comments

    - by TATWORTH
    This book is more than just  a book about an ASP.NET CMS system -  it has much practical advice and examples for the Dot Net web developer. I liked the use of JQuery to detect that JavaScript was not enabled. One chapter was about MemCached - this one chapter could justify the price of the book if you run a server farm and need to improve performance. Some links to get you started are: Windows Memcache at http://code.jellycan.com/memcached/ Dot Net Access Library at http://sourceforge.net/projects/memcacheddotnet/ The chapters on scripting, performance analysis and search engne optimisation all provide excellent examples. This certainly is a book that should be part of every Dot Net Web Development team library. Congratulations to the author and to Apress for publishing this book!

    Read the article

  • Recommend an open source CMS for single page web site

    - by RedMan
    Hi I want to create a single page web site like http://kiskolabs.com/ or http://www.carat.se to display my portfolio. I want to add new products after launching the site without having to edit the entire site. I've looked at opencart (too much for single page site), Magneto (more for ecommerce), Wordpress (couldn't find open source / free templates which i can start from). Can you suggest a CMS which will support the creation of a single page site and allow insertion of new products without having to edit the entire page? I would prefer a CMS which also has open source / free templates which I can tweak for my use. I can do php and mysql, xml. If it is an easier option I can do PSD to site (but don't know much about this at all).

    Read the article

  • Figuring our complex REST queries for SharePoint

    - by Sahil Malik
    SharePoint 2010 Training: more information A little while ago, I showed the REST query for a relatively complex query. Some readers have emailed me about how to figure out further queries, and especially for complex insert/delete/update scenarios. Well it is quite easy to figure out almost any query for SharePoint REST API. Okay, this is not just about SharePoint – you can apply what you read here for any REST API interface supported by Microsoft, like WCF data services. But, sometimes when you have many columns, or complex update operations, or are working with weird providers, it is tough to figure out the specific HTTP request you need to craft, error free, using REST. Well fear not, there is hope. As an example, what I did is, I created a SharePoint site at http://sp2010.winsmarts.internal/sampledata with 3 lists in it - 1. Artists (with one Column, Title) Read full article ....

    Read the article

  • Removing surrounding noises from voice recording

    - by Peak Reconstruction Wavelength
    I have a wave file whose frequency spectrum looks like this. http://i.stack.imgur.com/2rRaS.png It contains audio, which I want to keep while removing the rest. The problem is that the surround noise changes, just those distinct voice patterns remain. I marked the voice patterns for clarity: http://i.stack.imgur.com/eLkBl.png What could an algorithm look like / a workflow in adobe audition look like that removes everything but the voice patterns? I think that the main characteristic is the line-shaped form over time. Loudness alone is not enough as the noise is loud aswell.

    Read the article

  • My talk at the Launch of the Vancouver Silverlight User Group

    Last week I did a presentation for the Launch of the new Vancouver Silverlight User Group (http://vanslug.net). I had a touch screen laptop so I started my presentation with the Multi Touch Silverlight Puzzle demo, wow effect guaranteed!   So 80 people were Rickrolled this night! You can download and try this demo from this link: http://www.silverlight.net/community/samples/silverlight-4/html-puzzle/   My presentation was mostly inspired by this awesome MIX10 session: Microsoft...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Using todolist (abstract spoon) on ubuntu 11.10?

    - by Tal Galili
    I wish to run todolist 6.3.8 (http://www.codeproject.com/KB/applications/todolist2.aspx) on ubuntu 11.10. I have installed the latest wine and winetricks. I have tried running: http://www.codeproject.com/KB/applications/todolist2.aspx winetricks vcrun2005 winetricks vcrun2008 winetricks vcrun6 Which installed all of the components. When I went to run todolist.exe, it started fine with the "first time wizard" (asking me where to save definitions and what file to open first), and then it stopped saying "the program todolist.exe has encountered a serious problem and needs to close. we are sorry for the inconvenience". What can I do to make this (great) software work on ubuntu? Thanks.

    Read the article

  • Newly compiled (patched with -ck) 3.3 kernel won't boot up

    - by Shiki
    Screen is garbled. Reaaaally garbled. Like some sort of scrambling image from a movie. I can upload some pictures if needed. To be short with the description. Here is my 3.3 + ck .config file: http://pastebin.com/aP2F9RcH The dmesg log of the failed kernel: http://pastebin.com/esc2Rtka The computer is a ThinkPad T500 laptop. I have selected everything needed for the computer (as far as I can tell, but check the .config). The normal generic Ubuntu kernel boots and works perfect!

    Read the article

  • using curl command to download file in parts from different interfaces and to run the commands simultaneously from a script

    - by jsjain007
    i wanted to download a file using curl command simultaneously in different parts using ip aliasing(virtual Ethernet ports) so what i did was pasted the commands in a text file and run but the problem as obvious since it is in a file the commands will be executed one by one so is there a way to run all those commands simultaneously. here is the command curl --interface eth0:0 --range 0,38010880 http://wdl.cache.ijinshan.com/wps/download/Linux/unstable/kingsoft-office_9.1.0.4244~a12p3_i386.deb -o kinsoft-office.part1 curl --interface eth0:1 --range 38010880 ,- http://wdl.cache.ijinshan.com/wps/download/Linux/unstable/kingsoft-office_9.1.0.4244~a12p3_i386.deb -o kinsoft-office.part2 cat kinsoft-office.part*>kinsoft-office can anyone help me to run these above 2 commands simultaneously from the script so as to increase download speed

    Read the article

< Previous Page | 586 587 588 589 590 591 592 593 594 595 596 597  | Next Page >