Search Results

Search found 5380 results on 216 pages for 'webmasters'.

Page 200/216 | < Previous Page | 196 197 198 199 200 201 202 203 204 205 206 207  | Next Page >

  • XCart Skip Checkout Steps

    - by user900553
    I am using xcart. In the checkout process, skip the step that appears after a customer clicks the add to cart button on the product page. It takes them to a page that shows the product and the text Added to your shopping cart. I want to take them directly to the view cart page, right before the checkout page. I also tried: Enable Redirect customer to cart after adding a product option in X-Cart admin back-end -Settings - General settings. Still not working. Anybody have Idea about this?

    Read the article

  • A drop in SERP after following webmaster guidelines [on hold]

    - by digiwig
    So here's a puzzle for all you SEO gurus out there. I recently launched my own site. I had target keywords which were ranking very well for about 1 month, within the top five and even appearing in first place. In an attempt to maintain good positioning, I followed guidelines by adding robots.txt, an xml sitemap redirecting non-www to www redirecting index.php to root domain adding htaccess 301 redirect for old pages I added rich snippets created a google+ account, verified my picture to appear, I went through each of the webmaster issues with duplicate titles and meta descriptions and improved header tag document outlines i even created a few more blog posts to keep the content freshing and moving. So now my website appears on page 2 with my target keywords - and all because I followed the guidelines. What is happening? I see competitors with stagnant content superglued to position 1.

    Read the article

  • Remove home page from Google cache

    - by Steve
    The Google Webmaster Tools Remove URL section allows you to specify a page URL to be removed from the index, or cache, or both. However, I want to remove just the home page, which is / I want to remove it from the cache because it is indexed when the "under construction" page was up. This URL is not recognised by the Remove URL section as an individual page. Instead Google assumes you want to remove the entire website from the index. I've specified /index.php and /index.html to be removed from the cache, but this is not the URL listed in the search results for the home page I want removed from the cache.

    Read the article

  • Redirecting to a default (or last visited) subdirectory. Does Google like this?

    - by andufo
    i have a site that has 3 web applications, lets use this example: example.com/nicy example.com/mash example.com/zoken The main application is nicy, so if the user comes for the first time (or if Google indexes my site) that will be the default choice. This is the code placed inside example.com/index.php <? header('HTTP/1.1 301 Moved Permanently'); header('Location: http://example.com/nicy'); die(); ?> Is this solution SEO friendly for Google to index the nicy subdirectory as the main entrance page for the domain? (because of the 301 redirect). Thanks!

    Read the article

  • XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token)

    - by Steve
    I have installed Joomla 1.5.22 on a new hosting account, which doesn't have a domain yet, so it's public URL is http://cp-013.micron21.com/~annimac/ A message saying: XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token). The source code for this message is: <dl id="system-message"> <dt class="error">Error</dt> <dd class="error message fade"> <ul> <li>XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token)</li> </ul> </dd> </dl> There is nothing in /logs to indicate what the problem is. I have uploaded the following folders from a freshly unzipped copy of Joomla 1.5.22: administrator components includes language\en-GB libraries modules plugins templates\ja_purity xmlrpc and the issue remains. I have no custom or additional plugins, modules, or components installed. If I change templates, the problem remains. What is the problem?

    Read the article

  • How to specify ingredients in Schema.org (other than recipes or drugs)?

    - by guillem
    I'd like to use Schema.org for a series of product specs pages. The products are cosmetics (shampoos, creams, etc.) and ideally an INCI declaration will be included for every product. I'd like to use Schema.org in the declaration, but I cannot find a suitable category in Schema.org. Ingredients is meant for cooking recipes, and ActiveIngredient for (medical) drugs and dietary supplements. Did I overlook some part of the tree? Any ideas for a workaround? Or am I trying to do something out of scope?

    Read the article

  • Google Analitycs says source is Search Engine (100%)

    - by Angel Estrada
    I have a problem with my google analytics code, 'cause it says since 2 months that all traffic of my websites is directly from search engine, i'm really sure that must be a mistake 'cause i have made some email marketing and landing page campaing that could help to create referal traffic, maybe somethink goes wrong with the new version of the analytics or i forgot some settings. please any help is aceptable

    Read the article

  • External link tracking when opening the link in a new window in Google Analytics?

    - by evanmcd
    OK, so this seems like a really simply problem, but I have yet to find a solution that accomplishes the following: Opens the link in a new window Tracks the event in GA (obviously) Doesn't trigger pop-up blockers (uses target="_blank" instead of window.open) Most of the code I've seen, including Google's, doesn't take into account the case of opening in a new window - they just window location.href. Even GAAddons (http://gaaddons.com/), which charges for commercial use, doesn't seem to work for me. Perhaps, I'm missing something simple - I'd be relieved if so and would thank profusely whoever points it out to me! If no one is able to provide an example, I'll post some of the test cases I've created to illustrate the problem. Thanks.

    Read the article

  • .htaccess two different rules but only one per time

    - by dragon112
    I'm rather new to the whole .htaccess thing and I'm using the following right now to use 'pretty url's': <IfModule mod_rewrite.c> RewriteEngine On RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_FILENAME} !-f RewriteRule ^(.*)$ index.php?path=$1 [NS,L] </IfModule> Now i found my website a bit slow and decided to start gzipping my CSS files thru a php script I found somewhere on the web (the website). For this to work I need to rewrite the url to open the correct php file. That would look something like this: RewriteRule ^(.*).css$ /csszip.php?file=$1.css [L] But I only want the first to happen when the second doesn't and vice versa. In other words i'd like something like this: <IfModule mod_rewrite.c> RewriteEngine On if request doesn't contain .css do RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_FILENAME} !-f RewriteRule ^(.*)$ index.php?path=$1 [NS,L] else do RewriteRule ^(.*).css$ /csszip.php?file=$1.css [L] </IfModule> Can anyone help me with the proper code or a place where i can find a way to use some kind of conditional statement in htaccess files? Thanks in Advance!:)

    Read the article

  • Getting expired domain name - most effective route?

    - by kcdwayne
    There is a domain name I have been wanting for years that was used as a parked page (read: cybersquatted) that has entered into WHOIS's redemptionPeriod stage. The domain has been expired for 61 days now - after contacting the Registar, they informed me that it would stay in redemption for 75 days. After which, it would either be sold to resellers, or sent back to the public registry. (I have since sent a followup message requesting the reseller they use.) My question is: What is the best way to proceed? I know there is at least one competitor that would love to have this name, but I'm unsure if they even know it's expiring. I did not tell the Registar the domain in question, as they seem geared towards cybersquatting, and I do not trust them. Domain Front Running sucks. Should I use a backorder service? Should I just take my chances and try to grab it after75 days? I checked an auctionhouse by manually browsing their expired domains - it wasn't there.

    Read the article

  • Is hidden content (display: none;) -indexed- by search engines? [closed]

    - by user568458
    Possible Duplicate: How bad is it to use display: none in CSS? We've established on this site before (in this question) that, since there are so many legitimate uses for hiding content with display: none; when creating interactive features, that sites aren't automatically penalised for content that is hidden this way (so long as it doesn't look algorithmically spammy). Google's Webmaster guidelines also make clear that a good practice when using content that is initially legitimately hidden for interactivity purposes is to also include the same content in a <noscript> tag, and Google recommend that if you design and code for users including users with screen readers or javascript disabled, then 9 times out of 10 good relevant search rankings will follow (though their specific advice seems more written for cases where javascript writes new content to the page). JavaScript: Place the same content from the JavaScript in a tag. If you use this method, ensure the contents are exactly the same as what’s contained in the JavaScript, and that this content is shown to visitors who do not have JavaScript enabled in their browser. So, best practice seems pretty clear. What I can't find out is, however, the simple factual matter of whether hidden content is indexed by search engines (but with potential penalties if it looks 'spammy'), or, whether it is ignored, or, whether it is indexed but with a lower weighting (like <noscript> content is, apparently). (for bonus points it would be great to know if this varies or is consistent between display: none;, visibility: hidden;, etc, but that isn't crucial). This is different to the other questions on display:none; and SEO - those are about good and bad practice and the answers are discussions of good and bad practice, I'm interested simply in the factual 'Yes or no' question of whether search engines index, or ignore, content that is in display: none; - something those other questions' answers aren't totally clear on. One other question has an answer, "Yes", supported by a link to an article that doesn't really clear things up: it establishes that search engines can spot that text is hidden, it discusses (again) whether hidden text causes sites to be marked as spam, and ultimately concludes that in mid 2011, Google's policy on hidden text was evolving, and that they hadn't at that time started automatically penalising display:none; or marking it as spam. It's clear that display: none; isn't always spam and isn't always treated as spam (many Google sites use it...): but this doesn't clear up how, or if, it is indexed. What I will do will be to follow the guidelines and make sure that all the content that is initially hidden which regular users can explore using javascript-driven interactivity is also structured in way that noscript/screenreader users can use. So I'm not interested in best practice, opinions etc because best practice seems to be really clear: accessibility best practices boosts SEO. But I'd like to know what exactly will happen: whether any display: none; content I have alongside <noscript> or otherwise accessibility-optimised content will be be ignored, or indexed again, or picked up to compare against the <noscript> content but not indexed... etc.

    Read the article

  • SEO consideration for duplicate sites

    - by Malk
    I am building a brochure-ware website for a company that sells products all across the world. They need the site to ask the user what region they are in before using the site; there are 5 regions. This is because there are different products offered to different regions and each region may or may not want to customize their own content. However, at launch and likely forever, most of the pages will be the exact same minus what is listed in the footer and in the product selection menu. My question is how should I structure the sitemap for this site for best SEO? Should I be concerned with duplicate content penalties and/or cannibalizing the site's presence on the SERP? Some considerations: The client wants to be able to print links directly to regional specific content bypassing any prompt for the user to select a region (to ensure they land on the target page). The client cannot have a 'default' region so the user must have a region specified "Clean" urls are important, but there is wiggle room The client does not want each region to have its own domain There will be a link on the page to allow users to specify a different region The client is not concerned with localization ...at this time Some products are available in multiple regions A quick list of options I am considering: www.site.com/region/page region.site.com/page www.site.com/page?region (no cookie, pages require the parameter. If visited without; the user must select a region) www.site.com/page (using cookie and a splash screen if needed; could pass parameter in to set the region for direct linking) Thanks in advance for your advice.

    Read the article

  • Links don't work on new server

    - by Beko
    I've recently moved from server with php4 to php5. Now I cant open content in my page. Links look like this "page.com/news-it-is-rainy-day-1.html". But when I try to open link I just get the archive (news.html). I also use smarty template engine. My rewrite rule in .htacces looks like this: Options +FollowSymlinks RewriteEngine on RewriteRule ^news-(.*)\.html$ news.php?do=read&id=$1 RewriteRule ^news.html$ news.php RewriteRule ^index.html$ index.php AddType application/x-httpd-php .xml

    Read the article

  • Thumbnail preview rotator on mouse-over - what is it called?

    - by Gerben
    I've been searching for hours now but can't get any tutorial on this. I have these thumbnails on my homepage which are the first images of their corresponding posts. What I want is that when I mouse-over a post on the homepage... the corresponding thumbnail should also show/rotate the other images of that particular post... bit like a sneak-peak image rotator... Does anyone know where I can get a tutorial regarding this? How's this called? It seems like I'm searching for the wrong keywords on google as I can't find anything.

    Read the article

  • What is the best and cheapest CMS for an E-Commerce site?

    - by kf0l
    My client needs a site customers can use to pay the membership fees and join. It should also track when old members need to pay again. There are some members that will not want to pay online and they want to be able to use the site to manage all payments. The CMS must be cheap to host online, potentially with something like GitHub/Heroku. I am experienced with PHP and RoR. I am using a CMS so once I am done with this project it is easy for them to update. After reading through different posts I think WordPress with an e-commerce Plugin may be the right solution. Spree and WebGUI also looks promising. Thanks for the help.

    Read the article

  • Reliable method for google analytics tracking for print advertising campaign?

    - by chrisjlee
    A client is looking to track advertising clicks through a newspaper ad to measure success. They have rigid business requirements that it will be a unique domain... e.g. foowidgetsnews.net instead of foodwidgets.com/contact-form-page.php What is the most reliable method of building redirected url to a landing page so it will be tracked in google analytics as a direct hit from the newspaper? Finally, we would like to track the foowidgetsnews.net as the main url in google analytics because 301 redirect isn't tracked in google analytics like the way we would like it to.

    Read the article

  • How would I migrate a front page form based site to an IIS7 server? [on hold]

    - by Nick Gilpin
    I have an large html website on a very old server and I'm attempting to move it to a brand new server running IIS7. I've tried moving the entire site to the WWW directory, but then all my forms return "The HTTP verb POST used to access path is not allowed". Additionally, all the solutions I've found online only work on a page by page basis. Is there a simpler/faster way to migrate a large website containing many forms to IIS7? Here is the specific error message: The HTTP verb POST used to access path '/FormServer/Mig/_vti_bin/shtml.dll/admissions/askseaaggie.htm' is not allowed.

    Read the article

  • Google analytics shows wrong number of page views, asp.net website

    - by f_karlsson
    Sometimes it can be for example 4500 requests, after a few hours it shows a few thousand less. What is wrong? It looks like analytics corrects itself. I changed from classic to Universal a few months ago, do not know if it has anything to do with this. In masterpage: <script> (function (i, s, o, g, r, a, m) { i['GoogleAnalyticsObject'] = r; i[r] = i[r] || function () { (i[r].q = i[r].q || []).push(arguments) }, i[r].l = 1 * new Date(); a = s.createElement(o), m = s.getElementsByTagName(o)[0]; a.async = 1; a.src = g; m.parentNode.insertBefore(a, m) })(window, document, 'script', '//www.google-analytics.com/analytics.js', 'ga'); ga('create', 'UA-xxxxxxxx-1', 'xxxxx.se'); ga('send', 'pageview'); </script>

    Read the article

  • phpBB - Reducing Spam

    - by user44175
    I've installed phpBB Forums last week and the past 2 days I've been getting users sign up and posting spam chinese emails on each topic. I have:- Added captcha on registration Made sure users have to verify subscription by email before allowing to post What else can I do to stop this from happening? I've banned their IP addresses but this doesn't stop them from using a proxy to keep spamming the forums. I've read I can block all chinese IP addresses through ACP but is this the best step to block all this? Seems to be all chinese spam at the minute, any help would be much appreciated.

    Read the article

  • jQuery Mobile Frame Forwarding [on hold]

    - by Nizam
    I have a site that does a standard forward to another site [301 Redirect]. In the redirected site, I detect if the device is a mobile using the following code: if (/Android|webOS|iPhone|iPad|iPod|BlackBerry/i.test(navigator.userAgent)) { window.location.replace("Mobile/Login/Login.aspx") } else { window.location.replace("Apps/Login/Login.aspx") } It works and jQuery mobile makes the site fits device very well. To do so, I use the following code in ASPX page: <meta name="viewport" content="width=device-width, initial-scale=1, maximum-scale=1" /> The problem is that I am wanting to frame forward - instead of standard forward - (there are a lot of advantages), but the site is not fitted for the device anymore, and even the icon I have chosen for my page is not well defined anymore. The code I use to set the icon of page is: <link rel="apple-touch-icon" href="../../Apps/Imagens/Icone.png" /> My site is hosted by Mochahost My question is: Is there anything I could do to make it works?

    Read the article

  • What to use for an event listing site? [on hold]

    - by Vykintas
    I have a site which lets users buy & sell tickets, but it's built on wordpress & buddypress. So it's very heavy and messy. I would like to re-do the whole site on something lighter, cleaner and solid. The main functionality for user would have to be as follows: Register or login via Facebook. Create events and sell tickets to them. See ticket sales statistics Upload photos and associate those with events. Buy event tickets, print pdf ticket. Comment, favourite and like events. What would be your suggestions? PHP framework? CMS? CMF? I must say that I'm a front-end dev so building a system from scratch on my own would take a while. I'd be interested more in a "skeleton" app solution or something similar.

    Read the article

  • Free web mangement control panel

    - by Thorn007
    Hey guys I need some help. I will also apologize for not being able to be more specific. I'm looking for a specific web admin panel. that uses a login page via port 2222 or 4444. This is not vanilla forum or any forum. So the only way I know how to make this a legit question is to ask what "free" control panels do you use to manage your web sites. This means files and domains. Why do you use it? Where is it located?

    Read the article

  • Why am I getting this message "Some important page is blocked by robots.txt"?

    - by Rounak
    My site's URL is: www.hackinguniverse.org From some day in Google Webmaster tool a message is showing that says "Some important page is blocked by robots.txt". My robots.txt is: User-agent: Mediapartners-Google Disallow: User-agent: * Disallow: /search Allow: / Sitemap: http://www.hackinguniverse.org/feeds/posts/default?orderby=updated Now for other information - I hosted this website on blogger. I have some other sitemaps too where I had only included "/Search" in Disallow like in this robots.txt file. But those sites are ok. I mean no message is showing on those. So why am I getting that message telling that I have blocked some important page via robot.txt?

    Read the article

  • Blank lines between sourcecode [closed]

    - by manix
    I'm so confused with a strange behaviour. Actually I have edited some php files remotely with my PhpDesigner8 (a php editor). Everything goes right, but when my teammates reopen the files that I have edited the source code have blank lines like below: class AdminController extends Controller { function __construct() { parent::__construct(); if (!$this->session->can_admin()) { show_error('Solo para administradores.'); } $this->load->library('backend'); } } Instead of class AdminController extends Controller { function __construct() { parent::__construct(); if (!$this->session->can_admin()) { show_error('Solo para administradores.'); } $this->load->library('backend'); } } Did you have experience these kinds of problems?

    Read the article

< Previous Page | 196 197 198 199 200 201 202 203 204 205 206 207  | Next Page >