Search Results

Search found 10402 results on 417 pages for 'macbook pro'.

Page 231/417 | < Previous Page | 227 228 229 230 231 232 233 234 235 236 237 238  | Next Page >

  • How does delicious.com avoid being sued for copyright infringement?

    - by Stanish
    With the recent redesign of delicious.com, they've added a much more graphical home page. The site continues to be a service for people to bookmark and share websites they come across on the web. The delicious home is now made up of images taken from those linked sites. See for yourself at http://delicious.com I would like to know what in the law allows them to do this, considering the images represent the main content of the page, and they clearly do not own copyright to those images? I know there is some leeway given to search engines where it is considered fair use to use a small portion of the content if the aim is to lead people to the originating site. Does that apply here?

    Read the article

  • What is required to create local business rich-snippets complete with sitelinks AND breadcrumbs?

    - by Felix
    I have a local business directory site. I would like to markup my business listing 'profile' level pages for display as enhanced listings/rich-snippets complete with business names, addresses and phone numbers. I would also like to display site-links and path-based breadcrumbs to help users navigate site directory hierarchy (which is deep). Is there a limit to the amount of breadcrumbs a site can leave? Is there a separate limit on the number of breadcrumbs which Google/Bing will display in the SERP? What kind of markup language(s) would be needed to best position my site to show site-links AND breadcrumbs? For example: Find a business Browse by Location State City Zip or Find a business Choose Service Browse by location State City Thanks all!

    Read the article

  • Microformats, Reviews and Duplicate Content

    - by Nicholas
    Let's say I have a site that sells widgets, and the URL structure is like so: /[type-of-widget]/[sub-type]/[widget-name]/ So, a URL for a widget might be: /screwdrivers/philips-screwdrivers/acme-big-screwdriver/ We show reviews on the widget page, and use the appropriate microformat data so Google knows it's a review, etc. Now, what if I want to show random reviews in the "sub-type" and "type-of-widget" landing pages? Will Google ding me for duplicate content, or is it smart enough to know (based on microformat data/etc.) that this is not duplicate content?

    Read the article

  • Is it bad SEO to embed an img tag in the middle of a sentence?

    - by Offlein
    I recently received a web mockup that included a sidebar with a short paragraph of narrow text; a quote from a person. At the end of the quote, the mockup had a circular portrait of the person floated to the right, hanging off the edge of the block, with the text flowing around the edge of it. It looked like this, where "TEXT" is the text, and "o" is the portrait TEXTTEXTTEXTTEXTTEX TEXTTEXTTEXTTE TEXTTEXTTEXT ooooo TEXTTEXTTEXT ooooooo TEXTTEXTTEXT ooooooo ooooo The easiest way to do this would be to put the tag for the portrait somewhere in the markup of the text, and float it right. But the HTML will look like, in that case: TEXTTEXTTE<img src="..." style="float: right;">XTTEXTTEXT and I fear this would be bad for SEO. Does anyone know? The other option is putting the at the end, absolutely positioning it, and using manual line breaks -- which sucks for my responsive design.

    Read the article

  • How do spambots work?

    - by rlb.usa
    I have a forum that's getting hit a lot by forum spambots, and of course the best way to defeat something is to know thy enemy. I'll worry about defeating those spambots later, but right now I'd like to know more about them. Reading around, I felt surprised about the lack of thorough information on the subject (or perhaps my ineptness to input the correct search terms for better google results). I'm interested in learning all about spambots. I've asked on other forums and gotten brush-off answers like "Spambots are always users registering on your site." How do forum spambots work? How do they find the 'new user registration' page? (I'm especially surprised because some forums don't have a dedicated URL for this eg, www.forum.com/register.html , but instead use query strings or even other methods invisible to the URL bar) How do they know what to enter into each 'new user registration' field? How do they determine what's a page they can spam / enter data into and what is not? Do they even 'view' this page at all? ..If not, then I'd assume they're communicating with the server directly - how is - this possible? How do they do it? Can forum spambots break CAPTCHAs? Can they solve logic questions (how?)? Math questions? Do they reverse-engineer client-side anti-bot validation scripts? Server-side scripts? What techniques are still valid to prevent them? Where do spambots come from? Is someone sitting behind the computer snickering as they watch their bot destroy site after site? Or are they snickering as they simply 'release' it onto the internet somehow? Are spambots 'run' by an infected computer somewhere? Do they replicate themselves? etc

    Read the article

  • Are Web Safe Colors Still Relevant?

    - by VxJasonxV
    I still remember one of my high school teachers lecturing us about the "web safe colors". A set of 216-256 colors that you should confine your designs to use, and nothing else besides them. Last I knew, Photoshop still has the "web safe" yield icon[1] on it's color picker. Are web safe colors still a concern? Outside of the obvious application (accessibility, legacy software versions, etc.), how much consideration should I give to limiting my color choice for my general audience? [1] Or was it the cube? I never remember.

    Read the article

  • Do I need to physically host my website in separate countries for SEO?

    - by noelmcg
    I run an ecommerce store that is hosted in Ireland, and ranks ok with google.ie The market for this comapny is the Republic of Ireland and the UK. Is it beneficial for me to have a UK hosted version of my site (.co.uk) to rank higher in google.co.uk (and other localised search engines of course). If so, how would I prevent the site from being punished for duplicate content? Thanks in advance for any assistance on the above.

    Read the article

  • seo value of duplicating content externally

    - by Don
    I run a website that includes a blog which was hand-coded by myself and is hosted on the same domain. My partner in this endeavour thinks it would be a good idea to open up a blogger/wordpress blog and duplicate the on-site blog on this off-site blog. AFAIK the main reason for doing this is the SEO benefits of the inbound links that this off-site blog will create. I think this is a bad idea, because: Effectively what we're doing is creating a (very small scale) link farm We're more likely to be punished than rewarded (in SEO terms) for duplicating our content across domains This introduces a problem of synchronising our content across domains. For example, if a blog post is edited on the on-site blog, then ideally the off-site blog should be similarly updated. I know very little about SEO, so would be interested to hear what more informed readers have to say.

    Read the article

  • Alternating links and slide down content on a list of sub sections

    - by user27291
    I have a page for a doctor's practice. In the summary page for the practice there is a list of subsections such as Women's, Men's, Children, Sport, etc. Some of these sub sections are very large, others can be a paragraph or more with a short unordered list. In terms of content volume, the large subsections warrant their own separate page, the smaller one's not so much. I created a little plugin which enables me to use the list of subsections in 2 ways. When clicking on the title of a larger section, you'll be sent through to it's own page. For the smaller sections, a slide down box will open with the information. Is this a good way to handle my information architecture? Should I be giving the smaller sub sections their own page for SEO purposes?

    Read the article

  • Google Analytics - New & Old Domain Filtering

    - by DotNetStudent
    I have two domains associated with the same hosting account. Domain A was the one I had purchased when I first set up my website, but I didn't like it all that much and it was getting low ranks, and so I bought domain B and 301'd links from domain A to domain B. Now I would like to know if there is any way I can filter page views on Google Analytics based on the domain used to access the website, so that I can know if it is worth to keep the old domain associated with the account. Is there any easy way I can do this?

    Read the article

  • To change url to user friendly url

    - by German
    I'm re-factoring my asp.net application from asp.net 3.5 to 4.0. Also I'm changing url to user friendly url. Example /product.aspx?id=100 to /product-name/100 All my pages indexed by search engines and the site already 6 years online. I'm planning to do 301 redirect from old pages to new one. I want to make sure I won't loose the rank and traffic. Any suggestion how to do it properly?

    Read the article

  • 302 Redirect Issue for Joomla 2.5.7 version site

    - by DDD
    For my site i am using Joomla 2.5.7 version and FB comments tools for the articles in the site. i am getting the 302 redirect problem for the FB comments for the Articles to which i post. I have checked the url's here http://www.webconfs.com/http-header-check.php and got the following result with 302 redirect. for http://www.fijoo.com HTTP/1.1 302 Moved Temporarily = Date = Wed, 21 Nov 2012 09:46:39 GMT Server = Apache/2.2.22 (Unix) mod_ssl/2.2.22 OpenSSL/1.0.0-fips mod_auth_passthrough/2.1 mod_bwlimited/1.4 FrontPage/5.0.2.2635 mod_perl/2.0.6 Perl/v5.10.1 X-Powered-By = PHP/5.3.16 Set-Cookie = =en-GB; expires=Wed, 21-Nov-2012 10:46:40 GMT LOCATION = / Content-Length = 0 Connection = close Content-Type = text/html How to overcome this anyone please help.

    Read the article

  • SEO For A Doman Name That Is A Dictionary Work Misspell [closed]

    - by Miles
    Possible Duplicate: How should I handle search engines auto-correcting the spelling of a site's name? My domain name is radiuus.com. Since the correct spelling of the word is with one u, when I search for "radiuus" on google it thinks I'm misspelling the word, kindly lets me know, and returns results for "radius". Example: https://www.google.com/search?q=radiuus&oq=radiuus&sugexp=chrome,mod=2&sourceid=chrome&ie=UTF-8 What can I do to prevent this? And taking this one step further: what can I do to get my site to show up when someone searches for "radius" spelled the correct way? Thanks!!

    Read the article

  • phishing attack. Where do I start the cleanup?

    - by Suz
    I'm a newbie webmaster. I've got a domain and a site... and no clue about the web (I'm OK with files and programs... ) I got a message from google that my site is a possible phishing site, with a link the the suspect page: http://www.mydomain.com/~phishers/Paypal/us/Confirm.php needless to say, I didn't put that up. Can someone point me to a good tutorial on what to do now? I'd like to figure out what happened so I can defend against it the next time around. How do I identify what kind of attach this is? Also, what is the tilde doing in the URL path? I couldn't find any path like this on my hosting account, so I'm not entirely sure how to delete it.

    Read the article

  • What is the benefit of the "download will begin shortly" page?

    - by Fammy
    I've noticed many websites that host files for downloads have an interstitial page between the download link/button and the actual start of the download. Terminology on the page may include "Your download will begin shortly. If it does not, try this direct link". What is the purpose of this page? It seems to draw away from the general experience of downloading a file. Is this beneficial for bookmarking? Less experienced users? Analytics?

    Read the article

  • Will my traffic come back after my site redesign?

    - by Steve
    I screwed up. I launched my site after rebuilding it without setting up the proper 301's and traffic immediately dropped about 60%(it's not really something I thought about). After about a week and a half, I set the 301's back up yesterday and resubmitted my sitemap to google. Google has yet to index the whole thing, but traffic isn't getting any better. Is it likely to come back? If so, I. How long? Has this happened to you? Any info is appreciated. I am really anxious!

    Read the article

  • Comparisons of Javascript 'data grids'?

    - by Joe
    I've found plenty of questions between here and StackExchange of people asking for the 'best' data grid / data table, or one that has a particular feature, and plenty of lists out there (of various ages) listing the various data grid implementations ... but is anyone aware of any matrix of what features the various solutions implement? (eg, allow shift-click to select multiple; support checkboxes for selection; can update a regular table in-place; allow editing of cells; support websql or indexeddb for local caching; which browsers they support; infinite scroll; etc.) There's a generic 'javascript framework' comparison on wikipedia, which would be the sort of thing I'm looking for, but it doesn't go into detail on data grids. (which makes sense, as so many are extensions, not core features of those frameworks, and in the case of jQuery, there's lots of 'em.)

    Read the article

  • Hosting server application for global SME

    - by BBe
    We are planning to set up a complete ERP and CRM system for a medium-sized global company, that might turn into a essential tool for all locations once deployed. For now these locations include USA, Germany, China and Indonesia, but the list is growing quickly. My question is, where it is best to physically locate the server to ensure the access times are optimal from all (future) locations? On my mind, I am dealing with multiple connected servers (a cloud?), where each of our users is served by the physically closest server. Being in a very competitive field we would also like to rule out, that any data is stored in mainland China... Thanks for any advice and pointers!

    Read the article

  • How to configure Google sitemap links in Wordpress? (without editing its HTML or PHP source code)

    - by Alexander Farber
    I run a Wordpress 3.7.1–de_DE site, but don't have much experience with it yet. When my site comes up in a Google search, there are 2 Google sitemap links displayed underneath: One of them points to a non-existent webpage /imprint though and I had to add a page at that URL to workaround this (and I want the URL actually be /impressum anyway since the site is in German and has German URLs). How to configure Google sitemap links in Wordpress (without editing its HTML or PHP source code)?

    Read the article

  • Google Analytics: Track user usage and flow

    - by Quintin Par
    Can someone help to query Google analytics to track a specific user behavior and usage pattern? Currently I pass user id’s to GA as _setCustomVar(2, 'id', id, 1); This is session based. But I am yet to master how I can utilize this to view usage pattern & behavior for the passed id. Say, I need to understand the visualization flow for one id or the page view count for that id etc Rephrasing, can I filter all existing reports for a specific id that I can select?

    Read the article

  • Which Hosting for a Mobile Infotainment Portal? [closed]

    - by VenomVipes
    Possible Duplicate: How to find web hosting that meets my requirements? I am building up a portal for music and movie downloads. My target is 85% Mobile User. I expect (acc to our promotion plan) 1000 Visitors a day and at least 300 parallel downloads. Files are mp3 & 3gp format. Please suggest the type of hosting Cloud/VPS/Dedi. Please suggest config (suppose I have 50Gb Video & 30Gb Audio). We have a budget of around 70$/m for Server. Also suggest if I should keep all files in my HostServer or just put links to direct downlad frm other sites offering the same video. My goal is Fast Loading on Mobile Browser (2G GPRS). Also faster download of files.

    Read the article

  • how does private sales ecommerce site work on their SEO?

    - by 142857
    In a private sales ecommerce site, users need to sign up/in before they can access the pages of website. So, even if a user tries to directly navigate to a product page, he is redirected to sign in. I am wondering then how does these sites manage their SEO, as it would imply google too can't crawl these pages, or do they completely ignore the SEO benefit of allowing google to crawl the product and catalogue pages?

    Read the article

  • PHP Code (modules) included via MySQL database, good idea?

    - by ionFish
    The main script includes "modules" which add functionality to it. Each module is set up like this: <?php //data collection stuff //(...) approx 80 lines of code //end data collection $var1 = 'some data'; $var2 = 'more data'; $var3 = 'other data'; ?> Each module has the same exact variables, just the data collection is different. I was wondering if it's a reasonable idea to store the module data in MySQL like this: [database] |_modules |_name |_function (the raw PHP data from above) |_description |_author |_update-url |_version |_enabled ...and then include the PHP-data from the database and execute it? Something like, a tab-navigation system at the top of the page for each module name, then inside each of those tabs the page content would function by parsing the database-stored code of the module from the function section. The purpose would be to save code space (fewer lines), allow for easy updates, and include/exclude modules based on the enabled option. This is how many other web-apps work, some of my own too. But never had I thought about this so deeply. Are there any drawbacks or security risks to this?

    Read the article

  • IP address and SEO

    - by Joel
    Hello, I currently host 5 websites within a dedicated server I own. I have several questions: Does it matter if I host all my sites on 100.100.100.100 (the server's IP for example) or if I split them into 100.100.100.100, 100.100.100.101 ... 100.100.100.104 (that is, each site on its own IP). Does it matter if I use a C-Class for each website? Do search engines really care if your site has its own c-class? Do search engines penalize a website if it moves its IP? Thanks, Joel

    Read the article

  • Can I benefit from links to pages on my site which have a `noindex` meta tag?

    - by Noam
    I'm trying to understand if/how I can benefit from people linking to pages on my site which are with pages that have a noindex meta tag. 2 actions I'm considering to perform: Remove the robots.txt disallow to these pages, to make sure inner links get the propagated link juice. Adding a canonical tag to the most similar page that doesn't have a noindex meta tag Are these valid approaches that might help? Any others I should consider?

    Read the article

< Previous Page | 227 228 229 230 231 232 233 234 235 236 237 238  | Next Page >