Search Results

Search found 16682 results on 668 pages for 'search engines'.

Page 243/668 | < Previous Page | 239 240 241 242 243 244 245 246 247 248 249 250  | Next Page >

  • To change url to user friendly url

    - by German
    I'm re-factoring my asp.net application from asp.net 3.5 to 4.0. Also I'm changing url to user friendly url. Example /product.aspx?id=100 to /product-name/100 All my pages indexed by search engines and the site already 6 years online. I'm planning to do 301 redirect from old pages to new one. I want to make sure I won't loose the rank and traffic. Any suggestion how to do it properly?

    Read the article

  • The SEO Checklist

    How many times have we heard the complaint that somebody owned a fantastic website, did excellent onsite optimization, built up hundreds to thousands of relevant back links but had nothing to show for his results? He was simply not ranking anywhere on the face of any of the search engines for his targeted keywords.

    Read the article

  • Google App Engine - SiteMap Creation for a social network

    - by spidee
    Hi all. I am creating a social tool - I want to allow search engines to pick up "public" user profiles - like twitter and face-book. I have seen all the protocol info at http://www.sitemaps.org and i understand this and how to build such a file - along with an index if i exceed the 50K limit. Where i am struggling is the concept of how i make this run. The site map for my general site pages is simple i can use a tool to create the file - or a script - host the file - submit the file and done. What i then need is a script that will create the site-maps of user profiles. I assume this would be something like: <?xml version="1.0" encoding="UTF-8"?> <urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9"> <url> <loc>http://www.socialsite.com/profile/spidee</loc> <lastmod>2010-5-12</lastmod> <changefreq>???</changefreq> <priority>???</priority> </url> <url> <loc>http://www.socialsite.com/profile/webbsterisback</loc> <lastmod>2010-5-12</lastmod> <changefreq>???</changefreq> <priority>???</priority> </url> </urlset> Ive added some ??? as i don't know how i should set these settings for my profiles based on the following:- When a new profile is created it must be added to a site-map. If the profile is changed or if "certain" properties are changed - then i don't know if i update the entry in the map - or do something else? (updating would be a nightmare!) Some users may change their profile. In terms of relevance to the search engine the only way a google or yahoo search will find the users (for my requirement) profile would be for example by means of [user name] and [location] so once the entry for the profile has been added to the map file the only reason to have the search-bot re-index the profile would be if the user changed their user-name - which they cant. or their location - and or set their settings so that their profile would be "hidden" from search engines. I assume my map creation will need to be dynamic. From what i have said above i would imagine that creating a new profile and possible editing certain properties could mark it as needing adding/updating in the sitemap. Assuming i will have millions of profiles added/being edited how can i manage this in a sensible manner. i know i need a script that can append urls as each profile is created i know the script will prob be a TASK - running at a set freq - perhaps the profiles have a property like "indexed" and the TASK sets them to "true" when the profiles are added to the map. I dont see the best way to store the map - do i store it in the datastore i.e; model=sitemaps properties key_name=sitemap_xml_1 (and for my map sitemap_index_xml) mapxml=blobstore (the raw xml map or ror map) full=boolean (set true when url count is 50) # might need this as a shard will tell us To make this work my thoughts are m cache the current site map structure as "sitemap_xml" keep a shard of url count when my task executes 1. build the xml structure for say the first 100 urls marked "index==false" (how many could u run at a time?) 2. test if the current mcache sitemap is full (shardcounter+10050K) 3.a if the map is near full create a new map entry in models "sitemap_xml_2" - update the map_index file (also stored in my model as "sitemap_index" start a new shard - or reset.2 3.b if the map is not full grab it from mcache 4.append the 100 url xml structure 5.save / m cache the map I can now add a handler using a url map/route like /sitemaps/* Get my * as map name and serve the maps from the blobstore/mache on the fly. Now my question is does this work - is this the right way or a good way to start? Will this handle the situation of making sure the search bots update when a user changes their profile - possibly by setting the change freq correctly? - Do i need a more advance system :( ? or have i re-invented the wheel! I hope this is all clear and make some form of sense :-)

    Read the article

  • what is optimum length for html title tag in Unicode format?

    - by user1501256
    I have a website that generates its title tag dynamically. the title tag is in unicode format. the title tag is limited to 65 character but sometimes Google doesn't show title tag completely in SERP. I'd like to know what is the optimum length of title tag in terms of seo for unicode titles, and is there any difference between Unicode title and non-Unicode title tag? And what about other search engines Bing, Yahoo and so on.

    Read the article

  • JQUERY AJAX, Issues with spaces being sent to the server, y?

    - by nobosh
    I'm using the following code to post to the server which is then sent to a MYSQL query to find matches via search. $.ajax({ url: '/search/spotlight/', data: "q=" + $(this).val(), success: function(data) { } }); When Q's val has spaces in it, it's creating problems. I'm wondering if I'm handling this correctly? Do I need to encode the value in the AJAX call? Or is this a problem on my backend, which is ColdFusion Right now JQUERY is posting the following to the server: /search/spotlight/?q=FirstName%20LastName is this right?

    Read the article

  • Added a tableview in a viewcontroller. How do I load tableview cells

    - by user1509593
    I have a view controller In storyboard, added tableview to view controller. Created an IBOutlet for tableview to View controller's header file. The view controller's header file includes an resultsarray Changed the @interface to include delegates <\UITableViewDelegate, UITableViewDataSource View controller implementation file has mandatory tableview protocols (NSInteger)numberOfSectionsInTableView:(UITableView *)tableView (NSInteger)tableView:(UITableView *)tableView numberOfRowsInSection:(NSInteger)section (UITableViewCell *)tableView:(UITableView *)tableView cellForRowAtIndexPath:(NSIndexPath *)indexPath The view controller has search button in storyboard and IBAction in view controller header file. On click of search button, results are obtained. How do I load the array and redisplay tableview with results in search method. After I loaded the results in array, I tried [self viewdidload];, hoping the cells will be loaded. But didn't. I thought of calling [self.detailView cellForRowAtIndexPath:?indexpath], but dontknow what the value of index path is. Appreciate help from guru's to load the UItableviewcell Thanks

    Read the article

  • How to Build Links & Improve PageRank

    Before we start talking about building links to improve Google PageRank, let's clear up any confusion - PageRank is a calculation of Google's estimate of the importance of a page, but it is not the same as where your page "ranks" in the organic results. It can be very confusing since the words are so similar. Try thinking of it this way: a page can rank in any of the search engines, but you can only have PageRank in Google.

    Read the article

  • SEO - Creating Backlinks

    You will learn how can you create effective backlinks to your site so that Search Engines will give you a good relevancy score. This article also share with you where can you get such valuable links.

    Read the article

  • How to Opt For C Class IP Addresses

    There are a great number of SEO Hosting Services in the world today that are formulated with the intention of being able to create and develop winning strategies for their websites to rank well on the top search engines of the world. The need for SEO Hosting has come at a time when webmasters are seeking ways in which they can assign multiple C Class IP Addresses to all of their domains and mange as well as create some of the best services that are able to know where webmasters can control all their domains from one single cPanel.

    Read the article

  • SEO Tips For Bloggers

    You will learn how to optimize your blog for popular Search Engines like Google and Yahoo in this article. You will also pick up what are the important areas to look out for when optimizing your blog.

    Read the article

  • SEO Tips For Bloggers

    You will learn how to optimize your blog for popular Search Engines like Google and Yahoo in this article. You will also pick up what are the important areas to look out for when optimizing your blog.

    Read the article

  • How to get the records using order by and so on

    - by paulrajj
    I have a table categories containing categories id having the records of 1 to 20. when i am doing the search query by using the IN function in mysql i got the results. but i am struggling to get the results using order by. The limit may be vary for every search as this is one of the input value. For example I have tried this query to find out the search results, select * from categories where category in (20,16,12,8) order by rand(), id limit 0,6 this query is executed and the results are in random category_id. the results will be, 8 12 16 20 and following this, another two records must be 8 12 If category_id contains only one record for 8 then, it should follow from 12,16. How can i achieve this ? thanks in advance.

    Read the article

  • High Value DoFollow Backlinks and Their Importance to Your Site Rankings

    For a website to be popular, it needs to be ranked rather high in the search engines. There are multitudes of ways this can be achieved. Among the most popular would be the amassing of backlinks to boost the site's popularity. Of course, not all backlinks have the same value and those procuring links for their site may wish to look towards dofollow backlinks as the prime links to acquire.

    Read the article

  • Refining Solr searches, getting exact matches?

    - by thebluefox
    Afternoon chaps, Right, I'm constructing a fairly complex (to me anyway) search system for a website using Solr, although this question is quite simple I think... I have two search criteria, location and type. I want to return results that are exact matches to type (letter to letter, no exceptions), and like location. My current search query is as follows ../select/?q=location:N1 type:blue&rows=100&fl=*,score&debugQuery=true This firstly returns all the type blue's that match N1, but then returns any type that matches N1, which is opposite to what I'm after. Both fields are set as textgen in the Solr schema. Any pointers? Cheers gang

    Read the article

  • How SEO Services Work to Improve Your Online Popularity

    There are assorted SEO Services available to help boost your online popularity. You might not know it but all these Internet Marketing strategies work together to make sure you appear on search engines, increase your rankings to outdo your competition, and of course, make a name for yourself on the Web.

    Read the article

  • SQL Conditional Select from SharePoint

    - by user3320324
    I'm creating a search page on our SharePoint site that will allow users to find results from a SQL table by either performing a search from a text box @Param1, or selecting a value from a dropdown @SiteParam. My SQL SelectStatement is SELECT * FROM [Routing] WHERE [Site] LIKE CASE WHEN @Param1 IS NOT NULL THEN '%' + @Param1 + '%' ELSE @SiteParam This works great in SQL when I declare the parameters and set values for the search, but not so much in SharePoint. I've been able to get it to work if I don't do the case and just use 1 of the parameters, but I haven't found a way to get it to do either of them. Any help is appreciated! Thanks

    Read the article

  • Link Tracking

    Finding the correct way to utilize your link tracking software is actually a very simple task for even the green SEO learner. Everyone knows that you want to have as many links as possible pointing to your site to increase your rankings in the SERP's thus increasing the amount of traffic coming to your site via major search engines.

    Read the article

  • Enumerating all hamiltonian paths from start to end vertex in grid graph

    - by Eric
    Hello, I'm trying to count the number of Hamiltonian paths from a specified start vertex that end at another specified vertex in a grid graph. Right now I have a solution that uses backtracking recursion but is incredibly slow in practice (e.g. O(n!) / 3 hours for 7x7). I've tried a couple of speedup techniques such as maintaining a list of reachable nodes, making sure the end node is still reachable, and checking for isolated nodes, but all of these slowed my solution down. I know that the problem is NP-complete, but it seems like some reasonable speedups should be achievable in the grid structure. Since I'm trying to count all the paths, I'm sure that the search must be exhaustive, but I'm having trouble figuring out how to prune out paths that aren't promising. Does anyone have some suggestions for speeding the search up? Or an alternate search algorithm?

    Read the article

  • SEO Article Writing Tips

    Article writing can generate thousands of target web visitors for your website. Much of the traffic comes from the search engines. So whether you know it or not, you are writing articles for SEO.

    Read the article

  • How Do External Links Help SEO?

    Starting at the beginning because without search engines there would be no optimisation. Google was founded by two Stanford University students in 1996 as part of a doctoral research project by Larry Page and Sergey Brin.

    Read the article

  • The Job of SEO Spiders

    The World Wide Web, also known as the Internet, is a very complex world. Search engines like Google need a software program that can read what's on the web. The said software program is known as bot or spider or crawler.

    Read the article

< Previous Page | 239 240 241 242 243 244 245 246 247 248 249 250  | Next Page >