Super Solutions For The Robots.txt

1 comment
Use robots.txt can prevent effective inbound link

The only thing about using robots.txt to block indexing of search engines is not only that is quite inefficient, but may also reduce the flow of inbound links. Locking a page using robots.txt, search engines are not indexing the content (or links!) Page. This means that if you have inbound links to the page, the link juice can not flow to other pages. It creates an impasse.


While inbound links to the blocked page is probably some benefit to the general area, this value inbound links are not used to their full potential. You missed an opportunity to convey a certain value link internal page blocked in several important internal pages.

3 Big Sites with Blocked Opportunity in the Robots.txt File

*1 - Digg.com

*2 - Blogger.com or Blogspot.com

*3 - IBM

Super Solutions to the Robots.txt

Great site, for example above, we have covered the wrong robots.txt file. Some of the scenarios were not included. The following is a list of effective solutions to maintain the contents index of search engines with no link juice to lose.

Noindex

In most cases the best alternative to robots.txt robots exclusion meta tags. By adding "noindex" and make sure it does not add 'nofollow' your pages will remain in the results of search engines, but will link value.
301 Redirect

The robots.txt file is not a place to list the old worn pages. If the page has expired (delete, move, etc) is not only the block. Redirect this page through a 301 to replace the most relevant. Get more information about redirecting the Knowledge Centre.

Canonical Tag

Do not block your overlap since the versions in robots.txt. Using the tag canon to keep the additional versions of the index, and consolidate the link value. Where possible. Get more information at the Information Centre on canonization and use the rel = tag canon.

Password Protection

The robots.txt is not an effective way to keep the information confidential at the hands of others. If you have confidential information on the Internet, password protect. If you have a login screen, go ahead and add meta tag "noindex" page. If you expect a lot of incoming links on this page for users, be sure to link to some of the most important pages of internal login page. This way, you pass through the link juice.

Effective Robots.txt Usage

The best way to use a robots.txt file does not use it at all. Use it to report that robots will have full access to all files on the site and to control a robot in the sitemap.xml file. That's it.

Your robots.txt file should look like this:

-----------------

User-agent: *
Disallow:

Sitemap: http://www.yoursite.com/sitemap.xml

-----------------

Bad Bots

"Robots and instructions for the robots.txt file," which means that there are robots that do not follow the robots.txt at all. So when you do a good job of keep away with a good, you are doing a horrible job to keep away from "bad" against. In addition to filtering to allow access only to the Google bot Bing is not recommended for three reasons:

1. The engines change/update bot names frequently.
2. Engines employ multiple types of bots for different types of content.
3. New engines/content discovery technologies getting off the ground stand even less of a chance with institutionalized preferences for existing user agents only and search competition is good for the industry.

Competitors

If your competitors are warned SEO in any way whatsoever, they look at your robots.txt file to see what they can discover. Say you are working on a new design or an entirely new product and you have a line in your robots.txt file that disallows bots "index" it. If a competitor appears, check the file and see this folder called "/ newproducttest" when they just won the jackpot! Better to keep it on a staging server, or behind a login. Do not give all your secrets in a small file.

Handling Non-HTML & System Content

* It isn't necessary to block .js and .css files in your robots.txt. The search engines won't index them, but sometimes they like the ability to analyze them so it is good to keep access open.

* To restrict robot access to non-HTML documents like PDF files, you can use the x-robots tag in the HTTP Header.

* Images! Every website has background images or images used for styling that you don't want to have indexed. Make sure these images are displayed through the CSS and not using the tag as much as possible. This will keep them from being indexed, rather than having to disallow the "/style/images" folder from the robots.txt.

* A good way to determine whether the search engines are even trying to access your non-HTML files is to check your log files for bot activity.

Robots.txt High Impact Solutions

1 comment :

  1. Someone advise me to put "blog labels" into Robots.txt. I don't know why? Is this correct?

    Regard
    Radhika

    ReplyDelete

Blogger news

Related Posts Plugin for WordPress, Blogger...

Labels

seo tips seo beginners seo techniques seo starter guide seo beginners tips PPC Tips ppc advertising tips seo updates seo beginners guide seo helps Dharmesh Talaviya SEO Executive adwords tips google algorithm google update ppc campaign tips ppc search engine tips ppc tricks seo guidelines 5 seo tips SEO Executive black hat seo collaboration microhoo google news joomla seo joomla seo tips joomla seo tools joomla seo tutorial linkwheel sites on page seo tips search engine marketing search engine optimisation search engine updates seo advise seo loophole seo news seo tips for beginners seo tools seo update tppc tips yahoo and bing collaboration "panda" update 17th jan 301 redirect Dharmesh Google Microsoft Buys 8 SEO Domain Names Online PPC Advertising tips Penalty Recover Talaviya adwords tools articles go viral automatically shares links basic principles of seo best blog sites best web 2.0 sites best web 2.0 sites for linkwheel best web 2.0 sites for seo bing bing search engine algorithm bing updates blog helps blog tips blogger tips blogs tricks bookmarking sites canonical tag canonical url content easily linkable cornucopia of updates dofollow social bookmarking site list drupal seo tips dynamic link building facebook preferences fastest growing search engine free banner ads free blog site list free social bookmarking site list freelance link building freelancing seo getting inbound links google algorithm news google algorithm update google algorithm update 2010 google algorithm update analysis google blog search google instant google instant messenger google instant search google jan 2013 update google latest news google search engine algorithm google search results update google search tips google search tricks google seo helps google seo search google seo search tricks google seo tools google seo update google update 2010 google's farmer update google's popular search tool google’s market share growing search engine in america high pr social bookmarking site list high pr social bookmarking sites how to improve search engine results increase rss subscribers instant messaging keyword analysis keyword discovery keyword popularity research keyword ranking keyword research keyword research articles keyword search landing page seo link bait content link bait post type linkwheel linkwheel site list linkwheel sites list list of blog sites list of social bookmarking sites making link building meta tag tips microblogiing microsoft update most popular organic seo columns networking events no follow code on page seo factors on page seo techniques online ppc courses tips page load time pay per click advertising popular organic seo columns ppc online info ppc software ppc tools ppc tools google quality backlink rganic seo columns robot .txt file robots.txt tutorial search engine algorithm search engine in the us search engine marketing strategies search engine news search engine optimization techniques search results ranking algorithm seo seo advice seo algorithm seo algorithm 2010 seo algorithm update seo beginners guidelines seo checklist seo columns seo daily update seo factors seo first step guide seo freelancer seo games seo guide seo helper seo microblogging seo news 2013 seo news jan 2013 seo people seo practitioners writing seo quiz seo search results seo search terms seo standpoint seo strategy seo tactics seo target keywords seo tasks seo tchnique seo update 2010 seomoz social media works social network websites social networking site social segments for tracking software giant study traditional marketing target keywords the difference between a like and a link twitter preferences upgrade to bing search engine url shorteners use search engines web 2.0 web 2.0 site list web 2.0 sites web 2.0 sites list word tracker yahoo news yahoo product lines yahoo to use bing search engine yahoo updates
.