Meta information
<head>
for the Search engine
<meta>
<meta name=”title” content= “keyword”/>
Title Tag
The title element of a web page is meant to be an accurate and concise description of a page's content. This element is critical to both user experience and search engine optimization. It creates value in three specific areas: relevancy, browsing, and in the search engine results pages. It is not relevant for Google.
Code Sample
<head>
<title>Example Title</title>
</head>
<meta name=”keywords” content=”keywords1, keyword 2 that is not important for Google
<meta name=”description” content=”keyword” />
Meta descriptions are HTML attributes that provide concise explanations of the contents of web pages. Meta descriptions are commonly used on search engine result pages (SERPs) to display preview snippets for a given page.
Code Sample
<meta name="description" content="This is an example of a meta description. This will often show up in search results.">
Optimal Length for Search Engines
The description should optimally be between 150-160 characters. Roughly 155 Characters
Avoid Duplicate Meta Description Tags
As with title tags, it is important that meta descriptions on each page be unique. One way to combat duplicate meta descriptions is to create a dynamic and programmatic way to make unique meta descriptions for automated pages.
Not a Google Rank Factor
Google announced in September of 2009 that neither meta descriptions nor meta keywords factor into Google's ranking algorithms for web search. Google uses meta descriptions to return results when searchers use advanced search operators to match meta tag content, as well as to pull preview snippets on search result pages, but it's important to note that meta descriptions do not to influence Google's ranking algorithms for normal web search.
<meta name=”author” content=”autorenname”>
<a href=’http://plus.google.com/ProfilID’ rel=’author’>uber mich</a>
<a href=’http://plus.google.com/seinenlID’ rel=’author’>Estefaniaojea</a>
Google crawlers
Crawler" is a generic term for any program (such as a robot or spider) used to automatically discover and scan websites by following links from one webpage to another. Google's main crawler is called Googlebot.
https://developers.google.com/webmasters/control-crawl-index/docs/getting_started?csw=1
Robots meta tag
Some pages use multiple robots meta tags to specify directives for different crawlers, like this: <meta name="robots" content="nofollow"><meta name="googlebot" content="noindex">
In this case, Google will use the sum of the negative directives, and Googlebot will follow both the noindex and nofollow directives. More detailed information about controlling how Google crawls and indexes your site.
Sitemaps.xml
Sitemaps are a way to tell Google about pages on your site we might not otherwise discover. In its simplest terms, an XML Sitemap is a list of the pages on your website. Creating and submitting a Sitemap helps make sure that Google knows about all the pages on your site, including URLs that may not be discoverable by Google's normal crawling process.
HTTP status codes
302 Found
The robots exclusion protocol (REP), or robots.txt is a text file webmasters create to instruct robots (typically search engine robots) how to crawl and index pages on their website.
Allow all web crawlers from all content
User-agent: * Disallow:
Block all web crawlers from all content
User-agent: * Disallow: /
Block a specific web crawler from a specific folder
User-agent: Googlebot Disallow: /no-google/
Block a specific web crawler from a specific web page
User-agent: Googlebot Disallow: /no-google/blocked-page.html
Allow a specific web crawler to visit a specific web page
Disallow: /no-bots/block-all-bots-except-rogerbot-page.html User-agent: rogerbot Allow: /no-bots/block-all-bots-except-rogerbot-page.html
Sitemap Parameter
User-agent: * Disallow: Sitemap: http://www.example.com/none-standard-location/sitemap.xml
Important Rules
- Only one "Disallow:" line is allowed for each URL.
- Each subdomain on a root domain uses separate robots.txt files.
- Spacing is not an accepted way to separate query parameters. For example, "/category/ /product page" would not be honored by robots.txt.
- In order to exclude individual pages from search engine indices, the noindex meta tag <meta name="robots" content="noindex"> is actually superior to robots.txt.
1. Page Level Factors
- Keyword in Title Tag
- Title that starts with keyword
- Duplicate content is bad
- Rel=canonical
- Keyword in H2, H3 Tags
- Proper Grammar and Spelling
- Too many broken links
- URLs length
2. Site Level Factors
- Content provides value
- Domain Trust
- Presence of Sitemaps
- Presence of youtube videos
- Site usability
3. Backlinks factors
- link location on page
- Backlink age
- Schema.org microformas ( webs with microformats ranks better)
4. User interaction
5. Special algorithm rules
6. Social signes
7. Brand signals
8. On Site Webspam factors
9. Off Page Web Spam factors