Cloaking is a search engine optimization (SEO) technique in which the content presented to the search engine spider is different from that presented to the user’s browser.
Image Source: amitbhawani.com
Some examples of cloaking include:
Serving a page of HTML text to search engines, while showing a page of images or Flash to users
Inserting text or keywords into a page only when the User-agent requesting the page is a search engine, not a human visitor.
Inserting additional keywords or tags in the same background color so it is not shown to human but to the spiders and bots.
Link farm
A link farm is a group of websites where every website links to every other website, with the purpose of artificially increasing the PageRank of all the sites in the farm. They are considered as spamming technique and thus can get you penalized.
NoFollow
“Nofollow” provides a way for webmasters to tell search engines “Don’t follow links on this page” or “Don’t follow this specific link.”
Originally, the Nofollow attribute appeared in the page-level meta tag and instructed search engines not to follow (i.e., crawl) any outgoing links on the page. For example:
<meta name=”robots” content=”nofollow” />
You can also set ‘nofollow’ if you are linking to external links and do not want to endorse them or trust them such as:
<a href=”www.abc.com” rel=”nofollow”>ABC</a>
Sandbox
The Google sandbox refers to a commonly held belief that Google has a filter that places all new websites under restrictions for a certain amount of time to prevent them from ranking in searches. The idea behind the Google sandbox is that newer websites are not as relevant as older sites, and they are also more likely to be spam. For this reason, they are restricted and allowed to mature before being allowed to rank well.
Spiders/Bots
Spiders and bots are also called as web crawlers. These are computer programs that crawl the web on behalf of search engines, trying to discover new links and new pages. These crawlers help in indexing.
Indexing
Indexing is the search engines’ process for collecting and storing data across the web. The search engines are constantly scouring the web for updated and new pages to add to their massive databases of information. When the search engines do find new pages, they ‘index’ it, meaning they add a copy of it to their database, so that they can retrieve it during searches.
Duplicate Content
Duplicate content is content that appears on the Internet in more than one place. When there are multiple pieces of identical content on the Internet, it is difficult for search engines to decide which version is more relevant to a given search query. To provide the best search experience, search engines will rarely show multiple duplicate pieces of content and thus, are forced to choose which version is most likely to be the original.
Robots.txt
A robots.txt file is a file at the root of your site that indicates those parts of your site you don’t want accessed by search engine crawlers. The file uses the Robots Exclusion Standard, which is a protocol with a small set of commands that can be used to indicate access to your site by section and by specific kinds of web crawlers. You can see robots.txt file by appending robots.txt after domain name such as:
http://www.yourdomain.com/robots.txt
Web Directories
Web directories are collections of links which are organized by topic for the convenience of people browsing the Internet looking for information and resources. Using a web directory, Internet users can find lists of websites on related topics which complement each other. For example, someone who is interested in Blogging could use a web directory to find resources organized by type, such as Blogging tips, SEO and blogging platform etc.
Above the Fold
‘Above the fold” is the point on your website which the user sees without scrolling the browser window. Search engines place some priority on content above the fold since it will be seen right away by new visitors. Having too many ads above the fold can be seen as a negative issue, too.
Inbound Link
A link from one site into another. A link from another site will improve your SEO, especially if that site has a high PageRank.
Internal Link
A link from one page to another on the same website, such as from your homepage to your products page. Interlinking your posts and pages can help in decreasing the bounce rate of your blog.
Bounce Rate
It is also one of most important SEO terms. Bounce Rate tells what percentage of your visitors visit your blog but leave without going through your content or traversing other pages of your blog (i.e. they are just visiting landing page, and leave your site without opening second page of your site).
For example, if your site received 3, 00,000 visits in one month, out of which 1,50,000 bounced after visiting just one page, your bounce rate for that month would be 1,50,000 divided by 3,00,000, which equals to 0.50 (or 50%).
Long Tail Keyword
Google’s Hummingbird algorithm change has put Long Tail Keyword high up on your SEO practices. Long tail keywords usually contain more than three words and include specific words that users can use to fulfill their needs. Hummingbird changed the way search actually looks at keywords and phrases. Instead of picking apart the keywords within a phrase, it now understands entire questions or sentences within context.
Broken Link
Broken Link is a hyperlink which is not functioning, or a link that does not lead to the desired location. Most websites have some broken links, but if there are too many it can indicate poor content, poor upkeep of the site, or both. Too many broken links may cause search engines to rank a page as being less relevant.
Referrer String
A piece of information sent by an internet user when they navigate to your website from somewhere else on the internet. It includes information on where they came from previously, which helps webmasters understand how users are finding their website.
Landing Page
In SEO, a landing page is a website page that search engine users come to after clicking on an organic link. Creating persuasive, informative landing pages for strategically important keywords greatly improves lead generation. All your Landing Pages should be optimized to maximize Conversions.
Sitemap
A sitemap is a document created by a webmaster or a piece of software that provides a map of all the pages on a website to make it easier for a search engine to index that website. Search engines recommends specific sitemap attributes to enable it to better understand and rank a website. A properly structured sitemap has a large role in improving search engine visibility, especially for websites with many pages.
Disavow
Websites sometimes get links from harmful domains or low authority or spam links. While best practices are to try and get these links removed, this is not always possible. To help website owners, Google created a disavow tool that lets webmasters tell Google that they don’t want to recognize these links.
SEM
The term Search Engine Marketing refers to a broad range of activities relating to positioning a site within search engines ranks. SEM is divided into two main pillars: SEO and PPC. SEO is the practice of optimizing websites to make their pages appear in the organic search results. PPC stands for Pay-Per-Click, and it is the practice of purchasing clicks from search engines. The clicks come from sponsored listings in the search results.
Comments
SEO Services