I have sat back and watched how the industry began, has grown, and developed throughout the years. On one hand it's been fun to be a part of something big that started from one company's idea and development that turned out an entire industry as a result. On the other had, because it is still a very young industry, and that industry is dictated by the company that sparked it, we are all still developing standards, strategies, and learning every day. In fact, just the other day I saw a post on Facebook for a workshop on how to use the Google Disavow Tool. It scares me to see SEO's already taking advantage of a strategy that is to not be taken lightly and making money off of "teaching" people on how to do it themselves. It's like a surgeon trying to teach a child how to patch up a kidney. Any seriously wrong move and the patient could die, and any slightly overlooked part of the process then the kidney could fail over time not knowing if it was the surgery or that the patient drinks vodka all day long.
In trying to learn and keep up with the latest in this capricious industry we often find ourselves having to look-up and research what the "experts" are talking about - those SEO buzzwords - coming across contradicting opinions, and quite frequently second guessing ourselves (even the highly experienced SEO experts second guess themselves). I have too often seen people trying to do what they think is right, and completely messing up their own site, and even client's websites because of all of the hype and misinformation out there.
The truth is that it is all viable, it's all in how you approach it. Of course, hearing that probably doesn't help, so the following are some of the most common strategies and some SEO buzzwords and hopefully clear up any confusion you might have.
Link Building
Yeah, I started with the most common, yet most controversial buzzword of all. The term "Link Building" began with the birth of the almighty Google itself. What was a very simple and quick way to get rankings for a website for the most popular search engine, is slowly becoming an art-form within itself. The basic idea is that a link from one site pointing to another site is counted as a "vote". The more links pointing from other sites to one site the more votes, and thus higher rankings. With such an easy strategy to implement, and the growing popularity of the search engine that uses the algorithm, more and more spammers began to take advantage. By offering website owners to pay money fro a link pointing to a website (a.k.a. purchasing links), asking a website to link to a site in exchange for a link back (a.k.a.link exchanging), submitting a website to directories (a.k.a. directory submissions), commenting on blog posts (a.k.a. commenting), and even submitting articles with links in them to article distribution sites - all of these means of obtaining links tricked the search engines into ranking websites that might not have otherwise deserved the positions they were given. In December of 2007 Google began cracking down on such strategies not only with increasingly new algorithms that catch sites that might be purchasing links, but by allowing webmasters to report one another manually. In the years since, we have seen a dramatic increase in the quality of the websites appearing in search results as a result.
SEO Buzzwords from Link Building:
Text Links
Links that point a page that contain a descriptive keyword or phrase. Many SEO's have used this strategy in the past because they give a context and a relevance to a link. This means that the search engines can read and index a page with all the text links and assign a ranking based on the quality if the content, the links and the destination of the links.
With Google's latest updates, the search engine no longer looks at the text within the link itself, but rather the words and relevance around the link. By recognizing that a page on a dog breed website with a link to a pet related website contains terms like "puppy", "hound", "paws", and other pet related terms that the dog site pointing to the pet site is, in fact, related. What I have seen in the past is a automotive website with an article on candy that contains one text link for "chocolate bon bons" and points to a chocolate website just isn't going to count (believe me, I've seen it). In fact, it will hurt the website's rankings.
Link Bait
The idea behind ‘link bait’ is to encourage people to bookmark or link to your website from theirs. Personal blogs, social media sites, and other communities will usually link to a site if the site offers something useful. Because of this, the search engines place a high value on the link.The best way to obtain these types of links is to write articles or white papers, a very valuable blog posting, or any sort of information your audience will find relevant. The more they share, the better the website ranks. The trick is to not force it - don't go out hiring people to share your posts, just let them happen naturally.
The idea behind ‘link bait’ is to encourage people to bookmark or link to your website from theirs. Personal blogs, social media sites, and other communities will usually link to a site if the site offers something useful. Because of this, the search engines place a high value on the link.The best way to obtain these types of links is to write articles or white papers, a very valuable blog posting, or any sort of information your audience will find relevant. The more they share, the better the website ranks. The trick is to not force it - don't go out hiring people to share your posts, just let them happen naturally.
Link Juice
The ‘search equity’ that is passed to one page from another is called "link juice". The more relevant a page is, how often it has been shared, and how many times it is visited places a high value from the search engines. From that page (or website) the pages that link from it will also gain extra value because the original content is deemed useful to users.
The ‘search equity’ that is passed to one page from another is called "link juice". The more relevant a page is, how often it has been shared, and how many times it is visited places a high value from the search engines. From that page (or website) the pages that link from it will also gain extra value because the original content is deemed useful to users.
Internal Linking
Almost self explanatory, most individuals tend to overlook the importance of linking within their own website. In fact, in most cases, the link to a page from a homepage can be just as valuable, if not more, than an external link. This, of course, does not mean that you should go adding a link to every page of your website from your homepage; nor does it mean you should link to a few pages, then rotate them, so that every page gets a chance at a high vote. What it means is that the pages that are most relevant to your users and make the most sense to continue from the homepage to, are the ones you should link to, and are the second most valuable pages (next to your homepage) that the search engines will rank.
Almost self explanatory, most individuals tend to overlook the importance of linking within their own website. In fact, in most cases, the link to a page from a homepage can be just as valuable, if not more, than an external link. This, of course, does not mean that you should go adding a link to every page of your website from your homepage; nor does it mean you should link to a few pages, then rotate them, so that every page gets a chance at a high vote. What it means is that the pages that are most relevant to your users and make the most sense to continue from the homepage to, are the ones you should link to, and are the second most valuable pages (next to your homepage) that the search engines will rank.
Taxonomy
Categorizing a website with a hierarchy and linking to one another internally is one of the best ways to show the search engines which pages are most important, and where they should rank. If a website is about cupcakes and selling supplies, the site should be organized by types of cupcakes (perhaps flavors) and then categories of supplies. Then place pages within that category that make sense. From there, pages should link to one another where relevant to show the search engines that this is X category and a set of pages, and this is Y category with a set of pages.
Often overlooked by agencies simply due to the fact that so many clients will hire an agency to "optimize" their site only to tell them in the end that they don't have the resources to make the suggested changes, or that they simply just can't make changes (whether it be because of design, usability, business reasons, etc). Unfortunately this leaves agencies in the predicament that they have to please the client and do what they were hired to do (which is to get the website rankings and increase traffic) but left with no other choice but to start link building. But a good SEO knows that internal optimization is really the heart and soul into obtaining legitimate rankings that will stick throughout all of the spam algorithm updates like Panda and Penguin. Below is a quick list and brief explanations for internal optimization.
Title Tag - This often shows up as the title in your search engine result. The title tag should never be more than 70 characters, and should only contain your most broad term that describes your website.
Description Tag - The description tag will often appear in the search result as the description text if the key term searched is within the tag. If not, then the search engine will pull from the content on the page itself where the key term is located. If a page on your site is specific to a certain term, then this is a good time to get that term within the description.
Keyword Tag - The keyword meta tag was once the main source of how search engines determined what site would show up for what search. Now it isn't as relevant, but is still used by some meta crawler search engines (not Google - but Excite, and often Bing). List out a few of your target terms for the page you are optimizing to help you focus on what you want the page to rank for, and just in case a search engine is paying attention.
Keyword density in document text - simply put, search engines look at how often a term shows up within the content of a page. If a word is mentioned 10 times within 300 words on a page, then the page won't get very good rankings. If a word is mentioned 10 times within 1200 words and spread out once or perhaps twice in a paragraph or two, then that page is more likely to rank better. A quick way to check densities is to put the content of a page within Microsoft Word, do a search within the document (Find), type in the word, and click "Highlight All". it's a great visual to see where a term is placed.
Content around the anchor text - As mentioned earlier, the words and context around an internal link is representative of the relevance of that page. The more a page will have of terms similar in context to the term you are optimizing for, the better.
Unique content - Any content borrowed, rented, or just stolen is considered a felony in the SEO world. There are algorithms in place that look for not only content within a site that exists elsewhere on a site, but content that exists on other sites as well. A quick way to check to see if your site has unique content is by searching on copyscape.com. Content that you have on yoru site that exists on other pages (or every page) will simply just not get counted (sort of just overlooked by the search engine), so any key terms within duplicate content on your site won't count. Duplicate content outside of your website is another story. If a website has content that you have copied (in other words, they had it first) then your site will get penalized. If your site had content first, and then someone copied you, then they would get penalized.
Frequency of content change - Search engines don't know the difference between a blog, a new publication, or a brochure-ware site that remains static. The best way they have developed to recognize a cutting edge news site and a static site, is how often new content is generated. The more often a new page is created with a robust amount of text, the more the search engine will come back and index, and therefore the higher the priority those new pages will get. If your site is something that is updated often, and is generating new content regularly, then the search engines will adjust accordingly. If your site is static, then don't worry, let it be, and the age of the pages will determine where they belong in the world of rankings (mentioned later).
Anchor text has key term(s) in links - What was a solid strategy of obtaining rankings for key terms in the past, is now less relevant, and even considered bad SEO. It's more about keyword "essence" and the relevance of the terms around the anchor text, than the anchor text itself (as mentioned above). Some of the more experienced SEO's are even finding that linking the word "more" or "click here" are helping their rankings more so than putting the key term within the anchor text.
Duplicating content - As mentioned before in the "Unique Content" bullet item, duplicating content on a site, or from another site is a very bad technique.
Invisible text - Nope, don't use white text on a white background with a bunch of keywords in it that only the search engine can see. Even 1 pixel high div's with the overflow hidden set in the stylesheet is a bad thing. Not only will you not get rankings, but your site will get penalized for it.
Age of website - the older a domain (or website) is, the higher a priority it will get within search rankings. A typical spam strategy is to buy a new domain and optimize it as much as possible to obtain quick rankings. Because of this, search engines will tend to ignore a website until it has been around for a few weeks, sometimes even months or years. If you have an older domain, then don't go thinking you should change it because it's "stale", it's actually a good thing.
Poor coding and design - Search engines can't tell what good design is, but they can tell from the popularity of the website. Social sharing, articles, blog posts, and all of the buzz about a website will only happen when a website is easy for the visitor to use, and gives all of the value a user is looking for. So, make sure your website is easy on the eyes, gives a clear and concise value proposition with a call to action, and is easy to navigate.
Exact Match Domain - Many spammers create website with a descriptive key term in the domain in attempts to get rankings. Google announced in October of 2012 that they were updating with an algorithm that will weed out any exact match domains. For example: http://www.compareinterestrates.com/ or http://www.best-interest-mortgage-rates.com/
Keyword-rich URLs and filenames - Just as the exact match domain is taking a hit in the recent updates, the keyword rich URL and filename strategy is as well. SEO's used to put their keyword within the URL with dashed between words in order to obtain ranking for long tail terms.
Site Accessibility - it's not talked about often, but can be potentially beneficial when your website is designed with accessibility in mind. Someone that has poor vision, hard of hearing, or may have trouble clicking links and buttons, is going to have trouble with most websites. If your website audience contains users that might need some extra help, keep this in consideration. Search engines know, and it could help you rank over your competition that hasn't.
Website size - Big or small, size doesn't matter. Some SEO's stress that a website needs to have millions upon millions of pages, but I have often personally witnessed websites that get penalized for having too many pages. Don't let this happen to your site, keep the pages down to a manageable and reasonable number. If your site is a publication with thousands or even hundreds of thousands of pages with unique content, then you should be fine. Just watch your webmaster tools notifications. Most of the websites that trigger the warnings are ecommerce websites with masses of pages for each product. If you find your site is showing this kind of error, it's best to seek out an experienced professional to help you get your pages under control and managed properly.
Domains versus subdomains - A subdomain is a subset of a main domain. Often used as a place to store images, or for other purposes, a subdomain looks something like images.mysite.com. Too often websites will put their highly valuable unique content of their blog on a subdomain. Unfortunately search engines don't know the difference between a main domain and the subdomain. Because of this, they treat each one as a separate entity. In the past SEO's have taken advantage of this and tried to get multiple rankings on one page with multiple subdomains. Just this year (2012) Matt Cutts has announced that they no longer treat them separately for separate rankings, but rather as an extension of the main domain. Because of this, subdomains not only won't see rankings, but the content is still not counted as part of the main domain. When setting up a blog, or any section of your website, it's best to simply just add a new directory (ex: www.mysite.som/blog) so that any of the content within that directory supports the domain as a whole.
Hyphens in URLs - When creating URLs for your website, it's still considered best practice to separate each word with a hyphen rather than a space, or an underscore. For example, if you write a blog post or article titled "The ten best puppies everyone should own" the URL should be "www.mysite.com/the-ten-best-puppies-everyone-should-own.html" or to avoid getting pegged for keyword rich URLs and a set hierarchy, it should be "www.mysite.com/puppies/ten-best.html".
URL length - A URL that is too long is a red flag for a keyword rich URL. try to keep your URL simple, and keep that site hierarchy.
IP address - The IP address is the unique identifying number (like a phone number) of where the server that hosts your website is located. If you are targeting a local audience, or maybe even just focusing on one country, be aware of where your website is hosted. A website that targets users searching in Canada, and is hosted in the U.S. will have an IP that resides within the U.S. In this case, search engines will only rank the site for U.S. searchers, and not for their Canadian searchers. If you aren't' worried about focusing your SEO by location, then don't worry about your IP.
robots.txt - The robots.txt file is a very simple text file (like Notepad) that resides on the main server. The only case in which you need a robots.txt is when you want to block certain sections of your website. Some search engines will allow you to put links to your xml sitemap for better indexing. For more information on setting up your robots.txt you can visit robotstxt.org.
XML Sitemap - Sitemaps are an easy way to let search engines know about all of the pages within your website that you would like to see indexed.
Redirects (301 and 302) or Status Codes - 404, 301, 302... Each one of these numbers has a different meaning to a search engine. The most common is a 404 or "page not found" it basically means that the UIRL existed, and now it doesn't. In the SEO world, the 301 is another code that is mentioned often. A 301 one lets the search engine know that the URL existed and has been moved, so we let the search engine know by redirecting the old URL to the new URL. My favorite explanation of these codes is from a dear friend of mine Lindsay Wassell at SEOmoz in which she uses pictures to best explain the different codes, and what they mean.
Long Tail - A long tail is what most SEO refer to when talking about a 3-5 or more word term. When a user is looking to buy a computer and begins their search with the word "computers", they will often start to get specific as they search focusing on the specifics like "500 GB laptop computer". This is what a long tail key terms is - the more specific you can target your audience, the more likely they will be to convert as they find what they are looking for.
Indexed - Indexing is a term SEO's use when a search engine has crawled a website and it's pages, and then starts to display them within the search results. This doesn't effect rankings, but merely expresses that a page is within the database, and recognized by the search engine. A quick and easy way to see if your website is indexed is to search with site: before your domain. For example: search for "site:oceansofpets.com".
SERP - Simply meaning the "search engine results page" and rolled off of the tongue of SEO's quite often. Pronounced just as it looks (serp) the search engine results page is the page that the user sees after completing the search.
Snippet - A search snippet is what SEO's use to describe the title and description a search engine displays on the search results page.
I think that should just about do it to get you started. With SEO there is no standard way of doing things. There is no true right and no true wrong, there is only what we try, fail or succeed, and try again.
Please feel free to add anything I might have missed in the comments below. I'm hoping this will become a pretty comprehensive list that newbie SEO's can get started with.
Categorizing a website with a hierarchy and linking to one another internally is one of the best ways to show the search engines which pages are most important, and where they should rank. If a website is about cupcakes and selling supplies, the site should be organized by types of cupcakes (perhaps flavors) and then categories of supplies. Then place pages within that category that make sense. From there, pages should link to one another where relevant to show the search engines that this is X category and a set of pages, and this is Y category with a set of pages.
Internal Optimization
Metatags
Description Tag - The description tag will often appear in the search result as the description text if the key term searched is within the tag. If not, then the search engine will pull from the content on the page itself where the key term is located. If a page on your site is specific to a certain term, then this is a good time to get that term within the description.
Keyword Tag - The keyword meta tag was once the main source of how search engines determined what site would show up for what search. Now it isn't as relevant, but is still used by some meta crawler search engines (not Google - but Excite, and often Bing). List out a few of your target terms for the page you are optimizing to help you focus on what you want the page to rank for, and just in case a search engine is paying attention.
Content
Content around the anchor text - As mentioned earlier, the words and context around an internal link is representative of the relevance of that page. The more a page will have of terms similar in context to the term you are optimizing for, the better.
Unique content - Any content borrowed, rented, or just stolen is considered a felony in the SEO world. There are algorithms in place that look for not only content within a site that exists elsewhere on a site, but content that exists on other sites as well. A quick way to check to see if your site has unique content is by searching on copyscape.com. Content that you have on yoru site that exists on other pages (or every page) will simply just not get counted (sort of just overlooked by the search engine), so any key terms within duplicate content on your site won't count. Duplicate content outside of your website is another story. If a website has content that you have copied (in other words, they had it first) then your site will get penalized. If your site had content first, and then someone copied you, then they would get penalized.
Frequency of content change - Search engines don't know the difference between a blog, a new publication, or a brochure-ware site that remains static. The best way they have developed to recognize a cutting edge news site and a static site, is how often new content is generated. The more often a new page is created with a robust amount of text, the more the search engine will come back and index, and therefore the higher the priority those new pages will get. If your site is something that is updated often, and is generating new content regularly, then the search engines will adjust accordingly. If your site is static, then don't worry, let it be, and the age of the pages will determine where they belong in the world of rankings (mentioned later).
Anchor text has key term(s) in links - What was a solid strategy of obtaining rankings for key terms in the past, is now less relevant, and even considered bad SEO. It's more about keyword "essence" and the relevance of the terms around the anchor text, than the anchor text itself (as mentioned above). Some of the more experienced SEO's are even finding that linking the word "more" or "click here" are helping their rankings more so than putting the key term within the anchor text.
Duplicating content - As mentioned before in the "Unique Content" bullet item, duplicating content on a site, or from another site is a very bad technique.
Invisible text - Nope, don't use white text on a white background with a bunch of keywords in it that only the search engine can see. Even 1 pixel high div's with the overflow hidden set in the stylesheet is a bad thing. Not only will you not get rankings, but your site will get penalized for it.
Overall Website
Poor coding and design - Search engines can't tell what good design is, but they can tell from the popularity of the website. Social sharing, articles, blog posts, and all of the buzz about a website will only happen when a website is easy for the visitor to use, and gives all of the value a user is looking for. So, make sure your website is easy on the eyes, gives a clear and concise value proposition with a call to action, and is easy to navigate.
Exact Match Domain - Many spammers create website with a descriptive key term in the domain in attempts to get rankings. Google announced in October of 2012 that they were updating with an algorithm that will weed out any exact match domains. For example: http://www.compareinterestrates.com/ or http://www.best-interest-mortgage-rates.com/
Keyword-rich URLs and filenames - Just as the exact match domain is taking a hit in the recent updates, the keyword rich URL and filename strategy is as well. SEO's used to put their keyword within the URL with dashed between words in order to obtain ranking for long tail terms.
Site Accessibility - it's not talked about often, but can be potentially beneficial when your website is designed with accessibility in mind. Someone that has poor vision, hard of hearing, or may have trouble clicking links and buttons, is going to have trouble with most websites. If your website audience contains users that might need some extra help, keep this in consideration. Search engines know, and it could help you rank over your competition that hasn't.
Website size - Big or small, size doesn't matter. Some SEO's stress that a website needs to have millions upon millions of pages, but I have often personally witnessed websites that get penalized for having too many pages. Don't let this happen to your site, keep the pages down to a manageable and reasonable number. If your site is a publication with thousands or even hundreds of thousands of pages with unique content, then you should be fine. Just watch your webmaster tools notifications. Most of the websites that trigger the warnings are ecommerce websites with masses of pages for each product. If you find your site is showing this kind of error, it's best to seek out an experienced professional to help you get your pages under control and managed properly.
Domains versus subdomains - A subdomain is a subset of a main domain. Often used as a place to store images, or for other purposes, a subdomain looks something like images.mysite.com. Too often websites will put their highly valuable unique content of their blog on a subdomain. Unfortunately search engines don't know the difference between a main domain and the subdomain. Because of this, they treat each one as a separate entity. In the past SEO's have taken advantage of this and tried to get multiple rankings on one page with multiple subdomains. Just this year (2012) Matt Cutts has announced that they no longer treat them separately for separate rankings, but rather as an extension of the main domain. Because of this, subdomains not only won't see rankings, but the content is still not counted as part of the main domain. When setting up a blog, or any section of your website, it's best to simply just add a new directory (ex: www.mysite.som/blog) so that any of the content within that directory supports the domain as a whole.
Hyphens in URLs - When creating URLs for your website, it's still considered best practice to separate each word with a hyphen rather than a space, or an underscore. For example, if you write a blog post or article titled "The ten best puppies everyone should own" the URL should be "www.mysite.com/the-ten-best-puppies-everyone-should-own.html" or to avoid getting pegged for keyword rich URLs and a set hierarchy, it should be "www.mysite.com/puppies/ten-best.html".
URL length - A URL that is too long is a red flag for a keyword rich URL. try to keep your URL simple, and keep that site hierarchy.
IP address - The IP address is the unique identifying number (like a phone number) of where the server that hosts your website is located. If you are targeting a local audience, or maybe even just focusing on one country, be aware of where your website is hosted. A website that targets users searching in Canada, and is hosted in the U.S. will have an IP that resides within the U.S. In this case, search engines will only rank the site for U.S. searchers, and not for their Canadian searchers. If you aren't' worried about focusing your SEO by location, then don't worry about your IP.
robots.txt - The robots.txt file is a very simple text file (like Notepad) that resides on the main server. The only case in which you need a robots.txt is when you want to block certain sections of your website. Some search engines will allow you to put links to your xml sitemap for better indexing. For more information on setting up your robots.txt you can visit robotstxt.org.
XML Sitemap - Sitemaps are an easy way to let search engines know about all of the pages within your website that you would like to see indexed.
Redirects (301 and 302) or Status Codes - 404, 301, 302... Each one of these numbers has a different meaning to a search engine. The most common is a 404 or "page not found" it basically means that the UIRL existed, and now it doesn't. In the SEO world, the 301 is another code that is mentioned often. A 301 one lets the search engine know that the URL existed and has been moved, so we let the search engine know by redirecting the old URL to the new URL. My favorite explanation of these codes is from a dear friend of mine Lindsay Wassell at SEOmoz in which she uses pictures to best explain the different codes, and what they mean.
Some basic SEO buzzwords
Indexed - Indexing is a term SEO's use when a search engine has crawled a website and it's pages, and then starts to display them within the search results. This doesn't effect rankings, but merely expresses that a page is within the database, and recognized by the search engine. A quick and easy way to see if your website is indexed is to search with site: before your domain. For example: search for "site:oceansofpets.com".
SERP - Simply meaning the "search engine results page" and rolled off of the tongue of SEO's quite often. Pronounced just as it looks (serp) the search engine results page is the page that the user sees after completing the search.
Snippet - A search snippet is what SEO's use to describe the title and description a search engine displays on the search results page.
I think that should just about do it to get you started. With SEO there is no standard way of doing things. There is no true right and no true wrong, there is only what we try, fail or succeed, and try again.
Please feel free to add anything I might have missed in the comments below. I'm hoping this will become a pretty comprehensive list that newbie SEO's can get started with.