Featured Post

Update: SEO Issues - is it Penguin? Is it Panda? or is it me?

It was a little over a year ago that I posted the " SEO Issues - is it Penguin? Is it Panda? or is it me? " in which I detailed o...

Tuesday, February 21, 2023

Back to the Basics of SEO #2 - Black Hat SEO

Black-Hat Techniques


When you look at how the search engines go about finding relevant websites when a certain word, or phrase is typed in, it can be fairly simple to trick that system into finding your website before anyone else's.

These techniques are generally called “Black-Hat Techniques”. They can range anywhere from doorway pages to hidden text, and more.

Some Black-Hat Techniques are:

Artificial Traffic – Artificial traffic systems are setup to hit your website with different IP addresses several times a day. The theory is that the more traffic that comes to your site, the more popular it is. So, by hitting your website website several times a day with different IP's, the search engine's see this as many people visiting your website every day. - In the end driving up your rankings with the search engines.

Why is this bad?

In the constant battle to weed out the good from the bad, search engines have been developed to recognize the programs used to generate artificial traffic. So, while it may work in the meantime in generating better rankings, in the long run you may end up getting penalized, or even banned by one or more of the search engines.

Cloaking Scripts – essentially outwits the search engines to increase your listings, and increase your traffic. Search engines spiders go to each site, follow every link, and index what they find in the engine's results. Cloaking Scripts automatically generate thousands of pages just for the spiders. These pages are usually dynamically built from keyword lists not available to the user.

Why is this bad?

Cloaking scripts are generally unreadable, or even have hidden text (text that is the same color of the webpage background, or hidden by some other feature such as an image, table or div), or redirecting scripts (scripts that send the visitor to the actual website from the cloaking page when the visitor runs their mouse over something, or with the click of a button and/or link within the cloaking page). The whole point of the search engines is to find legitimate content and quality websites for visitors to see. If they can't see the content on the page, or are redirected to a different page, that webpage is essentially thrown out as a quality website.

Doorway Pages - Doorway pages are usually developed as a page that is spider friendly for the user to see that generates rankings, and then redirects the visitor to the actual website. Doorway pages are created to do well for particular phrases. They are also known as portal pages, jump pages, gateway pages, entry pages and by other names. Doorway pages can be used for sites that aren't getting indexed (usually a site developed with frames, or dynamically driven), or they can be completely different domains that direct traffic to the actual website. For example, a lawyer might create a specific website with a doorway page for divorce law, and another one for criminal law, and another one for personal injury law. All three websites are optimized for each specific term, and then redirect to the original website that holds more information about that law firm.

Why is this bad?

Doorway pages do not have any significant content, and only have one or possibly two pages to the site. While they may have links to the site, the content may be the exact same as the original site, and all will link to one another. Search engines view this as spamming them (or tricking), and will eventually either penalize all of the websites, or completely ban them altogether.

Duplicate or Similar Content – In an effort to generate more content quickly, webmasters will sometimes create multiple pages, or even multiple websites, and then simply place the exact same content on each page with different keywords worked in.

Why is this bad?

Search engines can see this as spam. When a website has duplicate content it is either viewed as sheer laziness, or in breech of another's copyrights. In the end, the website that was created originally may not get penalized (generally looking at the older website, or how long a website has been live) though the newer websites will start to drop in rankings as a result.

While the search robots may simply be computers, and computers are only as smart as the ones that created them, search engines such as Google, MSN, and Yahoo spend all of their time, and energy from a staff made of those holding PhD's in computer science, mathematics, and more working on optimizing the search tools, and outsmarting those who create websites, and those who work on optimizing websites to find “ the most comprehensive search” on the web . In the end, the PhDs are most likely to win.

While the search engines may not be able to pick up on these black hat techniques, they may have a little help from users, or even your competitors. Each one of the major search engines (Yahoo, Google, and MSN) all have a form to fill out when it comes to websites that spam. Each and every website is the researched thoroughly, and all that are involved may potentially be penalized as well. So, be wary of those that you link to, and techniques that search marketing companies may use.

Link Farms – The more links you have pointing to your website, the better chance you have of getting a ranking. Although they aren't as around as much as they have been in the past link farms promise to place your link on other websites for a fee.

Why is this bad?

Link Farms often link your page with Web sites that have nothing to do with your content. The repercussions of this action are that the major search engines penalize sites that participate in link farming, thereby reversing their intended effect. A Link Farm usually places your link on a Web page that is nothing more than a page of links to other sites.

Where to report spammers:

Google - http://www.google.com/contact/spamreport.html

Yahoo - http://add.yahoo.com/fast/help/us/ysearch/cgi_reportsearchspam

Alta Vista - http://www.altavista.com/help/contact/search



Yahoo Copyright reporting –


http://docs.yahoo.com/info/copyright/copyright.html

Thursday, January 10, 2019

Making Decision When to Block in Robots.txt, 301, 404 or Ignore Errors and Warnings

Countless times I have been asked by a Boss, CEO or a Client why we see so many errors and warnings in Google Search Console, Moz, Conductor, Brightedge or Botify Reports; and what should we do about them. Often times the solution to the issue could be more damaging than the issue itself.

When Do We Address Errors and Warnings, and How Best to Deal with Them?

It's not an answer that is very straightforward, really. As any SEO will respond with "It Depends". It depends on so many factors going into the issue from cause, amount pages, impact on traffic and revenue those pages have, and so on.

I'll walk you through a decision making process that might help you make the best decision on what to do in these situations.

What are Errors/Warnings?

First, take a look at the issue that is being brought up. Are they Warnings or Errors? 

If you are seeing Errors, then there is a strong chance the issue should receive a priority, but don't go running up and down the hallways of the office screaming "Fire" just yet. We need to look as what exactly is going on before sounding the alarms.

Using one of my own sites as an example, Google Search Console sent me an email that there is an increase in Crawl Errors. I logged in, clicked on "Coverage" to find that the errors had a substantial increase. At some point every SEO has been through this same scenario. The key is to dig through and understand exactly what the error is, and start to identify what the cause is.

The few most common errors you will find in Google Search Console are:
  • Server error (5xx)
  • Submitted URL not found (404)
  • Submitted URL blocked by robots.txt
  • Submitted URL has crawl issue
  • Submitted URL seems to be a Soft 404
  • Submitted URL marked ‘noindex’
Iit's good to familiarize yourself with your GSC Errors before any such notifications are sent. URLs marked "noindex", blocked by robots.txt and not founf (404) are usually known issues that you can ignore. If you are going through your errors for the first time, it's good to spot check directories or sets of pages generated dynamically to understand what might be causing the errors. If they are something to be alarmed by, then it's a good time to discuss this with engineering, or if you develop yourself, to understand what the cause is and come up a best solution to fix.

Server error (5xx), seems to be a Soft 404, and has a crawl issue are all errors that should be addressed immediately. However, you should fully understand how many pages are causing the errors, what directory those pages are in, and what impact those pages are having before bringing them up with engineering, your boss, CEO or client.

Understanding the Impact

Let's say we have a set of pages in a directory that are causing 500 or soft 404 errors. In this case we see one real estate listing and a set of pages under the "events" directory. This is telling us that the system building out those events is having connectivity issues (server errors are usually a connection from web to database). 

When looking at the server logs for these pages, there is an issue with a javascript call on those pages that is causing the 500 error. An easy fix, but what would be the impact for the company?

Logging into Google Analytics and pulling up all pages that include "events" the session percentage of the total for the event pages in a 3 month span is 18% and last year for the same 3 months was 23% with a avg revenue value and conversion rate higher than any other set of pages for the site, and in general. These numbers show that the impact on the server error is potentially effecting the percentage of traffic negatively and the value of those pages is too great to let the error go. If the value of the pages were lower, and the YoY numbers were the same, then a lower priority on the fix could be set. 

Soft 404s could have the same impact as a server error, with a different cause and more involved fix. Soft 404s are usually caused from pages that throw a 200 (page is ok) code, but the page has little to no content on it. Google translates this a page that is in error and gives it a soft 404, rather than saying is doesn't exist.

Understanding server codes and reading your Google Analytics is extremely important in these cases. Furthermore, understanding web log data and being able to identify issues is just as equally important in determining an issue, level of effort to resolve, and impact to the business both negatively and positively (once resolved). 

With this knowledge, you can go into any meeting with your engineering team, your boss, CEO or client and articulate what it is you're seeing, demonstrate your knowledge of the issue, impact on the business and level of effort it will take to resolve. 

Monday, September 3, 2018

Linkless link building


Since its early days, Search Engines consider links as authority of website but now both giant search engines Google and Bing stated that they are now considering brand mentions without links as rank factors too along with links.

Gary Illyes from Google mentioned in his keynote at Brighton SEO in September 2017 that “If you publish high-quality content that is highly cited on the internet — and I’m not talking about just links, but also mentions on social networks and people talking about your branding, crap like that. Then you are doing great.”

Duane Forrester, formerly senior product manager at Bing, pointed out at SMX West 2016 that unlinked mentions can be as strong a signal as backlinks, confirming that search engines can easily identify mentions and use them to determine site authority.

Google and other search engines are intelligent enough now that they know what your brand is and when it is mentioned without links pointing to your website. Let's use Nordstrom, for example. My SEO team had a strategy to work with the PR team to manage mentions in our releases. Brand mentions in social media without links pointing back to the website, and many other ways in which Nordstrom is mentioned all provides credit to the brand. The more the brand is mentioned on the web from reputable sources the better.

If you're a smaller brand than Nordstrom a great way to get such mentions is by hiring a small PR consultant or firm. Mentions in local news, brand publications as a subject matter expert giving opinion pieces are all great ways to get brand mentions from reputable sources. Video transcripts for Vlogs, Podcasts and News Video Snippets all count as linkless link building for your brand. Search Engines look at the keywords surrounding your brand in these mentions and accredit your website as the source for those searches. If you're an SEO Agency, become an SEO subject matter expert on podcasts, news, vlogs, and online publications. The mention of SEO with your Brand (or your personal name) will connect the two and give your brand credit for those terms.

Of course, it's much more complex than that, but it's a great place to start. So, stop focusing on old school link building and get out there and become an authority.