SEO has gone through extensive evolutionary changes over the years, and continues to do so every day.
While most traditional marketing tactics (for the most part) still hold true in digital marketing today, SEO changes have quite drastically changed the landscape.
Most, if not all, of these changes have helped improve the web – and search, in particular.
Yet, some people still cling to the “old ways” and try to use outdated SEO practices to improve their brand’s organic search visibility and performance.
Some of the tactics worked a few years ago, but now just aren’t as effective as they used to be.
Yet many novice marketers and/or small business owners are still using these “zombie” SEO techniques (tactics that should be dead, but aren’t for some godforsaken reason).
Not only are they ineffective, but many of the 12 outdated SEO practices below are potentially dangerous to the well-being of your brand, websites, and other digital properties.
1. Keyword Abuse
There are so many ways webmasters and “marketers” continue to misunderstand keywords’ role in general SEO initiatives, and how they should be used in the day-to-day strategy.
Let’s take a more granular look at specific types of keyword abuse and mismanagement, including irrelevant usage, writing for a specific keyword density, and keyword stuffing.
Irrelevant Keyword Targeting/Confusion
All too often, novice SEO practitioners try and fit their content and messaging within the confines of their keyword research (and not much else).
These “marketers” will shape the content and its metadata to represent keywords it’s not properly aligned with, nor the proper intent of the users conducting the searches for the high-volume keywords being targeted.
This causes brands to likely lose the attention of readers before ever having the chance to communicate a real message with them.
If the keywords marketed for don’t align with the content on the page, the disconnect will hinder the success of content even if it’s otherwise of good quality.
Don’t try to mislead users and direct them to content that is misrepresented by high-volume keywords in order for increased visibility.
Google knows what this looks like, and it can truly be defined as an obsolete SEO practice (as well as a “black hat” technique, in many instances).
Keyword Density
Writing for a specific “keyword density,” like many keyword-focused marketing tactics, is just missing the mark.
Google no longer depends on keyword density (or the ratio of specific keyword usage to the overall page copy) to determine whether a webpage is an effective source for answering a search query.
It is so much more advanced than simply crawling for keywords; search engines like Google use a multitude of signals to determine search results.
While keywords remain important to the topics and ideas they represent, they are not the lifeline for ranking for high-value search queries.
The quality of content and how the messaging is delivered are the lifeline for that.
Keyword Stuffing
This is probably the oldest trick in the book.
SEO is about keywords, right?
So, loading up our webpages with keywords — especially the same high-value keyword we are aggressively targeting throughout the website — is going to help us show up higher in search, thus outranking out competition?
Absolutely not.
Search engines have, for a long time, known what keyword stuffing is and what kind of text combinations are unnatural. They notice these as attempts to manipulate search results and demote the content as such.
Yes, there may still be valuable content that uses simple keyword stuffing, either intentionally or unintentionally, that is not demoted because of its actual value to users.
Back in the day, webmasters trying to game the system would go as far as putting every keyword variation of a high-value keyword in the website footer or, even more sketchily, make those keywords the same color as the site’s background, effectively hiding them from humans but not the search engine crawlers.
Webmasters have also tried this with links. (Don’t do anything like this.)
Remember, you’re writing for humans, not search engines.
2. Writing for Robots
It’s important to understand that writing unnatural is, well, not natural.
And search engines know it.
The belief is: writing for the web means we should repeat a subject by its proper name every time it is mentioned, working in variations and plural/non-plural versions of the word so that “all bases are covered.”
When crawled, the crawlers see the keyword repeated, and in several different versions, thus leading the page to rank well for the keyword variations used (over and over … and over again).
This isn’t going to work anymore.
Search engines are advanced enough to understand repeated keywords, their variations, and the unfavorable experience of generally bad content.
Write for humans, not search engine crawlers or any other robot.
3. Article Marketing & Article Directories
Any attempt to game the system doesn’t usually work out in the world of SEO.
But that doesn’t stop people from trying.
Especially when these tactics offer noticeable improvements to a brand, its website, and/or its associated digital properties.
Sure, article directories worked. And they worked pretty darn good for a long time, too.
Commonly considered one of earliest forms of digital marketing, article syndication was low-hanging fruit to those in the know. And it made sense since the idea was similar to other channels like TV and print that already use syndicated content regularly.
But Google eventually caught on, unleashing its game-changing Panda update in 2011.
Panda chewed up the search landscape, targeting content farms and directories, as well as other websites offering crap content (whether it was simply bad/false, horribly written, makes no sense, or stolen from someone else).
The idea behind article marketing doesn’t make sense in today’s world, where your high-quality content needs to be original and demonstrate expertise, authority, and trustworthiness.
4. Article Spinning
Typically done with software, article spinning is the black-hat tactic of trying to recreate quality content using different words, phrases, and organization.
Essentially the end result was a garbled mess of an article that made the same points as the source material.
It’s no surprise this isn’t effective anymore.
While AI is getting better all the time at creating content, anything generated by a machine is still of a lower quality than what a human can produce – something original, helpful, and of substance.
5. Buying Links
This one is still biting webmasters many years later.
Like most SEO tactics, if it seems shady, you probably shouldn’t do it.
Buying links is no different.
Once upon a time, it was routine practice to quickly pay to get a high volume of links pointing at your site.
Now we now that backlink profiles need to be maintained and optimized just like the websites we oversee, and low-quality domains with far too many backlinks pointing to a website may be dangerous to a website’s health.
Google can easily identify low-quality sites, and it will also identify when those sites are sending an abundance of links out that they shouldn’t be.
Today if you want to legitimately help boost the authority and visibility of your website, you need to earn links, not pay someone to build them manually.
6. Anchor Text
Internal linking is a characteristic of any good site structure and user experience.
This is typically done with anchor text, an HTML element that allows us to tell users what type of content they can expect if they click on a link.
There are various types of anchor text (branded, naked, exact-match, website/brand name, page title and/or headline, etc.), but some have most certainly become more favorable than others, depending on the usage and situation.
In the past, using exact-match and keyword-rich anchor text were standard SEO best practices.
Since Penguin, Google has been better at identifying over-optimized content.
This goes back to the Golden Rule about producing well-constructed content that is user-friendly and natural.
If you’re optimizing for search engines and not humans, you’re likely going to fail.
7. Obsolete Keyword Research Tactics
Keywords have certainly gone through some drastic changes over the last five to 10 years.
Marketers used to have a plethora of keyword-level data at their fingertips, allowing us to see what works well for our brand and what doesn’t, but also to get a better understanding of idea targeting and user intent.
Much of this went to the wayside with keyword “(not provided)”.
In the years following, tools popped up that tried to replicate keyword data. But to fully recreate it correctly is simply impossible.
And yet, even with that now-stripped keyword data, marketers are required to do keyword research of their own to get an understanding of the industry, the competition, the geographic region, etc.
To do this, many marketers turn to Google’s free Keyword Planner. While the data in there has been subject to some scrutiny over the years, it’s a free Google-owned product that gives us data we previously couldn’t really come by, so many of us continue to use it (myself included).
But it’s important to remember what the data actually represents for keywords.
“Competition” in the Keyword Planner pertains solely to paid competition and traffic, thus it is practically useless to build an organic search strategy around this data.
Some alternatives to this are the Moz Keyword Explorer tool and SEMrush’s Keyword Magic Tool, both of which are paid tools.
Google Trends is helpful for this type of competitive analysis, too, and it’s free.
8. Pages for All Keyword Variations
This was once a useful tactic to rank well for all the variations of high-value keywords targeted by your brand and its messaging.
Fortunately, algorithm updates like Hummingbird, RankBrain, and others have helped Google understand that variations of the same word are, in fact, all related to the same topic.
The best, most-useful content around these entities should be most visible due to the value it offers users on the topic, not just one variation of the word.
Aside from the fact that this will lead to brutal site self-cannibalization, it makes a website considerably harder to use and navigate since content will be so incredibly similar.
The negative user experience alone is reason enough not to do this. But the added fact that Google knows better than to overlook this practice makes it a no-brainer.
This tactic evolved and eventually helped lead to the inception of many content farms that were targeting traffic solely for their keyword value and visibility.
This was attributed to the “old way” of optimizing a website — for keywords and search engines, rather than users and their intent.
9. Targeting Exact-Match Search Queries
The tactic of targeting exact-match search queries in hopes to rank for those queries solely for the traffic numbers — and not because the search query or its answer actually pertained to the business optimizing for it — became a somewhat popular practice before the full deployment of the Google Knowledge Graph.
Marketers would strive to rank in the top spot for exact-match search queries to trigger a breakout box and an increased click-through rate for their sites.
10. Exact-Match Domains
Having high-value keywords in your URL makes sense. To some extent.
But when it becomes confusing or misleading (i.e., it results in a bad user experience), you have to draw the line.
A main best practice for domains is to keep it consistent with your brand.
Brand names should be short, concise, and somewhat meaningful.
Why wouldn’t you want the same from your domain?
Google would value exact-match domains a long time ago because it made sense to use it as a signal.
The behavioral data now has helped Google make changes like this (and many others) that are common sense, clean-up moves.
Run a good company and offer great products and/or services under the brand name, and Google will do work of making your brand visible when it’s relevant to the people searching for it.
11. XML Sitemap Frequency
We should never try to manipulate search engine crawlers so that our website is crawled more than others because it believed new content was published or substantial site changes were made.
But, since webmasters did that in the past, the sitemap is used quite differently than what was once intended.
Previously, webmasters could give a priority number to each page of a website listed in the sitemap ranging from 0.0 to 1.0.
Since that was never quite used correctly, crawlers don’t even honor the frequency rating.
Instead, search engines just crawl the content it deems it needs to crawl
Make sure you adhere to XML Sitemap best practices. Sitemaps are an incredibly important element for every website.
12. Bad Content
Face it. There was a time in our world when crappy content could still rank well.
Oh, how times have changed.
Stolen content, thin content, keyword-stuffed content, non-credible content — there was a time when all of this could get by search engine crawlers and regurgitated back to users as worthy results.
But no more.
We know what it takes to make quality content that is rewarded by search engines because they tell us what’s right and what’s wrong.
If you want to succeed at SEO today, you must do what’s right.