Monday, 10 July 2023

Advanced Technical SEO

 


CONTENT

CHAPTER 1 SEO-Friendly Hosting: 5 Things to Look for in a Hosting

Company

By Amelia Willson

CHAPTER 2 The Ultimate Guide for an SEO-Friendly URL Structure

By Clark Boyd

CHAPTER 3 How to Use XML Sitemaps to Boost SEO

By Jes Scholz

CHAPTER 4 Best Practices for Setting Up Meta Robots Tags &

Robots.txt

By Sergey Grybniak

CHAPTER 5 Your Indexed Pages Are Going

Down – 5 Possible Reasons Why

By Benj Arriola

CHAPTER 6 An SEO Guide to HTTP Status

Codes

By Brian Harnish

CHAPTER 7 404 vs. Soft 404 Errors: What’s

the Difference & How to Fix

Both

By Benj Arriola

CHAPTER 8 8 Tips to Optimize

Crawl Budget for

SEO

By Aleh Barysevich

INTRODUCTION Technical SEO Is a Necessity, Not an Option

By Andy Betts

Monitor

your website’s

performance

at scale

Monitor your sites technical health

with DeepCrawl. Maintain sound

site architecture, proper internal

linking and provide positive user

experiences for visitors across

both mobile and desktop devices.

Javascript rendered crawling

Customized dashboards

Automated workflows

Task management

@DeepCrawl

+44 (0) 207 947 9613

https://www.deepcrawl.com

Sign up for a free trial today

CONTENT

CHAPTER 10 HTTP or HTTPS? Why You Need a Secure Site

By Jenny Halasz

CHAPTER 11 How to Improve Page Speed for More Traffic &

Conversions

By Jeremy Knauff

CHAPTER 12 7 Ways a Mobile-First Index Impacts SEO

By Roger Montti

CHAPTER 13

The Complete Guide to

Mastering Duplicate

Content Issues

By Stoney G deGeyter

CHAPTER 14 A Technical SEO Guide to

Redirects

By Vahan Petrosyan

CHAPTER 15 SEO-Friendly Pagination:

A Complete Best Practices

Guide

By Jes Scholz

CHAPTER 16

What Is Schema

Markup & Why It’s

Important for SEO

By Chuck Price

CHAPTER 9 How to Improve Your Website Navigation: 7 Essential

Best Practices

By Benj Arriola

CONTENT

CHAPTER 18 Understanding JavaScript Fundamentals: Your Cheat

Sheet

By Rachel Costello

CHAPTER 19 An SEO Guide to URL Parameter Handling

By Jes Scholz

CHAPTER 20 How to Perform an In-Depth Technical SEO Audit

By Anna Crowe

CHAPTER 17 Faceted Navigation: Best Practices for SEO

By Natalie Hoben

Technical SEO Is a

Necessity, Not an

Option

Introduction

Written By

Executive & CMO Advisor

Andy Betts

The practice of SEO has

changed more than any other

marketing channel over the

last decade. NEW WAYS TO APPROACH TECHNICAL SEO: A

NECESSITY, NOT AN OPTION

Through a succession of algorithmic evolutions, SEO has also

remained the foundation of a successful digital strategy – 51 percent of

online traffic arrives at websites via organic search, after all.

SEO has gone mainstream.

Still, we must take stock of the fact that SEO in 2018 requires new

skills and approaches to succeed in an increasingly competitive

world.

With more than 5,000 devices integrated with Google Assistant and

voice search on the rise, the focal points of search have become

decentralized.

The SERP as we knew it is long gone; search is dynamic, visual, and

everywhere now.

This has a very significant impact on organizations, as SEO is

a collaborative discipline that requires a synthesis of multiple

specialisms to achieve optimal results. At the heart of this lies the

domain of technical SEO, which has remained the foundation upon

which any successful strategy is built.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

A Brief History of Technical SEO

All roads lead back to technical – it’s how you now use your skills that

has changed.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

SEO has always entailed driving high-quality traffic through organic

search.

The means of achieving this goal have altered significantly since the

early days of SEO, when technical skills were dominant.

Crawlability was then – as it is now – a foremost consideration

when setting up an SEO strategy.

Content was secondary – a vehicle to include keywords and improve

rankings. This evolved over time to encompass link building, based

on Google’s key early innovation of using links to evaluate and rank

content.

The goal of marketers remained constant: to attract organic search

traffic that converted on their website.

As a result, we endured a cat and mouse game with some marketers

doing whatever it took to gain high search rankings.

As soon as Google caught up with keyword cloaking, black hat SEO

practitioners moved on to link buying in an attempt to manipulate

their rankings.

The Panda

and Penguin

algorithm

updates put

paid to a lot

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

In the process, SEO has moved from an entirely rational discipline

to something more rounded, including the typically “right-brained”

domain of creative content. This has changed the shape of SEO

departments and demanded collaboration with other digital

marketing departments.

Technical SEO, for its part, now encompasses all search engine

best practices and allows no room for manipulation. This specialism

never went away, but it has seen a recent renaissance as senior

marketers realize that it drives performance as well as crawler

compliance.

of those murky tactics and even (briefly) raised the discussion of

whether SEO was dead.

This question missed one key point.

As long as people are using search as a means to discover

information, SEO will continue in rude health. Those discussions

are a distant memory as we embrace modern SEO, especially its

convergence with content marketing.

The industry has gone from strength to strength and the best

strategies are now justly rewarded with increased search presence.

There are four key areas to this:

Site Content: Ensuring that content can be crawled and indexed

by all major search engines, in particular making use of log file

analysis to interpret their access patterns and structured data to

enable efficient access to content elements.

Structure: Creating a site hierarchy and URL structure that allow

both search engines and users to navigate to the most relevant

content. This should also facilitate the flow of internal link equity

through the site.

Conversion: Identifying and resolving any blockages that prevent

users from navigating through the site.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

Performance: A key development has been the evolution of

technical SEO into a performance-related specialism. This has

always been the case, but marketers of all stripes have realized

that technical SEO is about a lot more than just “housekeeping.”

Getting the three areas above in order will lead to better site

performance through search and other channels, too.

Within this context, it is worth questioning whether “SEO” is even an

adequate categorization for what we do anymore.

A New Approach: Site, Search &

Content Optimization

The term “search engine optimization” is arguably no longer fit

for purpose, as we extend our remit to include content marketing,

conversion rate optimization, and user experience.

According to research from BrightEdge, only 3 percent of

250 marketers surveyed believe SEO and content are separate

disciplines.

Our work includes:

Optimizing the site for users.

Ensuring accessibility of content for all major search engines and

social networks.

Creating content that engages the right audience across multiple

marketing channels.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

We should therefore be looking at this set of skills as site, search,

and content optimization – especially as the role of a search engine

continues to evolve beyond the 10 blue links of old.

Our responsibility is to get in front of consumers wherever they are

searching, which is an ever-changing set of contexts. This would

be a more fitting depiction of a marketing channel that plays an

increasingly pivotal role in digital and business strategy.

After all, when major technological trends develop, technical SEO

pros are often at the forefront of innovation. This looks set to be

further entrenched by recent industry developments.

Now that Accelerated Mobile Pages (AMP) and Progressive

Web Apps (PWAs) are center stage, brands must ensure that their

web presence meets the highest standards to keep pace with the

modern consumer.

Being “mobile-first” has big implications for how we engage our

audiences, but it is also a technological consideration. PWAs will

soon be coming to Google Chrome on desktop, which is a further

manifestation of the “mobile-first” approach to site experiences that

we all need to adopt.

It would be hard to argue that these fit uniquely under the remit of

‘Search Engine Optimization’, and yet it is likely SEO pros that will

lead to change within their respective organizations.

Brands need to think beyond search engines and imagine the

new ways their content could – and should – be discovered by

customers.

A different approach to SEO is required if we are to tap into the

full potential of emerging consumer trends. That approach should

expand to include site experience optimization, as well as traditional

SEO techniques.

There are plentiful new opportunities for those who adapt; a

process that can be accelerated by creating a collaborative working

environment.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

6 Thinking Hats & SEO

However we choose to label it, it should be clear that SEO has never

existed in a vacuum. From its early symbiosis with web development

to its latter-day convergence with content, SEO has always been

about collaboration.

It is therefore helpful to consider frameworks that can bring this idea

to life and bring together the specialist skills required for a modern

organic search campaign.

We typically talk only about black hat and white hat in SEO (with the

occasional mention of gray), but Edward de Bono’s Six Thinking

Hats approach can add structure to collaboration.

Each hat reflects a way of thinking and separates out the different

functions required to achieve successful outcomes. These could

be entirely different departments, different

individuals, or even different mindsets for

one person.

The objective is to improve the

collaborative process, but also to

erode the fallibility of subjectivity

by approaching every

challenge from all angles

before progressing.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

1. White Hat

A well-known term for most SEO pros, White Hat thinking in this

context depends purely on facts, statistics, and data points. This is

the most objective way of approaching a situation.

Who Should Wear This

Hat?

Data analysts and analytics

specialists are typically

naturals at adopting this

approach.

Why Is It Needed for

SEO?

Looking purely at the data

is a perfect starting point

for discussion. It keeps

everyone focused on the

objective truths of crosschannel

performance.

Data without context is

meaningless, of course, so

this approach in isolation

lacks the color needed to

understand consumers.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

2. Yellow Hat

The Yellow Hat approach brings optimism to the table, focusing

on the potential benefits a strategy may bring for brands and the

consumer.

Who Should Wear This

Hat?

Anyone can be an optimist,

so this could be a mindset

that all parties take on for a

period of time. Equally, this

could be handed to one

person as a responsibility;

the key thing is to maintain

some structure.

Why Is It Needed for

SEO?

We tend to have a lot

of ideas, so it is easy to

jettison some of them

before their full potential

has been explored. Taking

an alternative view allows

for full exploration of an

idea, even if only to retain

some of its components.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

3. Black Hat

The Black Hat is anathema to advanced SEO pros, but the

concept does have value in this particular context. We can use

this interchangeably with the “devil’s advocate” approach, where

someone purposefully points out obstacles and dangers for the

project.

Who Should Wear This Hat?

No one really, but be aware of the dangers of people offering SEO

solutions and little transparency into the how. Keep an eye out for

negative SEO attacks.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

4. Red Hat

The Red Hat approach relates to feelings and emotions, often based

on the gut reaction to an idea. This can be very beneficial for a digital

project, as we can sometimes be overly rational in our data-driven

approach.

Who Should Wear This

Hat?

It can be helpful to assign

this role to someone who

works closely with the target

audience, or who analyzes

and interprets a lot of

audience data.

Why Is It Needed for

SEO?

When fighting for vital –

and dwindling – consumer

attention, first impressions

matter. Content marketing

campaigns can depend

on getting this right, so

it’s worth listening to gut

instinct sometimes.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

5. Green Hat

The Green Hat brings creativity and spontaneity to the process,

tackling challenges from a new perspective when possible. Where

others see obstacles, this approach will see new opportunities.

Who Should Wear This

Hat?

Anyone can be creative.

However, it may be best to

assign this role to someone

who feels comfortable

sharing their ideas with

a group and is not easily

disheartened if they don’t

take off!

Why Is It Needed for

SEO?

There are best practices,

but those only take us so

far. They are a leveling

force; new ideas are what

really make the difference.

In an industry, as nascent

as ours, there are plenty

of trails yet to be explored.

The Green Hat brings that

element of innovation to a

discussion.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

6. Blue Hat

The Blue Hat organized the thinking process and takes ultimate

responsibility for bringing together the different strands into a

cohesive whole.

Who Should Wear This

Hat?

The project lead or the

person closest to the

brand’s objectives can help

keep things focused. Project

managers also have a

natural affinity for this role.

Why Is It Needed for

SEO?

SEO is an increasingly

diverse set of disciplines,

which makes this role

indispensable. To maximize

the organic search

opportunity, various

departments need to

be working in tandem

on an ongoing basis.

The Blue Hat keeps this

collaboration going.

Actual hats are optional, but may help the adoption of this approach.

Regardless, these ways of thinking have a range of

benefits across any organization:

Opportunities to integrate more digital functions into the SEO

process.

Ways to learn new skills, both by doing and by observing.

Integration of SEO best practices across more digital channels.

A central role for SEO, without reducing the importance of other

specialists.

Technical SEO Is Important Now

More Than Ever

SEO has evolved to be part of something bigger and technical skills

must be applied in a different manner.

If anything, it has expanded into a much more sophisticated and

nuanced digital channel that has outgrown the “Search Engine

Optimization” category.

The core tenets of organic search remain firmly in place, with

technical SEO given overdue prominence as a driver of web, mobile

and device performance.

SEO professionals are often at the forefront of technological

innovations and this looks unlikely to change in a world of voice

search, digital assistants, and Progressive Web Apps.

New approaches

are required if we

are to maximize

this opportunity,

however. That

begins with the

definition of what

exactly SEO entails

and extends to

the ways we lead

collaboration within our organizations.

The level of technical acumen needed for success has changed

back to the levels it once was.

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

NEW WAYS TO APPROACH TECHNICAL SEO: A NECESSITY, NOT AN OPTION

However, where and how you apply that knowledge

is key to technical success. Focus your skills on

optimizing:

Your site.

Mobile and desktop devices.

Mobile apps.

Voice search.

VR.

Agents.

Vertical search engines (it’s not just Google anymore – think

Amazon for example).

The AI revolution is begging for more help from technical SEO

professionals and data scientists to help drive it forward.

Mastering SEO fundamentals is only the bare minimum. If you really

want to win against the competition, you must go beyond the basics.

If you act now and take a slightly different viewpoint on your role,

organic search can assume a central role in both business strategy

and cross-channel digital marketing.

SEO-Friendly Hosting:

5 Things to Look for in

a Hosting Company

Chapter 1

Written By

Owner at AWCopywriting

Amelia Willson

1

As SEO professionals, we

have no shortage of things to

worry about.

I

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

There are the old standbys: backlinks, content creation, sitemaps

and robots.txt files.

And there’s new(er) stuff to get excited about as well: voice search,

featured snippets, the mobile-first index.

Amidst the noise, one factor often goes overlooked, even though it can

impact your site’s uptime and your page speed – both of which are

essential elements for maintaining positive organic performance.

I’m talking about web hosting, folks.

The web host you choose determines the overall consistency of

the site experience you offer organic visitors (and all visitors, for

that matter).

If you want to prevent server errors and page timeouts – and stop

users from bouncing back to Google – you need a solid web host

you can rely on.

Ultimately, you want a web host that supports your organic efforts,

rather than impeding them. Let’s look at five key features that define

an SEO-friendly web hosting company.

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

1

Your host’s uptime guarantee is arguably the most important factor in

whether they’re SEO-friendly.

Uptime refers to the percentage of the time your site is online and

accessible. The higher your uptime, the less likely visitors will visit

your site only to discover it’s down, sending them back to the search

engines and potentially risking your rankings in the process.

Better, more reliable hosts offer higher uptime guarantees.

For best results, choose a host with at least 99.9 percent uptime

guarantee (or higher, if you can get it). That translates to roughly 1.44

minutes of downtime a day and 8.8 hours per year. Not bad.

However, be wary of any host that claims 100 percent uptime. There’s

always going to be some

downtime. The key is

to keep it as short as

possible. That way, it

won’t affect your SEO

performance.

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

1

1. High Uptime Guarantee

While uptime refers to your site content being accessible to users,

your server location may dictate how quickly it’s accessible to them.

If you’re on a shared, VPS, or dedicated server hosting plan, your site

lives on a physical server in a data center somewhere (as opposed to

cloud hosting, where your data is housed in the cloud).

Ideally, you want that data center located as close as possible to the

majority of your site visitors. The farther away your server is, the longer

it can take for your site to load.

Server location can also look fishy to search engines, which may

affect your SEO. If you operate in one country but use a host located

halfway around the world, there may be something nefarious going

on.

It goes without saying that servers themselves should also be fast,

and that the host should further boost performance through a

Content Delivery Network (CDN).

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

1

2. Server Location

We all like options. You should enjoy them with your web hosting, too.

Beyond hosting itself, many hosting companies offer optional valueadds

that can upgrade your site.

Here are some of the SEO-friendly ones you’ll want to

see:

Automatic backups: If something ever goes wrong, you want a

site backup you can quickly restore from. See if your host offers

automatic backups for free or for an added cost.

SSL: HTTPS has been a ranking factor for years now. If you

haven’t already transitioned to a secure site, you need to get your

act together. Make sure your host supports SSL. Some even

include them for free with your hosting package.

Multiple hosting plans: As your site grows, your hosting needs

are likely to change (this is a good thing!). Eventually, your traffic

numbers may be big enough to warrant switching to your own

dedicated server. This transition will be easier (and cheaper) if

you don’t have to switch hosting providers at the same time.

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

1

3. Multiple Options

Alright, let’s say you’re actually using this list to compare hosts. By this

point, you’ve read through their hosting features, and it appears they’re

checking off all the right things.

Now it’s time to validate that the marketing claims are true. Before you

sign up with a host, take a few minutes to read their online reviews.

A caveat: The hosting space tends to attract more unhappy reviews

than most.

If a barista messes up your coffee, you’re unlikely to be bothered

enough to write a scathing review for the cafe on Yelp.

But if your site goes down, even for a moment, or even if you were at

fault (as can happen if you choose an inappropriate hosting plan for

your traffic needs), you are going to be extremely angry with your host

and tweet, post, and blog about it loudly and vociferously.

Unfortunately, that’s just the nature of the business.

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

1

4. Good Reviews

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

1

Having said that, you can still gather a lot of valuable information

from reviews. Look for hosts that appear again and again on Top

Web Hosts lists, and read the reviews to verify that the hosting plan

you’re considering is likely to give you what you need.

You won’t have trouble finding these lists. A quick Google search for

[best web hosting] delivered a slew of results from PCMag, CNET,

and more:

While you’re reading through the reviews, pay special attention to how

people talk about their support.

In the unlikely event that your site does go down, you want to be able

to fix it immediately. Most often, that will involve speaking to a support

person.

A good host will offer 24/7 support for free. Verify the operating hours

of your potential host’s support team, and see how exactly you’ll be

able to get in touch with them. Is there a phone number, live chat, or

email?

Check out their social profiles, too. Web hosts who care about helping

their customers tend to make customer support widely available on

social media, perhaps even via dedicated support Twitter accounts.

Here’s an example from Squarespace:

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

1

5. Responsive Support Team

Bonus: Easy-to-Use CMS

What Defines an SEO-Friendly Web

Host?

This one’s not exactly related to hosting, but it’s important

nonetheless. Being able to easily create outstanding content is key for

your SEO success. You know that.

So, you want a host that integrates with a CMS you’re either already

familiar with or you can easily learn. Otherwise, you’re just making

things hard on yourself!

Fortunately, most hosts today offer their own drag-and-drop content

creation tools. Many also integrate with WordPress and other popular

content management systems.

Good, reliable web hosting is one of those things that runs in the

background without you ever having to think about it. That, in

essence, is an SEO-friendly web host.

SEO-FRIENDLY HOSTING: 5 THINGS TO LOOK FOR IN A HOSTING COMPANY

1

The Ultimate Guide for

an SEO-Friendly URL

Structure

Chapter 2

Written By

Founder , Candid Digital

Clark Boyd

2

First impressions count.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

And when it comes to your website, your URLs are often the first thing

Google and customers will see.

URLs are also the building blocks of an effective site hierarchy, passing

equity through your domain and directing users to their desired

destinations.

They can be tricky to correct if you don’t plan ahead, as you can end

up with endless redirect loops. Neither Google nor your site visitors will

appreciate those.

So they are worth getting right. But getting URL structure right involves

a complex blend of usability and accessibility factors, along with some

good old-fashioned SEO.

Although there is no one-size-fits-all approach, there are some rules we

can all follow to get the most out of our URLs and set our sites up for

future SEO success.

Every time you launch a page on your domain, it should have a

purpose. Whether transactional, informational, or administrative, its

reason for existence should be clear at the outset.

You’ll want this page to be discovered by the right people (and

crawlers), so you will incorporate some keyword research and include

the relevant terms. The most descriptive of these — the term that gets

to the nub of what this page is about — should be included in the

URL, close to the root domain.

We’ll deal with multiple pages that broadly tackle the same topic later,

but for now, let’s assume the simple example of a page that clearly

handles one topic. Let’s go for whiskey.

Generic example:

https://example.com/topic

Whiskey-based example:

https://example.com/whiskey

Even this isn’t quite as simple as it seems, though.

Should we use

“whiskey” or

“whisky” as our

standard spelling?

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

1. Use Your Keywords

The search volume is with “whiskey” and our site is based in the U.S.,

so let’s run with that.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

Both are valid, with the former being an Irish spelling and the latter

Scottish. The Irish spelling has been adopted in the U.S., but we’ll

need more proof before proceeding with that as our chosen variation.

The Moz Keyword Explorer is great for this sort of predicament,

as it groups keywords together to give an estimate of the search

volume for particular topics. In this era of vague keyword-level search

volumes, it provides a nice solution.

Perhaps the biggest challenge we all face when defining a sitewide

URL hierarchy is ensuring that it will still fit our purpose for years to

come.

It is for this reason that some websites end up as a patchwork quilt

of sub-domains and conflicting paths to arrive at similar products.

This is poor from a user’s perspective, but it also sends confusing

signals to Google about how you categorize your product offering.

An example of this would be:

https://example.com/whiskey/irish-whiskey/jameson

https://example.com/bushmills

The first URL flows

logically from domain

to category to subcategory

to product.

The second

URL goes from

domain to product.

Hierarchically, both

products should sit at

the same level in the site

and the Jameson example is

better for SEO and users.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

2. Build a Sound Structure for the

Future

We encounter this a lot, though. Why?

It can be a simple lack of communication, with a product team

launching a new item straight onto the site without consulting other

parties. It can also be down to a failure of future planning.

Either way, it’s essential to lay out your structure in advance. Work

together with different teams to understand the future direction

of the business, then add your SEO knowledge to shape the site

architecture. It will rarely be perfect, but the more you plan, the fewer

errors you will have to undo down the line.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

As a rule of thumb, make sure a user can understand what your

page is about by looking at the URL. That means you don’t need to

include every single preposition or conjunction.

Words like “and” or “the” are just distractions and can be stripped

out of the URL altogether. Just as users can understand what a

topic is about without these short words, Google will derive all the

meaning it requires too.

You should also avoid keyword repetition within URLs. Adding the

same keyword multiple times in the hope of increasing your ranking

chances will only lead to a spammy URL structure.

An example of this unnecessary repetition would be:

https://domain.com/whiskey/irish-whiskey/jameson-irish-whiskey/

jameson-irish-whiskey-history

The first two uses of the main keyword make sense, but the third

and fourth are overkill.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

3. Avoid Superfluous Words &

Characters

A few additional points to bear in mind on this topic:

Case Sensitivity: It is surprisingly common to find multiple

versions of the same URL, with one all in lower case and the

others using occasional capital letters. Use canonical tags to

mark the lower-case URL as the preferred version or, if possible,

use permanent redirects.

Hashes: These can be useful to send users to a specific section

of a page, but restrict their use in other circumstances if possible.

If the content users are sent to after the # symbol is unique, make

it available via a simple URL instead.

Word Delimiters: Stick with hyphens to separate words within

your URL strings. Underscores will serve to join two words

together, so be wary of using these.

URL Length: After 512 pixels, Google will truncate your URL in

search results pages. A good rule of thumb is to keep y0ur URLs

as short as you can, without losing their general meaning.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

This one can be harder than it sounds, depending on the content

management system you use.

Some e-commerce platforms will automatically spit out character

strings that leave you with URLs like: https://domain.com/

cat/?cid=7078.

These are a bit unsightly and they also go against the rules we’ve

been outlining above. We want static URLs that include a logical

folder structure and descriptive keywords.

Although search engines have no problem crawling or indexing

either variant, for SEO-based reasons it’s better to use static URLs

rather than dynamic ones. The thing is, static URLs contain your

keywords and are more user-friendly since one can figure out what

the page is about just by looking at the static URL’s name.

So how do we get around this? You can use rewrite rules if your web

server runs Apache, and some tools like this one from Generate It

are helpful. There are different fixes for different platforms (some more

complex than others).

Some web developers

make use of relative URLs,

too. The problem with

relative URLs for SEO is

that they are dependent on

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

4. Minimize Dynamic URL Strings

the context in which they occur. Once the context changes, the URL

may not work. For SEO, it’s better to use absolute URLs instead of

relative ones, since the former are what search engines prefer.

Now, sometimes different parameters can be added to the URL

for analytics tracking or other reasons (such as sid, utm, etc.) To

make sure that these parameters don’t make the number of URLs

with duplicate content grow over the top, you can do either of the

following:

Ask Google to disregard certain URL parameters in Google Search

Console in Configuration > URL Parameters.

See if your content management system allows you to solidify URLs

with additional parameters with their shorter counterparts.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

As a rule, there are two major versions of your domain indexed in

search engines: the www and the non-www version of it. We can

add to this the complexity of having a secure (https) and non-secure

(HTTP) version too, with Google giving preference to the former.

Most SEOs use the 301 redirect to point one version of their site to

the other (or vice versa).

This tells search engines that a particular URL has moved

permanently to another destination.

Alternatively (for instance, when you can’t do a redirect), you

can specify your preferred version in Google Search Console in

Configuration > Settings > Preferred Domain. However, this has

certain drawbacks:

This takes care of Google only.

This option is restricted to root domains only.

If you have an example.wordpress.com site,

this method is not for you.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

5. Consolidate the Different

Versions of Your Site

But why worry about the www vs non-www issue in the first place?

The thing is, some of your backlinks may be pointing to your www

version, while some could be going to the non-www version.

To ensure all versions’ SEO value is consolidated, it’s better to

explicitly establish this link between them. You can do this via the

301 redirect, in Google Search Console, or by using a canonical tag,

the latter of which we will look at in more detail below. THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

So, canonical tags. These are a very helpful piece of code when you

have multiple versions of what is essentially the same page. By adding

a canonical tag, you can tell Google which one is your preferred

version.

Note: The canonical tag should be applied only with the purpose of

helping search engines decide on your canonical URL. For redirection

of site pages, use redirects. And, for paginated content, it makes

sense to employ rel=”next” and rel=”prev” tags in most cases.

Canonical tags are useful for just about any website, but they are

particularly powerful for online retailers.

For example, on Macy’s website, I can go to the Quilts & Bedspreads

page directly by using the URL (https://www.macys.com/shop/

bed-bath/quilts-bedspreads), or I can take different routes from the

homepage:

I can go to Homepage >> Bed& Bath >> Quilts & Bedspreads. The

following URL with my path recorded is generated:

https://www.macys.com/shop/bed-bath/quiltsbedspreads?

id=22748&edge=hybrid

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

6. Make Correct Use of Canonical

Tags

As you see, for each of these URLs, a canonical URL is specified,

which is the cleanest version of all the URLs in the group:https://

www.macys.com/shop/bed-bath/quilts-bedspreads?id=22748

What this does is, it funnels down the SEO value each of these three

URLs might have to one single URL that should be displayed in the

search results (the canonical URL). Normally search engines do a

pretty good job identifying canonical URLs themselves, but, as Susan

Moskwa once wrote at Google Webmaster Central:

“If we aren’t able to detect all the duplicates of a particular page, we

won’t be able to consolidate all of their properties. This may dilute

the strength of that content’s ranking signals by splitting them across

multiple URLs.”

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

Or I can go to Homepage >> For the Home >> Bed & Bath >>

Bedding >> Quilts & Bedspreads. The following URL is generated:

https://www.macys.com/shop/bed-bath/quiltsbedspreads?

id=22748&cm_sp=us_hdr-_-bed-%26-bath-_-22748_

quilts-%26-bedspreads_COL1

Now, all three URLs lead to the same content. And if you look into

the code of each page, you’ll see the following tag in the head

element:

In Google’s own Search Quality Evaluators Guidelines (a mustread

document for all SEOs!), there are clear references to both main

content and supplementary content.

Main content will be your lead page in each section that really sets

out what your category is all about. It will set out your stall as a

relevant source for a topic. Supplementary content provides, as the

name suggests, additional information that helps users navigate the

topic and reach informed decisions.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

7. Incorporate Topical Authority

URL structure is an essential component of getting this right.

So, let’s go back to our whiskey example to see how we might

tackle this. Our site is e-commerce focused and we want to sell the

product, of course. However, going for the jugular and only pushing

out product pages is tantamount to SEO tunnel vision.

Our initial research from Moz Keyword Explorer is a great resource

as we make these plans. Below, I have exported the keyword list

and reduced it to the highest-volume topics. From here, we can

start to decide what might qualify as a topic for a main content or

supplementary content page.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

This is a simplified example and just a first step, of course.

However, it is worth noting that this approach goes further than just

category > sub-category > product. By thinking in terms of main

content and supplementary content, a product is just as likely to

qualify as main content as a category is. The question is more about

which topics consumers want us to elaborate on to help them make

choices.

From here, we can dig into some of these topics and start to flesh

out what each hub might look like.

Some clear opportunities already stand out to create content and

rank via rich snippets. People want to know how whiskey is made,

what different varieties exist, and of course, whether it’s spelled

‘whiskey’ or ‘whisky’. This could be the beginning of a business

case to create a whiskey tasting guide or a ‘history of whiskey’

content hub on the site.

Combined with ranking difficulty metrics, business priorities, and

content production capabilities, this approach will soon take shape

as a site hierarchy and opportunity analysis.

For our whiskey example, it might start to comprise the following

structure:

https://domain.com/whiskey/whiskey-tasting-guide

https://domain.com/whiskey/whiskey-tasting-guide/how-to-tastewhiskey

https://domain.com/whiskey/whiskey-tasting-guide/how-iswhiskey-

made

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

https://domain.com/whiskey/whiskey-tasting-guide/barley-whiskey

Again, there are decisions to make.

In the last URL, one could argue that the tasting guide page for

barley whiskey should sit under the barley whiskey sub-category

page in the site hierarchy. Barley whiskey has been earmarked as

‘main content’ in my spreadsheet, after all. The choice here comes

down to where we want to consolidate value; dispersing that value

would reduce our chances of ranking for any ‘tasting guide’ terms.

These are exactly the kinds of decisions that can lead to a confused

structure if a consistent logic is not followed.

All of this will contribute to your topical authority and increase site

visibility.

This type of content often already exists on site, too. I am not

claiming anything revolutionary by saying a website should have

lots of useful information, after all. However, the structure of this

content and how entities are semantically linked to each other

makes the difference between success and failure.

This can be used as a ‘quick win’ tactic and it tends to be received

well by all parties. Updating and moving existing content will always

be an easier sell than asking for an all-new content hub.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

Once you’ve ticked off all of the above, you’ll want to make sure

search engines know what’s going on with your website. That’s where

sitemaps come in handy — particularly XML sitemaps.

An XML Sitemap is not to be confused with the HTML sitemap. The

former is for the search engines, while the latter is mostly designed

for human users (although it has other uses t00).

So what is an XML Sitemap? In plain words, it’s a list of your site’s

URLs that you submit to the search engines.

This serves two purposes:

1. This helps search engines find your site’s pages more easily.

2. Search engines can use the sitemap as a reference when

choosing canonical URLs on your site.

Picking a preferred

(canonical) URL becomes

necessary when search

engines see duplicate

pages on your site, as we

saw above.

So, as they don’t want any

duplicates in the search

results, search engines use

a special algorithm to identify

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

8. Create an XML Sitemap

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

duplicate pages and pick just one URL to represent the group in the

search results. Other web pages just get filtered out.

Now, back to sitemaps. One of the criteria search engines may use

to pick a canonical URL for the group of web pages is whether this

URL is mentioned in the website’s sitemap.

So, what web pages should be included in your sitemap? For

purely SEO reasons, it’s recommended to include only the web

pages you’d like to show up in search. You should include a more

comprehensive account of your site’s URLs within the HTML

sitemap.

Summary

An SEO-friendly URL structure is the following things:

Easy to read: Users and search engines should be able to

understand what is on each page just by looking at the URL.

Keyword-rich: Keywords still matter and your target queries

should be within URLs. Just be wary of overkill; extending URLs just

to include more keywords is a bad idea.

Consistent: There are multiple ways to create an SEO-friendly

URL structure on any site. It’s essential that, whatever logic you

choose to follow, it is applied consistently across the site.

Static: Dynamic parameters are rarely an SEO’s best friend, but

they are quite common. Where possible, find a solution that allows

your site to render static URLs instead.

Future-proof: Think ahead when planning your site structure. You

should minimize the number of redirects on your domain, and it’s

easier to do this if you don’t require wholesale changes to URLs.

Comprehensive: Use the concepts of main content and

supplementary content to ensure you have adequate coverage for

all relevant topics. This will maximize your site’s visibility.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

Supported by data: It normally requires buy-in from a lot of

stakeholders to launch or update a particular site structure.

Numbers talk, so make use of search and analytics data to support

your case.

Submitted to search engines: Finally, create an XML sitemap

containing all of the URLs that you want to rank via SEO and

submit it to search engines. That will ensure all your hard work

gets the reward it deserves.

THE ULTIMATE GUIDE FOR AN SEO-FRIENDLY URL STRUCTURE

2

How to Use XML

Sitemaps to Boost SEO

Chapter 3

Written By

International Digital Director, Ringier

Jes Scholz

3

As the web evolves, so too

does Google and SEO.

HOW TO USE XML SITEMAPS TO BOOST SEO

This means what is considered best practice is often in flux. What may

have been good counsel yesterday, is not so today.

This is especially true for sitemaps, which are almost as old as SEO itself.

The problem is, when every man and their dog has posted answers

in forums, published recommendations on blogs and amplified

opinions with social media, it takes time to sort valuable advice from

misinformation.

3

So while most of us share a general understanding that

submitting a sitemap to Google Search Console is important,

you may not know the intricacies of how to implement them in a way

that drives SEO key performance indicators (KPIs).

Let’s clear up the confusion around best practices for sitemaps today.

In this article we cover:

What is an XML sitemap

XML sitemap format

Types of sitemaps

XML sitemap indexation optimization

XML sitemap best practice checklist

HOW TO USE XML SITEMAPS TO BOOST SEO

3

What Is an XML Sitemap

In simple terms, an XML sitemap is a list of your website’s URLs.

It acts as a roadmap to tell search engines what content is available

and how to reach it.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

In the example above, a search engine will find all nine pages in a

sitemap with one visit to the XML sitemap file.

On the website, it will have to jump through five internal links to find

page 9.

This ability of an XML sitemap to assist crawlers in faster

indexation is especially important for websites that:

Have thousands of pages and/or a deep website architecture.

Frequently add new pages.

Frequently change content of existing pages.

Suffer from weak internal linking and orphan pages.

Lack a strong external link profile.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Side note: Submitting a sitemap with noindex URLs can also

speed up deindexation. This can be more efficient than removing

URLs in Google Search Console if you have many to be deindexed.

But use this with care and be sure you only add such URLs

temporarily to your sitemaps.

Key Takeaway

Even though search engines can technically find your URLs

without it, by including pages in an XML sitemap you’re indicating

that you consider them to be quality landing pages.

While there is no guarantee that an XML sitemap will get your

pages crawled, let alone indexed or ranked, submitting one

certainly increases your chances.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

XML Sitemap Format

A one-page site using all available tags would have this XML sitemap:

But how should an SEO use each of these tags? Is all the metadata

valuable?

Loc (a.k.a. Location) Tag

This compulsory tag contain the absolute, canonical version of the

URL location.

It should accurately reflect your site protocol (http or https) and if you

have chosen to include or exclude www.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

For international websites, this is also where you can implement

your hreflang handling.

By using the xhtml:link attribute to indicate the language and region

variants for each URL, you reduce page load time, which the other

implementations of link elements in the <head> or HTTP headers

can’t offer

Yoast has an epic post on hreflang for those wanting to learn more.

Lastmod (a.k.a. Last Modified) Tag

An optional but highly recommended tag used to communicate the

file’s last modified date and time.

John Mueller acknowledged Google does use the lastmod metadata

to understand when the page last changed and if it should be

crawled. Contradicting advice from Illyes in 2015.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

The last modified time is especially critical for content sites as it

assists Google to understand that you are the original publisher.

It’s also powerful to communicate freshness, be sure to update

modification date only when you have made meaningful changes.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Trying to trick search engines that your content is fresh, when

it’s not, may result in a Google penalty.

Changefreq (a.k.a. Change Frequency) Tag

Once upon a time, this optional tag hinted how frequently content on

the URL was expected to change to search engines.

But Mueller has stated that “change frequency doesn’t really play

that much of a role with sitemaps” and that “it is much better to just

specify the time stamp directly”.

Priority Tag

This optional tag that ostensibly tells search engines how important

a page is relative to your other URLs on a scale between 0.0 to 1.0.

At best, it was only ever a hint to search engines and both Mueller

and Illyes have clearly stated they ignore it.

Key Takeaway

Your website needs an XML sitemap, but not necessarily the

priority and change frequency metadata.

Use the lastmod tags accurately and focus your attention on

ensuring you have the right URLs submitted.

Types of Sitemaps

There are many different types of sitemaps. Let’s look at the ones you

actually need.

XML Sitemap Index

XML sitemaps have a couple of limitations:

A maximum of 50,000 URLs.

An uncompressed file size limit of 50MB.

Sitemaps can be compressed using gzip (the file name would

become something similar to sitemap.xml.gz) to save bandwidth for

your server. But once unzipped, the sitemap still can’t exceed either

limit.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Whenever you exceed either limit, you will need to split your URLs

across multiple XML sitemaps.

Those sitemaps can then be combined into a single XML sitemap

index file, often named sitemap-index.xml. Essentially, a sitemap

for sitemaps.

For exceptionally large websites who want to take a more granular

approach, you can also create multiple sitemap index files.

For example:

sitemap-index-articles.xml

sitemap-index-products.xml

sitemap-index-categories.xml

But be aware that you cannot nest sitemap index files.

For search engines to easily find every one of your

sitemap files at once, you will want to:

Submit your sitemap index(s) to Google Search Console and Bing

Webmaster Tools.

Specify your sitemap index URL(s) in your robots.txt file. Pointing

search engines directly to your sitemap as you welcome them to

crawl.

You can also submit sitemaps by pinging them to Google.

But beware:

Google no longer pays attention to hreflang entries in “unverified

sitemaps”, which Tom Anthony believes to mean those submitted via

the ping URL.

XML Image Sitemap

Image sitemaps were designed to improve the indexation of image

content.

In modern-day SEO, however, images are embedded within page

content, so will be crawled along with the page URL.

Moreover, it’s best practice to utilize JSON-LD schema.org/

ImageObject markup to call out image properties to search engines

as it provides more attributes than an image XML sitemap.

Because of this, an XML image sitemap is unnecessary for most

websites. Including an image sitemap would only waste crawl budget.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

The exception to this is if images help drive your business, such

as a stock photo website or ecommerce site gaining product page

sessions from Google Image search.

Know that images don’t have to to be on the same domain as your

website to be submitted in a sitemap. You can use a CDN as long as

it’s verified in Search Console.

XML Video Sitemap

Similar to images, if videos are critical to your business, submit an

XML video sitemap.If not, a video sitemap is unnecessary.

Save your crawl budget for the page the video is embedded into,

ensuring you markup all videos with JSON-LD as a schema.org/

VideoObject.

Google News Sitemap

Only sites registered with Google News should use this sitemap.

If you are, include articles published in the last two days, up to a limit

of 1,000 URLs per sitemap, and update with fresh articles as soon as

they’re published.

Contrary to some online advice, Google News sitemaps don’t

support image URL.

Google recommends using schema.org image or og:image to

specify your article thumbnail for Google News.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Mobile Sitemap

This is not needed for most websites.

Why? Because Mueller confirmed mobile sitemaps are for feature

phone pages only. Not for smartphone-compatibility.

So unless you have unique URLs specifically designed for featured

phones, a mobile sitemap will be of no benefit.

HTML Sitemap

XML sitemaps take care of search engine needs. HTML sitemaps

were designed to assist human users to find content.

The question becomes, if you have a good user experience and well

crafted internal links, do you need a HTML sitemap?

Check the page views of your HTML sitemap in Google Analytics.

Chances are, it’s very low. If not, it’s a good indication that you need

to improve your website navigation.

HTML sitemaps are generally linked in website footers. Taking link

equity from every single page of your website.

Ask yourself. Is that the best use of that link equity? Or are you

including HTML sitemap as a nod to legacy website best practices?

If few humans use it. And search engines don’t need it as you have

strong internal linking and an XML sitemap. Does that HTML sitemap

have a reason to exist? I would argue no.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Dynamic XML Sitemap

Static sitemaps are simple to create using a tool such as Screaming

Frog.

The problem is, as soon as you create or remove a page, your

sitemap is outdated. If you modify the content of a page, the sitemap

won’t automatically update the lastmod tag.

So unless you love manually creating and uploading sitemaps for

every single change, it’s best to avoid static sitemaps.

Dynamic XML sitemaps, on the other hand, are automatically update

by your server to reflect relevant website changes as they occur.

To create a dynamic XML sitemap:

Ask you developer to code a custom script, being sure to provide

clear specifications

Use a dynamic sitemap generator tool

Install a plugin for your CMS, for example the Yoast SEO plugin for

Wordpress

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Key Takeaway

Dynamic XML sitemaps and a sitemap index are modern best

practice. Mobile and HTML sitemaps are not.

Use image, video and Google News sitemaps only if improved

indexation of these content types drive your KPIs.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

XML Sitemap Indexation

Optimization

Now for the fun part. How do you use XML sitemaps to drive SEO

KPIs.

Only Include SEO Relevant Pages in XML Sitemaps

An XML sitemap is a list of pages you recommend to be crawled,

which isn’t necessarily every page of your website.

A search spider arrives at your website with an “allowance” for how

many pages it will crawl.

The XML sitemap indicates you consider the included URLs to

be more important than those that aren’t blocked but aren’t in the

sitemap.

You are using it to tell search engines “I’d really appreciate it if you’d

focus on these URLs in particular”.

Essentially, it helps you use crawl budget effectively.

By including only SEO relevant pages, you help search engines crawl

your site more intelligently in order to reap the benefits of better

indexation.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

You should exclude:

Non-canonical pages.

Duplicate pages.

Paginated pages.

Parameter or session ID based URLs.

Site search result pages.

Reply to comment URLs.

Share via email URLs.

URLs created by filtering that are unnecessary for SEO.

Archive pages.

Any redirections (3xx), missing pages (4xx) or server error pages

(5xx).

Pages blocked by robots.txt.

Pages with noindex.

Resource pages accessible by a lead gen form (e.g. white paper

PDFs).

Utility pages that are useful to users, but not intended to be

landing pages (login page, contact us, privacy policy, account

pages, etc.).

HOW TO USE XML SITEMAPS TO BOOST SEO

3

I want to share an example from Michael Cottam about prioritising

pages:

Say your website has 1,000 pages. 475 of those 1,000 pages are SEO

relevant content. You highlight those 475 pages in an XML sitemap,

essentially asking Google to deprioritize indexing the remainder.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Now, let’s say Google crawls those 475 pages, and algorithmically

decides that 175 are “A” grade, 200 are “B+”, and 100 “B” or “B-”.

That’s a strong average grade, and probably indicates a quality

website to which to send users.

Contrast that against submitting all 1,000 pages via the XML

sitemap. Now, Google looks at the 1,000 pages you say are SEO

relevant content, and sees over 50 percent are “D” or “F” pages.

Your average grade isn’t looking so good anymore and that may

harm your organic sessions.

But remember, Google is going to use your XML sitemap only as a

clue to what’s important on your site.

Just because it’s not in your XML sitemap doesn’t necessarily

mean that Google won’t index those pages.

When it comes to SEO, overall site quality is a key factor.

To assess the quality of your site, turn to the sitemap related

reporting in Google Search Console (GSC).

Key Takeaway

Manage crawl budget by limiting XML sitemap URLs only to SEO

relevant pages and invest time to reduce the number of low quality

pages on your website.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Fully Leverage Sitemap Reporting

The sitemaps section in the new Google Search Console is not as

data rich as what was previously offered.

It’s primary use now is to confirm your sitemap index has been

successfully submitted.

If you have chosen to use descriptive naming conventions, rather

than numeric, you can also get a feel for the number of different types

of SEO pages that have been “discovered” - aka all URLs found by

Google via sitemaps as well as other methods such as following links.

In the new GSC, the more valuable area for SEOs in regard to

sitemaps is the Index Coverage report.

The report will default to “All known pages”. Here you

can:

Address any “Error” or “Valid with warnings” issues. These often

stem from conflicting robots directives. One solved, be sure to

validate your fix via the Coverage report.

Look at indexation trends. Most sites are continually adding

valuable content, so “Valid” pages (aka those indexed by Google)

should steadily increase. Understand the cause of any dramatic

changes.

Select “Valid” and look in details for the type “Indexed, not

submitted in sitemap”. These are pages where you and Google

disagree on their value. For example, you may not have submitted

your privacy policy URL, but Google has indexed the page. In

such cases, there’s no actions to be taken. What you need to

be looking out for are indexed URLs which stem from poor

pagination handling, poor parameter handling, duplicate

content or pages being accidently left out of sitemaps.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Afterwards, limit the report to the SEO relevant URLs you have

included in your sitemap by changing the drop down to “All

submitted pages”. Then check the details of all “Excluded” pages.

Reasons for exclusion of sitemap URLs can be put into

four action groups:

1. Quick wins: For duplicate content, canoncials, robots

directives, 40X HTTP status codes, redirects or legalities

exclusions put in place the appropriate fix.

2. Investigate page: For both “Submitted URL dropped” and

“Crawl anomaly” exclusions investigate further by using the Fetch

as Google tool.

3.Improve page: For “Crawled - currently not indexed” pages,

review the page (or page type as generally it will be many URLs

of a similar breed) content and internal links. Chances are, it’s

suffering from thin content, unoriginal content or is orphaned.

4. Improve domain: For “Discovered - currently not indexed”

pages, Google notes the typical reason for exclusion as they “tried

to crawl the URL but the site was overloaded”. Don’t be fooled. It’s

more likely that Google decided “it’s not worth the effort” to crawl

due to poor internal linking or low content quality seen from the

domain. If you see a larger number of these exclusions, review

the SEO value of the page (or page types) you have submitted via

sitemaps, focus on optimising crawl budget as well as review your

information architecture, including parameters, from both an link

and content perspective.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Whatever your plan of action, be sure to note down benchmark KPIs.

The most useful metric to assess the impact of sitemap optimisation

efforts is the “All submitted pages” indexation rate - calculated by

taking the percentage of valid pages out of total discovered URLs.

Work to get this above 80 percent.

Why not to 100 percent? Because if you have focussed all your

energy on ensuring every SEO relevant URL you currently have is

indexed, you likely missed opportunities to expand your content

coverage.

Key Takeaway

In addition to identifying warnings and errors, you can use the

Index Coverage report as an XML sitemap sleuthing tool to isolate

indexation problems.

Note: If you are a larger website who has chosen to break their

site down into multiple sitemap indexes, you will be able to filter by

those indexes.

This will not only allow you to:

1. See the overview chart on a more granular level.

2. See a larger number of relevant examples when investigating a

type of exclusion.

3. Tackle indexation rate optimisation section by section.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

XML Sitemap Best Practice

Checklist

Do invest time to:

Compress sitemap files using gzip

Use a sitemap index file

Use image, video and Google news sitemaps only if indexation

drives your KPIs

Dynamically generate XML sitemaps

Ensure URLs are included only in a single sitemap

Reference sitemap index URL(s) in robots.txt

Submit sitemap index to both Google Search Console and Bing

Webmaster Tools

Include only SEO relevant pages in XML sitemaps

Fix all errors and warnings

Analyse trends and types of valid pages

Calculate submitted pages indexation rate

Address causes of exclusion for submitted pages

Now, go check your own sitemap and make sure you’re doing it

right.

HOW TO USE XML SITEMAPS TO BOOST SEO

3

Best Practices for

Setting Up Meta

Robots Tags & Robots.

txt

Chapter 4

Written By

Founder, Opporty

Sergey Grybniak

4

First-rate website optimization is

fundamental to success in search, but

forgetting about the technical part of

SEO can be a serious mistake. BEST PRACTICES FOR SETTING UP META ROBOTS TAGS &

ROBOTS.TXT

Experienced digital marketers and SEO professionals understand the

importance of proper search engine indexing. For that reason, they do

their best to help Google crawl and index their sites properly, investing

time and resources in on-page and off-page optimization.

Content, links, tags, meta descriptions, image optimization, and website

structure are essential for SEO, but if you have never heard about robots.

txt, meta robots tags, XML sitemaps, microformats, and X-Robot tags, you

could be in trouble.

But do not panic.

In this chapter, I will explain how to use and set up robots.txt and meta

robots tags. I will provide several practical examples as well.

Let’s start!

4

What Is Robots.txt?

Robots.txt is a text file used to instruct search engine bots (also

known as crawlers, robots, or spiders) how to crawl and index

website pages.

Ideally, a robots.txt file is placed in the top-level directory of your

website so that robots can access its instructions right away.

Why Is Robots.txt Important?

Correct robots.txt operation ensures that search engine bots are

routed to required pages, disallowing content duplicates that lead to

a fall in position. For that reason, you should make sure your site has

a thoughtfully created robot.txt file.

If a robots.txt file is set up incorrectly, it can cause multiple indexing

mistakes. So, every time you start a new SEO campaign, check your

robots.txt file with Google’s robots texting tool.

Do not forget: If everything is correctly set up, a robots.txt file will

speed up the indexing process.

Robots.txt on the Web

Yet, do not forget that any robots.txt file is publicly available on

the web. To access a robots.txt file, simply type: www.websiteexample.

com/robots.txt.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

This availability means that:

You cannot secure or hide any data within it.

Bad robots and malicious crawlers can take advantage of a robots.

txt file, using it as a detailed map to navigate your most valuable web

pages.

Also, keep in mind that robots.txt commands are actually directives.

This means that search bots can crawl and index your site, even if

you instruct them not to.

The good news is that most search engines (like Google, Bing,

Yahoo, and Yandex) honor robots.txt directives.

Robots.txt files definitely have drawbacks. Nonetheless, I strongly

recommend you make them an integral part of every SEO campaign.

Google recognizes and honors robots.txt directives and, in most

cases, having Google under your belt is more than enough.

Robots.txt Basics

The robots.txt file should:

Contain the usual text in the UTF-8 encoding, which consists of

records (lines), divided by symbols.

Be situated at the root of the website host to which it applies.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Be unique.

Contain not more than 1,024 rules.

Be under 500KB.

Google bots find all the content available for indexing if:

There is no robots.txt file.

A robots.txt file isn’t shown in the text format.

They do not receive the 200 OK response.

Note:

You can, but are not allowed to, mention the byte order mark (BOM)

at the beginning of the robots.txt file, as it will be ignored by bots.

The standard recommends the use of a newline before each Useragent

directive.

If your encoding contains symbols beyond the UTF-8, bots may

analyze the file incorrectly. They will execute the valid entry only,

ignoring the rest of your content without notifying you about the

mistake.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Robots.txt Structure

Robots.txt File consists of:

One or several User-agent directives, meant for robots of various

search engines.

Disallow and Allow directives that allow or restrict indexing.

Sitemap directives.

Disallow directives forbid indexing, Allow directives allow

indexing.

Each record consists of the directory field (allow, disallow, host or

user-agent), two-spot and a value. Empty spaces are not required,

but recommended for better readability. You can place comments

anywhere in the file and mark them with the # symbol.

“#” is the symbol meant for comment descriptions.

Google bots do not count everything mentioned between

the # symbol and the next newline.

The general format is: <field>:<value><#comment (optional)>.

Empty spaces at the beginning and the end will be ignored.

Letter case for <field> element does not matter.

Letter case might be important for the <value> element, depending

on the <field> element.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

What to Hide with Robots.txt

Obviously, you do not want to show search engines your private

technical page, customers’ personal data, and duplicate content.

Robots.txt files can be used to exclude certain directories,

categories, and pages from search. To that end, use the “disallow”

directive.

Here are some pages you should hide using a robots.txt

file:

Pages with duplicate content

Pagination pages

On-site search pages

Dynamic product and service pages

Account pages

Admin pages

Shopping cart

Chats

Thank-you pages

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

How to Use Robots.txt

Robots.txt files are pretty flexible and can be used in many ways.

Their main benefit, however, is that they enable SEO experts to

“allow” or “disallow” multiple pages at once without having to

access the code of page by page.

Here is an example of how I instruct Googlebot to avoid crawling

and indexing all pages related to user accounts, cart, and multiple

dynamic pages that are generated when users look for products in

the search bar or sort them by price, and so on.

For example, you can block all search crawlers from content, like

this:

User-agent: *

Disallow: /

Or hide your site’s directory structure and specific categories, like

this:

User-agent: *

Disallow: /no-index/

It’s also useful for excluding multiple pages from search.

Just parse URLs you want to hide from search crawlers. Then, add

the “disallow” command in your robots.txt, list the URLs and, voila! –

the pages are no longer visible to Google.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

More important, though, is that a robots.txt file allows you to

prioritize certain pages, categories, and even bits of CSS and JS

code. Have a look at the example below:

Here, we have disallowed WordPress pages and specific

categories, but wp-content files, JS plugins, CSS styles, and blog

are allowed. This approach guarantees that spiders crawl and index

useful code and categories, firsthand.

One more important thing: A robots.txt file is one of the possible

locations for your sitemap.xml file. It should be placed after Useragent,

Disallow, Allow, and Host commands. Like this:

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Note: You can also add your robots.txt file manually to Google

Search Console and, in case you target Bing, Bing Webmaster

Tools.

Even though robots.txt structure and settings are pretty

straightforward, a properly set up file can either make or break

your SEO campaign.

Be careful with settings: You can easily “disallow” your entire site

by mistake and then wait for traffic and customers to no avail.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Typical Robots.txt Mistakes

1. The File Name Contains Upper Case

The only possible file name is robots.txt, nor Robots.txt or ROBOTS.

TXT.

2. Using Robot.Txt Instead of Robots.txt

Once again, the file must be called robots.txt.

3. Incorrectly Formatted Instructions

For example: Disallow: Googlebot

The only correct option is:

User-agent: Googlebot

Disallow: /

4. Mentioning Several Catalogs in Single ‘Disallow’

Instructions

Do not place all the catalogs you want to hide in one ‘disallow’ line,

like this:

Disallow: /css/ /cgi-bin/ /images/

The only correct option is:

Disallow: /css/

Disallow: /cgi-bin/

Disallow: /images/

5. Empty Line in ‘User-Agent’

Wrong option:

User-agent:

Disallow:

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

The only correct option is:

User-agent: *

Disallow:

6. Using Upper Case in the File

This is wrong and is treated as a bad style:

USER-AGENT: GOOGLEBOT

DISALLOW:

7. Mirror Websites & URL in the Host Directive

To state which website is the main one and which is the mirror

(replica), specialists use 301 redirect for Google and ‘host’ directive

for Yandex.

Although the links to http://www.site.com, http://site.com, https://

www.site.com, and https://site.com seem identical for humans,

search engines treat them as four different websites.

Be careful when mentioning ‘host’ directives, so that search engines

understand you correctly:

Wrong

User-agent: Googlebot

Disallow: /cgi-bin

Host: http://www.site.com/

Correct

User-agent: Googlebot

Disallow: /cgi-bin

Host: www.site.com

If your site has https, the correct option is

User-agent: Googlebot

Disallow: /cgi-bin

Host: https:// www.site.com

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

8. Listing All the Files Within the Directory

9. Absence of Disallow Instructions

The disallow instructions are required so that search engines bots

understand your intents.

10. Redirect 404

Even if you are not going to create and fill out robots.txt. file for your

website, search engines may still try to reach the file. Consider

creating at least an empty robots.txt. to avoid disappointing search

engines with 404 Not Found pages.

Wrong

User-agent: *

Disallow: /AL/Alabama.html

Disallow: /AL/AR.html

Disallow: /Az/AZ.html

Disallow: /Az/bali.html

Disallow: /Az/bed-breakfast.html

Wrong

User-agent: *

Disallow: /AL/Alabama.html

Disallow: /AL/AR.html

Disallow: /Az/AZ.html

Disallow: /Az/bali.html

Disallow: /Az/bed-breakfast.html

Correct

Just hide the entire directory:

User-agent: *

Disallow: /AL/

Disallow: /Az/

Correct

Just hide the entire directory:

User-agent: *

Disallow: /AL/

Disallow: /Az/

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

11. Using Additional Directives in the * Section

If you have additional directives, such as ‘host’ for example, you

should create separate sections.

12. Incorrect HTTP Header

Some bots can refuse to index the file if there is a mistake in the

HTTP header.

Wrong

User-agent: *

Disallow: /css/

Host: www.example.com

Wrong

Content-Type: text/html

Correct

User-agent: *

Disallow: /css/

User-agent: Googlebot

Disallow: /css/

Host: www.example.com

Correct

Content Type: text/plain

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

The main panel shows us all the pages that are blocked.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Checking Pages Blocked with Robots.txt

Let’s use Screaming Frog to check the web pages that are blocked

with our robots.txt file.

1. Go to the right panel and choose ‘Overview’ (1), ‘Response Codes’

(2), ‘Blocked by Robots.txt’ (3).

2. Check to ensure that no pages with essential content are

occasionally hidden from search engines.

3. Choose ‘User Agent’ to test robots.txt for various search engines.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

4. Specify which search engine bots the tool should imitate.

5. You may test various robots.txt sections by repeating the entire

process and pressing ‘Start.’

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

What Are Meta Robots Tags?

Meta robots tags (REP tags) are elements of an indexer directive that

tell search engine spiders how to crawl and index specific pages on

your website.

They enable SEO professionals to target individual pages and

instruct crawlers on what to follow and what not to follow.

Meta Robots Tags Basics

You may hide pages from indexing in several ways, including

meta robots tags implementation. Here you can use the following

directives:

all – No limitations for indexing and content demonstration. This

directive is being used by default and has no impact on the search

engines’ work, unless otherwise specified.

noindex – Do not show this page and the ‘Saved Copy’ link in the

SERPs.

nofollow – Do not allow following the on-page links.

none – The same as noindex, and nofollow meta tags.

noarchive – Do not show the ‘Saved Copy’ link in the SERPs.

nosnippet – Do not show the extended description version of this

page in the SERPs.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

notranslate – Do not offer this page’s translation in the SERPs.

noimageindex – Do not index the on-page images.

unavailable_after: [RFC-850 date/time] – Do not show this

page in the SERPs after specified date/time. Use RFC 850 format.

How to Use Meta Robots Tags

Meta robots tags are pretty simple to use.

It does not take much time to set up meta robots tags. In four simple

steps, you can take your website indexation process up a level:

1. Access the code of a page by pressing CTRL + U.

2. Copy and paste the <head> part of a page’s code into a separate

document.

3. Provide step-by-step guidelines to developers using this

document. Focus on how, where, and which meta robots tags to

inject into the code.

4. Check to make sure the developer has implemented the tags

correctly. To do so, I recommend using The Screaming Frog SEO

Spider.

The screenshot below demonstrates how meta robots tags may

look (check out the first line of code):

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Meta robots tags are recognized by major search engines: Google,

Bing, Yahoo, and Yandex. You do not have to tweak the code for each

individual search engine or browser (unless they honor specific tags).

Main Meta Robots Tags Parameters

As I mentioned above, there are four main REP tag

parameters: follow, index, nofollow, and noindex. Here is

how you can use them:

index, follow: allow search bots to index a page and follow its links

noindex, nofollow: prevent search bots from indexing a page and

following its links

index, nofollow: allow search engines to index a page but hide its

links from search spiders

noindex, follow: exclude a page from search but allow following its

links (link juice helps increase SERPs)

REP tag parameters vary. Here are some of the rarely used

ones:

none

noarchive

nosnippet

unavailabe_after

noimageindex

nocache

noodp

notranslate

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Meta robots tags are essential if you need to optimize specific

pages. Just access the code and instruct developers on what to do.

If your site runs on an advanced CMS (OpenCart, PrestaShop) or

uses specific plugins (like WP Yoast), you can also inject meta tags

and their parameters directly into page templates. This allows you to

cover multiple pages at once without having to ask developers for

help.

Robots.txt & Meta Robots

Tags Non-Compliance

Incoherence between directives in robots.txt and on-page meta tags

is a common mistake.

For example, the robots.txt file hides the page from indexing, but the

meta robots tags do the opposite.

In such cases, Google will

pay attention to what is

prohibited by the robots.

txt file. Most likely, bots will

ignore the directives that

encourage indexing of the

content.

Pay attention to the

fact that robots.txt is

a recommendation by

Google, but not a demand.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Therefore, you still have a chance to see your page indexed, as long

as there are external links that lead to them.

If robots.txt does not hide the page, but the directives do – Google

bots will accomplish the most restricting task and will not index the

on-page content.

The conclusion is simple: eliminate non-compliance between meta

robots tags and robots.txt to clearly show Google which pages

should be indexed, and which should not.

Another noteworthy example is incoherence between on-page meta

tags.

Yandex search bots opt for positive value when they notice conflicts

between the meta tags on a page:

<meta name= “robots” content=”all”/>

<meta name=”robots” content=”noindex, follow”/>

<!–Bots will choose the ‘all’ value and index all the links and texts.–>

By contrast, Google bots opt for the strongest directive, indexing only

links and ignoring the content.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

The sitemap.xml, robots.txt and meta robots tags instructions

complement one another when set up correctly.

The major rules are:

Sitemap.xml, robots.txt and meta robots tags should not be

conflicting.

All the pages that are blocked in robots.txt and meta robots tags

must be excluded from sitemap.xml as well.

All the pages that are opened for indexing must be included in the

sitemap.xml as well.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

The Sitemap.xml Role

The sitemap.xml, robots.txt and meta robots tags instructions

complement one another when set up correctly.

The major rules are:

Sitemap.xml, robots.txt and meta robots tags should not be

conflicting.

All the pages that are blocked in robots.txt and meta robots tags

must be excluded from sitemap.xml as well.

All the pages that are opened for indexing must be included in the

sitemap.xml as well.

To Sum It Up

Knowing how to set up and use a robots.txt file and meta robots

tags is extremely important. A single mistake can spell death to your

entire campaign.

I personally know several digital marketers who have spent months

doing SEO, only to realize that their websites were closed to

indexation in robots.txt. Others abused the “nofollow” tag so much

that they lost backlinks in droves.

Dealing with robots.txt files and REP tags is pretty technical, which

can potentially lead to many mistakes. Fortunately, there are several

basic rules that will help you implement them successfully.

Robots.txt

1. Place your robots.txt file in the top-level directory of your website

code to simplify crawling and indexing.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

However, a few exceptions exist:

Starting the second pagination page, you should add ‘noindex,

follow’ to the meta robots tags, leaving those pages open for

indexing in robots.txt.

Consider adding all the pagination pages to the si temap.xml, so

all the links can be re-indexed.

2. Structure your robots.txt properly, like this: User-agent - Disallow

- Allow - Host - Sitemap. This way, search engine spiders access

categories and web pages in the appropriate order.

3. Make sure that every URL you want to “Allow:” or “Disallow:” is

placed on an individual line. If several URLs appear on one single

line, crawlers will have a problem accessing them.

4. Use lowercase to name your robots.txt. Having “robots.txt” is

always better than “Robots.TXT.” Also, file names are case sensitive.

5. Do not separate query parameters with spacing. For instance, a

line query like this “/cars/ /audi/” would cause mistakes in the robots.

txt file.

6. Do not use any special characters except * and $. Other

characters are not recognized.

7. Create separate robots.txt files for different subdomains. For

example, “hubspot.com” and “blog.hubspot.com” have individual files

with directory- and page-specific directives.

8. Use # to leave comments in your robots.txt file. Crawlers do not

honor lines with the # character.

9. Do not rely on robots.txt for security purposes. Use passwords

and other security mechanisms to protect your site from hacking,

scraping, and data fraud.

Meta Robots Tags

Be case sensitive. Google and other search engines may recognize

attributes, values, and parameters in both uppercase and lowercase,

and you can switch between the two if you want. I strongly

recommend that you stick to one option to improve code readability.

Avoid multiple <meta> tags. By doing this, you will avoid conflicts

in code. Use multiple values in your <meta> tag, like this: <meta

name=“robots” content=“noindex, nofollow”>.

Do not use conflicting meta tags to avoid indexing mistakes.

For example, if you have several code lines with meta tags like

this <meta name=“robots” content=“follow”> and this <meta

name=“robots” content=“nofollow”>, only “nofollow” will be taken

into account. This is because robots put restrictive values first.

Note: You can easily implement both robots.txt and meta robots

tags on your site. However, be careful to avoid confusion between

the two.

The basic rule here is, restrictive values take precedent. So, if you

“allow” indexing of a specific page in a robots.txt file but accidentally

“noindex” it in the <meta>, spiders will not index the page.

Also, remember: If you want to give instructions specifically to

Google, use the <meta> “googlebot” instead of “robots”, like this:

<meta name=“googlebot” content=“nofollow”>. It is similar to

“robots” but avoids all the other search crawlers.

BEST PRACTICES FOR SETTING UP META ROBOTS TAGS & ROBOTS.TXT

4

Your Indexed Pages

Are Going Down – 5

Possible Reasons Why

Chapter 5

Written By

SEO Director, Myers Media Group

Benj Arriola

5

Getting your webpages

indexed by Google (and other

search engines) is essential.

Pages that aren’t indexed

can’t rank.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

How do you see how many pages you have indexed? You can:

Use the site: operator.

Check the status of your XML Sitemap Submissions in Google

Search Console.

Check your overall indexation status.

Each will give different numbers, but why they are different is another

story.

For now, let’s just talk about analyzing a decrease in the number of

indexed pages reported by Google.

5

If your pages aren’t being indexed, this could be a sign that Google

may not like your page or may not be able to easily crawl it.

Therefore, if your indexed page count begins to decrease,

this could be because either:

You’ve been slapped with a Google penalty.

Google thinks your pages are irrelevant.

Google can’t crawl your pages.

Here are a few tips on how to diagnose and fix the issue of

decreasing numbers of indexed pages.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

5

Make sure they have the proper 200 HTTP Header Status.

Did the server experience frequent or long downtime? Did the

domain recently expire and was renewed late?

Action Item

You can use a free HTTP Header Status checking tool to

determine whether the proper status is there. For massive sites,

typical crawling tools like Xenu, DeepCrawl, Screaming Frog, or

Botify can test these.

The correct header status is 200. Sometimes some 3xx (except

the 301), 4xx, or 5xx errors may appear – none of these are good

news for the URLs you want to be indexed.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

5

1. Are the Pages Loading

Properly?

Sometimes a change in CMS, backend programming, or server

setting that results in a change in domain, subdomain, or folder may

consequently change the URLs of a site.

Search engines may remember the old URLs but, if they don’t

redirect properly, a lot of pages can become de-indexed.

Action Item

Hopefully a copy of the old site can still be visited in some way

or form to take note of all old URLs so you can map out the 301

redirects to the corresponding URLs.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

5

2. Did Your URLs Change

Recently?

Fixing duplicate content often involves implementing canonical tags,

301 redirects, noindex meta tags, or disallows in robots.txt. All of

which can result in a decrease in indexed URLs.

This is one example where the decrease in indexed pages might be

a good thing.

Action Item

Since this is good for your site, the only thing you need to do is to

double check that this is definitely the cause of the decrease of

indexed pages and not anything else.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

5

3. Did You Fix Duplicate Content

Issues?

Some servers have bandwidth restrictions because of the

associated cost that comes with a higher bandwidth; these servers

may need to be upgraded. Sometimes, the issue is hardware related

and can be resolved by upgrading your hardware processing or

memory limitation.

Some sites block IP addresses when visitors access too many

pages at a certain rate. This setting is a strict way to avoid any

DDOS hacking attempts but it can also have a negative impact on

your site.

Typically, this is monitored at a page’s second setting and if the

threshold is too low, normal search engine bot crawling may hit the

threshold and the bots cannot crawl the site properly.

Action Item

If this is a server bandwidth limitation, then it might be an

appropriate time to upgrade services.

If it is a server processing/memory issue, aside from upgrading

the hardware, double check if you have any kind of server caching

technology in place, this will give less stress on the server.

If an anti-DDOS software is in place, either relax the settings or

whitelist Googlebot to not be blocked anytime. Beware though,

there are some fake Googlebots out there; be sure to detect

googlebot properly. Detecting Bingbot has a similar procedure.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

5

4. Are Your Pages Timing Out?

Sometimes what search engine spiders see is different than what

we see.

Some developers build sites in a preferred way without knowing the

SEO implications.

Occasionally, a preferred out-of-the-box CMS will be used without

checking if it is search engine friendly.

Sometimes, it might have been done on purpose by an SEO

who attempted to do content cloaking, trying to game the search

engines.

Other times, the website has been compromised

by hackers, who cause a different page

to be shown to Google to promote

their hidden links or cloak the

301 redirections to their own

site.

The worse situation would

be pages that are infected

with some type of malware

that Google automatically

deindexes the page

immediately once detected.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

5

5. Do Search Engine Bots See

Your Site Differently?

Action Item

Using Google Search Console’s fetch and render feature is

the best way to see if Googlebot is seeing the same content as

you are.

You may also try to translate the page in Google Translate

even if you have no intention to translate the language or check

Google’s Cached page, but there are also ways around these

to still cloak content behind them.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

5

Index Pages Are Not Used as

Typical KPIs

A Decrease in Indexed Pages Isn’t

Always Bad

Key Performance Indicators (KPIs), which help measure the success

of an SEO campaign, often revolve around organic search traffic and

ranking. KPIs tend to focus on the goals of a business, which are tied

to revenue.

An increase in indexed pages pages may increase the possible

number of keywords you can rank for that can result in higher

profits. However, the point of looking at indexed pages is mainly just

to see whether search engines are able to crawl and indexed your

pages properly.

Remember, your pages can’t rank when search engines can’t see,

crawl, or index them.

Most of the time, a decrease in indexed pages could mean a bad

thing, but a fix to duplicate content, thin content, or low-quality

content might also result in a decreased number of indexed pages,

which is a good thing.

Learn how to evaluate your site by looking at these five possible

reasons why your indexed pages are going down.

YOUR INDEXED PAGES ARE GOING DOWN – 5 POSSIBLE REASONS WHY

5

An SEO Guide to HTTP

Status Codes

Chapter 6

Written By

SEO Director, Site Objective

Brian Harnish

6

One of the most important

assessments in any SEO audit

is determining what hypertext

transfer protocol status codes

(or HTTP Status Codes) exist on a

website.

AN SEO’S COMPLETE GUIDE TO HTTP: STATUS CODES

These codes can become complex, often turning into a hard puzzle that

must be solved before other tasks can be completed.

For instance, if you put up a page that all of a sudden disappears with a

404 not found status code, you would check server logs for errors and

assess what exactly happened to that page.

6

If you are working on an audit, other status codes can be a

mystery, and further digging may be required.

These codes are segmented into different types:

1xx status codes are informational codes.

2xx codes are success codes.

3xx redirection codes are redirects.

4xx are any codes that fail to load on the client side, or client error

codes.

5xx are any codes that fail to load due to a server error.

AN SEO GUIDE TO HTTP STATUS CODES

6

1xx Informational Status Codes

These codes are informational in nature and usually have no realworld

impact for SEO.

100 – Continue

Definition: In general, this protocol designates that the initial serving

of a request was received and not yet otherwise rejected by the

server.

SEO Implications: None

Real World SEO Application: None

101 - Switching Protocols

Definition: The originating server of the site understands, is willing

and able to fulfill the request of the client via the Upgrade header

field. This is especially true for when the application protocol on the

same connection is being used.

SEO Implications: None

Real World SEO Application: None

102 – Processing

Definition: This is a response code between the server and the

client that is used to inform the client side that the request to the

server was accepted, although the server has not yet completed the

request.

SEO Implications: None

Real World SEO Application: None

AN SEO GUIDE TO HTTP STATUS CODES

6

2xx Client Success Status Codes

This status code tells you that a request to the server was

successful. This is mostly only visible server-side. In the real world,

visitors will never see this status code.

SEO Implications: A page is loading perfectly fine, and no action

should be taken unless there are other considerations (such as

during the execution of a content audit, for example).

Real-World SEO Application: If a page has a status code of 200

OK, you don’t really need to do much to it if this is the only thing you

are looking at. There are other applications involved if you are doing

a content audit, for example.

However, that is beyond the scope of this article, and you should

already know whether or not you will need a content audit based on

initial examination of your site.

How to find all 2xx success codes on a website via Screaming Frog:

There are two ways in Screaming Frog that you can find 2xx HTTP

success codes: through the GUI, and through the bulk export

option.

Method 1 – Through the GUI

1. Crawl your site using the settings that you are comfortable with.

2. All of your site URLs will show up at the end of the crawl.

3. Look for the Status Code column. Here, you will see all 200 OK,

2xx based URLs.

AN SEO GUIDE TO HTTP STATUS CODES

6

Method 2 – The Bulk Export Option

1. Crawl your site using the settings that you are comfortable with.

2. Click on Bulk Export

3. Click on Response Codes

4. Click on 2xx Success Inlinks

AN SEO GUIDE TO HTTP STATUS CODES

6

201 – Created

This status code will tell you that the server request has been

satisfied and that the end result was that one or multiple resources

were created.

202 – Accepted

This status means that the server request was accepted to be

processed, but the processing has not been finished yet.

203 – Non-Authoritative Information

A transforming proxy modified a successful payload from the origin

server’s 200 OK response.

204 – No Content

After fulfilling the request successfully, no more content can be sent

in the response payload body.

205 – Reset Content

This is similar to the 204 response code, except the response

requires the client sending the request reset the document view.

206 – Partial Content

Transfers of one or more components of the selected page that

corresponds to satisfiable ranges that were found in the range

header field of the request. The server, essentially, successfully

fulfilled the range request for said target resource.

207 – Multi-Status

In situations where multiple status codes may be the right thing, this

multi-status response displays information regarding more than one

resource in these situations.

AN SEO GUIDE TO HTTP STATUS CODES

6

3xx Redirection Status Codes

How Can 301 Redirects Impact

SEO?

Mostly, 3xx Redirection codes denote redirects. From temporary to

permanent. 3xx redirects are an important part of preserving SEO

value.

That’s not their only use, however. They can explain to Google

whether or not a page redirect is permanent, temporary, or

otherwise.

In addition, the redirect can be used to denote pages of content that

are no longer needed.

301 – Moved Permanently

These are permanent redirects. For any site migrations, or other

situations where you have to transfer SEO value from one URL to

another on a permanent basis, these are the status codes for the

job.

Google has said several things about the use of 301 redirects and

their impact. John Mueller has cautioned about their use.

“So for example, when it comes to links, we will say well, it’s this link

between this canonical URL and that canonical URL- and that’s how

we treat that individual URL.

AN SEO GUIDE TO HTTP STATUS CODES

6

In that sense it’s not a matter of link equity loss across redirect

chains, but more a matter of almost usability and crawlability. Like,

how can you make it so that Google can find the final destination

as quickly as possible? How can you make it so that users don’t

have to jump through all of these different redirect chains. Because,

especially on mobile, chain redirects, they cause things to be really

slow.

If we have to do a DNS lookup between individual redirects, kind of

moving between hosts, then on mobile that really slows things down.

So that’s kind of what I would focus on there.

Not so much like is there any PageRank being dropped here. But

really, how can I make it so that it’s really clear to Google and to users

which URLs that I want to have indexed. And by doing that you’re

automatically reducing the number of chain redirects.”

It is also important to note here that not all 301 redirects will pass

100 percent link equity. From Roger Montti’s reporting:

“A redirect from one page to an entirely different page will result in no

PageRank being passed and will be considered a soft 404.”

John Mueller also mentioned previously:

“301-redirecting for 404s makes sense if you have 1:1 replacement

URLs, otherwise we’ll probably see it as soft-404s and treat like a

404.”

AN SEO GUIDE TO HTTP STATUS CODES

6

302 – Found

Also known as temporary redirects, rather than permanent

redirects. They are a cousin of the 301 redirects with one important

difference: they are only temporary.

You may find 302s instead of 301s on sites where these redirects

have been improperly implemented.

Usually, they are done by developers who don’t know any better.

The other 301 redirection status codes that you may come across

include:

300 – Multiple Choices

This redirect involves multiple documents with more than one

version, each having its own identification. Information about these

documents is being provided in a way that allows the user to select

the version that they want.

303 – See Other

A URL, usually defined in the location header field, redirects the user

agent to another resource. The intention behind this redirect is to

provide an indirect response to said initial request.

The matching of the topic of the page in this instance is what’s

important. “the 301 redirect will pass 100 percent PageRank only if

the redirect was a redirect to a new page that closely matched the

topic of the old page.”

AN SEO GUIDE TO HTTP STATUS CODES

6

304 – Not Modified

The true condition, which evaluated false, would normally have

resulted in a 200 OK response should it have evaluated to true.

Applies to GET or HEAD requests mostly.

305 – Use Proxy

This is now deprecated, and has no SEO impact.

307 – Temporary Redirect

This is a temporary redirection status code that explains that the

targeted page is temporarily residing on a different URL. It lets the

user agent know that it must NOT make any changes to the method

of request if an auto redirect is done to that URL.

308 – Permanent Redirect

Mostly the same as a 301 permanent redirect.

AN SEO GUIDE TO HTTP STATUS CODES

6

4xx Client Error Status Codes

4xx client error status codes are those status codes that tell us that

something is not loading – at all – and why.

While the error message is a subtle difference between each code,

the end result is the same. These errors are worth fixing and should

be one of the first things assessed as part of any website audit.

Error 400 Bad Request

403 Forbidden

404 Not Found

These statuses are the most common requests an SEO will

encounter – the 400, 403 and 404 errors. These errors simply mean

that the resource is unavailable and unable to load.

Whether it’s due to a temporary server outage, or other reason,

it doesn’t really matter. What matters is the end result of the bad

request – your pages are not being served by the server and is

There are two ways to find 4xx errors that are plaguing a site in

Screaming Frog – through the GUI, and through bulk export.

Screaming Frog GUI Method:

1. Crawl your site using the settings that you are comfortable with.

2. Click on the down arrow to the right.

3. Click on response codes.

4. Filter by Client Error (4xx).

AN SEO GUIDE TO HTTP STATUS CODES

6

Screaming Frog Bulk

Export Method:

1. Crawl your site with the

settings you are familiar

with.

2. Click on Bulk Export.

3. Click on Response

Codes.

4. Click on Client error

(4xx) Inlinks.

AN SEO GUIDE TO HTTP STATUS CODES

6

These are other 4xx errors that you may come across, including:

401 - Unauthorized

402 - Payment Required

405 - Method Not Allowed

406 - Not Acceptable

407 - Proxy Authentication Required

408 - Request Timeout

409 - Conflict

410 - Gone

411 - Length Required

412 - Precondition Failed

413 - Payload Too Large

414 - Request-URI Too Long

415 - Unsupported Media Type

416 - Requested Range Not Satisfiable

417 - Expectation Failed

418 - I’m a teapot

421 - Misdirected Request

422 - Unprocessable Entity

423 - Locked

424 - Failed Dependency

426 - Upgrade Required

428 - Precondition Required

429 - Too Many Requests

431 - Request Header Fields Too Large

444 - Connection Closed Without Response

451 - Unavailable For Legal Reasons

499 - Client Closed Request

AN SEO GUIDE TO HTTP STATUS CODES

6

5xx Server Error Status Codes

All of these errors imply that there is something wrong at the server

level that is preventing the full processing of the request.

The end result will always (in most cases that serve us as SEOs) be

the fact that the page does not load and will not be available to the

client side user agent that is viewing it.

This can be a big problem for SEOs.

Again, using Screaming Frog, there are two methods you can use

to get to the root of the problems being caused by 5xx errors on a

website. A GUI method, and a Bulk Export method.

Screaming Frog GUI Method for Unearthing 5xx Errors

1. Crawl your site using the settings that you are comfortable with.

2. Click on the dropdown arrow on the far right.

3. Click on “response codes”.

4. Click on Filter > Server Error (5xx)

5. Select Server Error (5xx).

6. Click on Export

AN SEO GUIDE TO HTTP STATUS CODES

6

Screaming Frog Bulk Export Method for Unearthing 5xx Errors

AN SEO GUIDE TO HTTP STATUS CODES

6

1. Crawl your site using the settings you are comfortable with.

2. Click on Bulk Export.

3. Click on Response Codes.

4. Click on Server Error (5xx) Inlinks.

This will give you all of the 5xx errors that are presenting on your

site.

AN SEO GUIDE TO HTTP STATUS CODES

6

There are other 5xx http status codes that you may come across,

including the following:

500 - Internal Server Error

501 - Not Implemented

502 - Bad Gateway

503 - Service Unavailable

504 - Gateway Timeout

505 - HTTP Version Not Supported

506 - Variant Also Negotiates

507 - Insufficient Storage

508 - Loop Detected

510 - Not Extended

511 - Network Authentication Required

599 - Network Connect Timeout Error

AN SEO GUIDE TO HTTP STATUS CODES

6

Making Sure That HTTP Status

Codes Are Corrected On Your Site

Is a Good First Step

When it comes to making a site that is 100 percent crawlable, one of

the first priorities is making sure that all content pages that you want

the search engines to know about are 100 percent crawlable. This

means making sure that all pages are 200 OK.

Once that is complete, you will be able to move forward with more

SEO audit improvements as you assess priorities and additional

areas that need to be improved.

“A website’s work is never done” should be an SEO’s mantra. There

is always something that can be improved on a website that will

result in improved search engine rankings.

If someone says that their

site is perfect, and that they

need no further changes,

then I have a $1 million

dollar bridge to sell you in

Florida.

AN SEO GUIDE TO HTTP STATUS CODES

6

404 vs. Soft 404

Errors: What’s the

Difference & How to

Fix Both

Chapter 7

Written By

SEO Director, Myers Media Group

Benj Arriola

7

Every page that loads in a

web browser has a response

code included in the HTTP

headers, which may or may

not be visible on the web

page itself.

404 VS SOFT 404 ERRORS: WHAT’S THE DIFFERENCE & HOW TO FIX BOTH

There are many different response codes a server gives to

communicate the loading-status of the page; one of the most well-known

codes is the 404-response code.

Generally, any code within 400 to 499 indicates that the page didn’t load.

The 404-response code is the only one that carries a specific meaning

– that the page is actually gone and probably isn’t coming back anytime

soon.

7

What’s a Soft 404 Error?

Potentially Misidentified as Soft

404

A soft 404 error isn’t an official response code sent to a web

browser. It’s just a label Google adds to a page within their index.

As Google crawls pages, it allocates resources carefully ensuring

that no time is wasted by crawling missing pages which do not

need to be indexed.

However, there are some servers that are poorly configured and

their missing page loads a 200 code when it should display a

404-response code. If the invisible HTTP header displays a 200

code even if the web page clearly states that the page isn’t found,

the page might be indexed, which is a waste of resources for

Google.

To combat this issue, Google notes the characteristics of 404 pages

and attempts to discern whether the 404 page really is a 404 page.

In other words, Google learned that if it looks like a 404, smells like a

404, and acts like a 404, then it’s probably a genuine 404 page.

There are also cases wherein the page isn’t actually missing, but

certain characteristics have triggered Google to categorize it as a

missing page.

Some of these characteristics include a small amount or lack of

content on the page and having too many similar pages on the site.

404 VS. SOFT 404 ERRORS: WHAT’S THE DIFFERENCE & HOW TO FIX BOTH

7

These characteristics are also similar to the factors that the Panda

algorithm tackles. The Panda update considers thin and duplicate

content as negative ranking factors.

Therefore, fixing these issues will help avoid both soft 404s and

Panda issues.

404 errors have two main causes:

An error in the link, directing users to a page that doesn’t exist.

A link going to a page that used to exist and suddenly disappeared.

Linking Error

If the cause of the 404 is a linking error, you just have to fix the links.

The difficult part of this task is finding all the broken links on a site.

It can be more challenging for large, complex sites that have

thousands or millions of pages. In instances like this, crawling

tools come in handy. You can try using software such as Xenu,

DeepCrawl, Screaming Frog, or Botify.

A Page That No Longer Exists

When a page no longer exists, you have two options:

Restore the page if it was accidentally removed.

301 redirect it to the closest related page if it was removed on

purpose.

404 VS. SOFT 404 ERRORS: WHAT’S THE DIFFERENCE & HOW TO FIX BOTH

7

First, you have to locate all the linking errors on the site. Similar to

finding all errors in linking for a large scale website, you can use

crawling tools. However, crawling tools may not find orphaned

pages, which are pages that are not linked from anywhere within

the navigational links or from any of the pages.

Orphaned pages can exist if they used to be part of the website,

then after a website redesign, the link going to this old page

disappeared, but external links from other websites might still be

linking to them. To double check if these kinds of pages exist on

your site, you can use a variety of tools.

Google Search Console

Search console will report 404 pages as Google’s crawler goes

through all the pages it can find. This can include links from other

sites going to a page that used to exist on your website.

Google Analytics

You won’t find a missing page report in

Google Analytics by default. However,

you can track them in a number of ways.

For one, you can create a

custom report

and segment

out pages

that have a

page title

mentioning Error

404 – Page Not Found.

404 VS. SOFT 404 ERRORS: WHAT’S THE DIFFERENCE & HOW TO FIX BOTH

7

Another way to find orphaned pages within Google Analytics is to

create custom content groupings and to assign all 404 pages to a

content group.

Site: Operator Search Command

Searching Google for “site:example.com” will list all pages of

example.com that are indexed by Google. You can then individually

check if the pages are loading or if they’re giving 404s.

To do this at scale, I like using WebCEO, which has a feature to

run the site: operator not only on Google, but also on Bing, Yahoo,

Yandex, Naver, Baidu, and Seznam.

Since all the search engines will only give you a subset, running it

on multiple search engines can help give a larger list of pages of

your site. This list can be exported and run on tools for a mass 404

check. I simply do this by adding all URLs as links within an HTML

file and loading it on Xenu to massively check for 404 errors.

Other Backlink Research Tools

Backlink research tools like Majestic, Ahrefs, Moz Open Site

Explorer, Sistrix, LinkResearchTools, and CognitiveSEO can also

help.

Most of these tools will export a list of backlinks linking to your

domain. From there, you can check all the pages that are being

linked to and look for 404 errors.

404 VS. SOFT 404 ERRORS: WHAT’S THE DIFFERENCE & HOW TO FIX BOTH

7

How to Fix Soft 404 Errors

Crawling tools won’t detect a soft 404 because it isn’t really a 404

error. But you can use crawling tools to detect something else.

Here are a few things to find:

Thin Content: Some crawling tools not only report pages that

have thin content, but also show a total word count. From there,

you can sort URLs based on your content’s number of words.

Start with pages that have the least amount of words and evaluate

whether the page has thin content.

Duplicate Content: Some crawling tools are sophisticated

enough to discern what percentage of the page is template

content. If the main content is nearly the same as many other

pages, you should look into these pages and determine why

duplicate content exists on your site.

Aside from the crawling tools, you can also use Google Search

Console and check under crawl errors to find pages that are listed

under soft 404s.

Crawling an entire site to find issues that cause soft 404s allows

you to locate and correct problems before Google even detects

them.

After detecting these soft 404 issues, you will need to correct them.

404 VS. SOFT 404 ERRORS: WHAT’S THE DIFFERENCE & HOW TO FIX BOTH

7

Most of the time, the solutions appear to be common sense. This

can include simple things like expanding pages with thin content

or replacing duplicate content with new and unique ones.

Throughout this process, here are a few things to

consider:

Consolidate Pages: Sometimes thin content is caused by being

too specific with the page topic, which can leave you with little

to say. Merging several thin pages into one page can be more

appropriate if the topics are related. Not only does this solve thin

content issues, but it can fix duplicate content issues as well. For

example, an e-commerce site selling shoes that come in different

colors and sizes may have a different URL for each size and color

combination. This leaves a large number of pages with content

that is thin and relatively identical. The more effective approach

is to put this all on one page instead and enumerate the options

available.

Find Technical Issues That Cause Duplicate Content: Using

even the simplest web crawling tool like Xenu (which doesn’t look

at content but only URLs, response codes, and title tags), you can

still find duplicate content issues by looking at URLs. This includes

things like www vs non-www URLs, http and https, with index.html

and without, with tracking parameters and without, etc. A good

summary of these common duplicate content issues found in

URLs patterns can be found on slide 6 of this presentation.

404 VS. SOFT 404 ERRORS: WHAT’S THE DIFFERENCE & HOW TO FIX BOTH

7

Google Treats 404 Errors & Soft

404 Errors the Same Way

A soft 404 is not real 404 error, but Google will deindex those pages

if they aren’t fixed quickly. It is best to crawl your site regularly to

see if 404 or soft 404 errors occur. Crawling tools should be a major

component of your SEO arsenal. 404 VS. SOFT 404 ERRORS: WHAT’S THE DIFFERENCE &

HOW TO FIX BOTH

7

8 Tips to Optimize

Crawl Budget for SEO

Chapter 8

Written By

Founder and CMO, SEO PowerSuite

Aleh Barysevich

8

When you hear the words

“search engine optimization,”

what do you think of? 8

TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

My mind leaps straight to a list of SEO ranking factors, such as proper

tags, relevant keywords, a clean sitemap, great design elements, and a

steady stream of high-quality content.

However, a recent article by my colleague, Yauhen Khutarniuk, made

me realize that I should be adding “crawl budget” to my list.

While many SEO experts overlook crawl budget because it’s not very well

understood, Khutarniuk brings some compelling evidence to the table –

which I’ll come back to later in this chapter – that crawl budget can, and

should, be optimized.

This made me wonder: how does crawl budget optimization overlap with

SEO, and what can websites do to improve their crawl rate?

8

First Things First – What Is a Crawl

Budget?

Web services and search engines use web crawler bots, aka

“spiders,” to crawl web pages, collect information about them, and

add them to their index. These spiders also detect links on the

pages they visit and attempt to crawl these new pages too.

Examples of bots that you’re probably familiar with include

Googlebot, which discovers new pages and adds them to the

Google Index, or Bingbot, Microsoft’s equivalent.

Most SEO tools and other web services also rely on spiders to

gather information. For example, my company’s backlink index, SEO

PowerSuite Backlink Index, is built using a spider called BLEXBot,

which crawls up to 7.1 billion web pages daily gathering backlink

data.”

The number of times a search engine spider crawls your website

in a given time allotment is what we call your “crawl budget.” So

if Googlebot hits your site 32 times per day, we can say that your

typical Google crawl budget is approximately 960 per month.

You can use tools such

as Google Search

Console and Bing

Webmaster

Tools to figure

out your

8 TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

8

Is Crawl Budget Optimization the

Same as SEO?

Yes – and no. While both types of optimization aim to make your

page more visible and may impact your SERPs, SEO places a

heavier emphasis on user experience, while spider optimization is

entirely about appealing to bots.

So how do you optimize your crawl budget specifically? I’ve

gathered the following nine tips to help you make your website as

crawlable as possible.

8 TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

8

website’s approximate crawl budget. Just log in to Crawl > Crawl

Stats to see the average number of pages crawled per day.

How to Optimize Your Crawl

Budget

1. Ensure Your Pages Are Crawlable

Your page is crawlable if search engine spiders can find and follow

links within your website. You’ll have to configure your .htaccess and

robots.txt so that they don’t block your site’s critical pages.

You may also want to provide text versions of pages that rely heavily

on rich media files, such as Flash and Silverlight.

Of course, the opposite is true if you do want to prevent a page from

showing up in search results.

However, it’s not enough to simply set your Robots.txt to “Disallow,”

if you want to stop a page from being indexed. According to

Google: “Robots.txt Disallow does not guarantee that a page will

not appear in results.”

If external information (e.g., incoming links) continue to direct traffic

to the page that you’ve disallowed, Google may decide the page is

still relevant.

In this case, you’ll need to manually block the page from being

indexed by using the noindex robots meta tag or the X-Robots-Tag

HTTP header.

noindex meta tag: Place the following meta tag in the <head>

section of your page to prevent most web crawlers from indexing

your page:

8 TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

8

noindex” />

X-Robots-Tag: Place the following in your HTTP header response to

tell crawlers not to index a page:

X-Robots-Tag: noindex

Note that if you use noindex meta tag or X-Robots-Tag, you should

not disallow the page in robots.txt, The page must be crawled

before the tag will be seen and obeyed.

2. Use Rich Media Files Cautiously

There was a time when Googlebot couldn’t crawl content like

JavaScript, Flash, and HTML. Those times are mostly past (though

Googlebot still struggles with Silverlight and some other files).

However, even if Google can read most of your rich media files,

other search engines may not be able to, which means that you

should use these files judiciously, and you probably want to avoid

them entirely on the pages you want to be ranked.

You can find a full list of the files that Google can index here.

3. Avoid Redirect Chains

Each URL you redirect to wastes a little of your crawl budget.

When your website has long redirect chains, i.e., a large number of

301 and 302 redirects in a row, spiders such as Googlebot may drop

off before they reach your destination page, which means that page

won’t be indexed.

8 TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

8

Best practice with redirects is to have as few as possible on your

website, and no more than two in a row.

4. Fix Broken Links

When asked whether or not broken links affect web ranking,

Google’s John Mueller once said:

If what Mueller says is true, this is one of the fundamental

differences between SEO and Googlebot optimization, because

it would mean that broken links do not play a substantial role in

rankings, even though they greatly impede Googlebot’s ability to

index and rank your website.

That said, you should take Mueller’s advice with a grain of salt –

Google’s algorithm has improved substantially over the years, and

anything that affects user experience is likely to impact SERPs.

5. Set Parameters on Dynamic URLs

Spiders treat dynamic URLs that lead to the same page as separate

pages, which means you may be unnecessarily squandering your

crawl budget.

8 TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

8

You can manage your URL parameters by going to your Google

Search Console and clicking Crawl > Search Parameters.

From here, you can let Googlebot know if your CMS adds

parameters to your URLs that doesn’t change a page’s content.

6. Clean Up Your Sitemap

XML sitemaps help both your users and spider bots alike, by making

your content better organized and easier to find.

Try to keep your sitemap up-to-date and purge it of any clutter

that may harm your site’s usability, including 400-level pages,

unnecessary redirects, non-canonical pages, and blocked pages.

The easiest way to clean up your sitemap is to use a tool like

Website Auditor (disclaimer: my tool). You can use Website Auditor’s

XML sitemap generator to create a clean sitemap that excludes all

pages blocked from indexing.

Plus, by going to Site Audit, you can easily find and fix all 4xx status

pages, 301

and 302

redirects,

and noncanonical

pages.

8 TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

8

7. Build External Links

Link building is still a hot topic – and I doubt it’s going away anytime

soon. As SEJ’s Anna Crowe elegantly put it:

“Cultivating relationships online, discovering new communities,

building brand value – these small victories should already be

imprints on your link-planning process. While there are distinct

elements of link building that are now so 1990s, the human need to

connect with others will never change.”

Now, in addition to Crowe’s excellent point, we also have evidence

from Yauhen Khutarniuk’s experiment that external links closely

correlate with the number of spider visits your website receives.

In his experiment, he used our tools to measure all of the internal

and external links pointing to every page on 11 different sites. He

then analyzed crawl stats on each page and compared the results.

This is an example of what he found on just one of the sites he

analyzed:

While the data set couldn’t prove any conclusive connection

between internal links and crawl rate, Khutarniuk did find an overall

“strong correlation (0,978) between the number of spider visits and

the number of external links.”

8 TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

8

8. Maintain Internal Link Integrity

While Khutarniuk’s experiment proved that internal link building

doesn’t play a substantial role in crawl rate, that doesn’t mean you

can disregard it altogether.

A well-maintained site structure makes your content easily

discoverable by search bots without wasting your crawl budget.

A well-organized internal linking structure may also improve user

experience – especially if users can reach any area of your website

within three clicks.

Making everything more easily accessible in general means visitors

will linger longer, which may improve your SERPs.

8 TIPS TO OPTIMIZE CRAWL BUDGET FOR SEO

8

Conclusion: Does Crawl Budget

Matter?

By now, you’ve probably noticed a trend in this article – the bestpractice

advice that improves your crawlability tends to improve

your searchability as well.

So if you’re wondering whether or not crawl budget

optimization is important for your website, the

answer is YES – and it will probably go

hand-in-hand with your SEO efforts

anyway.

Put simply, when you make

it easier for Google to

discover and index your

website, you’ll enjoy

more crawls, which

means faster updates

when you publish new content.

You’ll also improve overall user experience,

which improves visibility, which ultimately

results in better SERPs rankings.

How to Improve Your

Website Navigation:

7 Essential Best

Practices

Chapter 9

Written By

SEO Director, Myers Media Group

Benj Arriola

9

Website navigation, when

done right, is great for

your users and your SEO

performance. HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7

ESSENTIAL BEST PRACTICES

Good website navigation makes it easy for your visitors to find what they

want and for search engines to crawl.

The result: more conversions and greater search visibility.

But how do you actually do it? By using these website navigation best

practices.

9

What is Website Navigation?

Website navigation (a.k.a., internal link architecture) are the links

within your website that connect your pages. The primary purpose

of website navigation is to help users easily find stuff on your site.

Search engines use your website navigation to discover and index

new pages. Links help search engines to understand the content

and context of the destination page, as well as the relationships

between pages.

Users come first. This is the underlying objective of website

navigation you must always remember.

Satisfy users first. Make navigation easy. Then, optimize for search

engines without hurting the user experience.

If you more basic information on website navigation, you’ll

find these SEJ posts helpful:

Internal Linking Guide to Boost Your SEO by Syed Balkhi

Your Essential Guide to Internal Content Linking by Julia

McCoy

The remainder of this post will maintain a broader focus on website

navigation best practices, outlining various internal linking situations

that can cause issues for your website visitors and search engines.

This topic will be especially relevant and important for anyone

working on large websites.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Website Navigation & Content

Hierarchies

When searching for a specific page within a book, you can simply

read through the table of contents or the index. When you walk

around the grocery store, the aisles are labeled with general section

categories and more subcategories are listed on the shelves

themselves. Both provide an efficient way to navigate through a lot of

content.

Content hierarchies exist to simplify the process of locating content.

When a mass amount of content exists, it can be broken down into a

few broad categories.

Within those broad categories, you can create even narrower

classifications; this builds differing hierarchical levels that users

can easily navigate. Utilizing content hierarchies organizes pages

of a website in a way that makes sense to the user and the search

engine.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Importance of Content Hierarchies & Website Navigation

The categorization and sub-categorization of content help pages

improve in rank for general head terms and for specific long-tail

terms.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Problems Caused by Content Hierarchies

Categorization of content and building hierarchies create content

silos, like clusters of closely related topics. Google will crawl different

pages at different rates, following links from different sites.

Some content silos are more popular than others. These pages may

get more external links and traffic than others and, as a result, earn

more prominent positions in organic search.

When content is too siloed and fails to get links and traffic, it might

not perform as well – even if your other content silos perform

extremely well. The content hierarchies can isolate certain popular

page clusters that may be located too deep within the site.

This is where horizontal linking comes into play.

As much as link relevancy helps in ranking, the lack of cross-linking

between content silos can be detrimental to your overall rankings.

There are always ways to create relationships that horizontally link

categories to one another.

The fact that all pages belong to the same website already indicates

that these pages are not completely irrelevant to each other.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Action Items: Linking Between Content Categories

Categorize content in a way that forms category hierarchies that

make sense to the user and interlink these pages properly, going up

and down the hierarchy. These are the majority of the links.

Create cross-linking between pages that are under different

categories but still have similarities.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Links Between Product & Content

Marketing Pages

Companies selling more than one product or service will do

everything mentioned above on categorizing the pages, creating

content silos, and interlinking them.

However, many SEO teams and content teams also create assets

that are designed to be compelling and shareable. Oftentimes, this

comes in the form of a blog, with posts containing links to specific

products and services.

Blog posts can be useful because they direct more traffic toward

product pages. However, many sites fail to link the product pages

back to the blog pages. Using this type of horizontal linking helps

inform users about your product or service and increases your SEO

performance.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Action Items: Linking Between Product and Content

Pages

Product pages should also link back to related content marketing

pages. This may include blog posts, FAQs, and product manuals.

Website Navigation Using

JavaScript Effects

Occasionally, links and web pages are written in JavaScript. This is

a problem because search engines have difficulty locating internal

links that are created in JavaScript.

Although Google has improved in recent years in terms of reading

JavaScript, SEO specialists have concluded that results are

inconsistent. Other search engines still have no capabilities when it

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

comes to reading JavaScript. This means your internal linking could

be completely lost when search engines crawl your content.

The SEO world is divided over whether using JavaScript is practical.

On one hand, some SEO experts avoid JavaScript altogether. On

the other hand, web designers and usability experts claim that

JavaScript is essential to the user experience. I believe there is a

middle ground where JavaScript can be used while avoiding any

SEO issues.

Links That Display and Hide Content Already on the Page

JavaScript can be used to display and hide certain content on a

page without actually changing the page you are on. When this

happens, all of your content is pre-loaded to the page.

In this case, search engines are still able to crawl all of your content,

even when some of it is hidden. This is only successful when the

amount of content that is hidden remains minor; it can become

problematic when the entire page changes but the URL remains the

same.

Problems arise because

of the fact that when you

hide too much content within

one URL, it dilutes the content

focus of what that page is all about. A

completely different topic should have

its own page.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Action Items: Links That Display and Hide Content

For small amounts of content, remove the anchor tag and replace

with a JavaScript onclick event handler.

Use CSS to control the cursor and change from an arrow to a

hand pointer.

For large amounts of content, including single-page parallax

scrolling websites, not all content should be pre-loaded.

Only pre-load content directly related to the URL.

For all anchor tags, there should be an href value and an

onclick setting.

This href value leads to a new URL that only pre-loads

the content related to this new URL.

The onclick function will prevent the new URL from

loading but will allow content from the destination URL

to load.

Use the pushState function to update the URL even if

that page did not load.

A more in-depth presentation of how this can be specifically

implemented on websites is explained well in this presentation

done at seoClarity in 2016. It specifically talks about AngularJS,

a popular JavaScript framework, and its SEO issues and

solutions. However, the lessons here are also applicable to

almost any JavaScript framework.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Using Tracking Parameters in the

URL

Usability experts and conversion optimization specialists track user

behavior in different ways.

Sometimes, this involves using tracking parameters in URLs within

the site. This causes duplicate content issues due to linking to

different URLs that have the exact same content. This can be

resolved in a number of ways.

Action Items: Tracking Parameters in URLs

Avoid using tracking parameters in the URL. Instead, track these by

using JavaScript tracking onclick event handlers on links that will

pass the same tracking parameters. If using Google Analytics, this

can be done with event tracking.

Always using a self-referencing canonical tag is a good practice to

have to avoid many kinds of duplicate content issues.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

The First Link Priority

in search engine crawling where only the first link is considered and

the duplicate link is disregarded. This has been discussed in forums

and tested in 2008 by a number of people, including Rand Fishkin

and myself.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

First Link Priority as Illustrated on Moz, by Rand Fishkin

A few things worth mentioning:

In 2014 Matt Cutts, former head of Google’s spam team, said this is

no longer an issue. I have yet to test this again and I haven’t seen

any other SEO professionals test this recently.

When this was first tested and detected to be an issue, the HTML

version was 4.1, XHTML 1.1 was on the rise, and HTML 5 did not

yet exist. Today, HTML 5 exists with tags like <header>, <article>,

and <sidebar>. Maybe this time Google treats links in the header,

sidebar, and article tags.

SEO Issues That Arise From the First Link Priority

Top-bar navigation and left side-bar often comes first within the

source code before the main content. Additionally, navigational

elements in these menus often have short anchor text. They tend to

be less keyword focused and more design focused.

Links within the main content of a page have a tendency to be

more keyword focused, with surrounding content that supports

the keyword. They are also more flexible in length, with longer,

more specific anchor text; this longer text increases the variety of

keywords that a page can potentially rank for. However, because of

first link priority issues, these links are often overlooked by search

engines.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Action Items: First Link Priority Issue

Consider code order. Prioritize the main content before the sidebar

and top bar navigation. CSS can be used to control float direction,

from left to right or right to left to make the sidebar navigation load

after the main content. The top bar navigation can be controlled

with absolute positioning.

Handling Navigation in Large

Websites

For large websites (those with hundreds of thousands or millions

pages), website navigation can be a huge challenge. The natural

site navigation within categorized menus generally links to all pages

of the site, and an XML sitemap can help index all pages. However,

the lack of cross-linking between content silos can create distance

between pages.

On a large site, it can be difficult to identify all possible links between

product pages and the corresponding product marketing pages.

Some sections of large sites may not be receiving much of the link

love they need from other pages. Additionally, other issues like the

first link priority and issues with JavaScript could be hard to detect

across millions of pages.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Here are three solutions to these challenges:

1. Delegate to Different Departments

Large companies have proportionately large websites with multiple

employees belonging to different departments. Many departments

may correspond to different sections of the website.

Make sure that everyone involved in maintaining the different

website sections abides by the same SEO principles and practices.

Then, distribute the labor in optimizing navigation across the whole

website.

2. Use Tools or Build Tools

Automation always makes manual processes more scalable. Unless

you have your own proprietary tool, there may not be a single tool to

identify and fix all issues mentioned above.

Crawling tools like Xenu, Screaming Frog, DeepCrawl, or Botify

can analyze your existing links, determine the issues, and provide a

description of the site architecture. If you want to visualize the site

architecture, tools like DynoMapper and PowerMapper can help

achieve this.

Link research tools like Moz’s Open Site Explorer, Ahrefs, Majestic,

Sistrix, LRT, and CognitiveSEO can analyze which pages get the

most backlinks externally then add cross-links from these pages

leading to more important pages of the site. The proprietary tool we

use automates the process of crawling the page and determining

which pages link to one another.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

3. Use a Phased Approach

Large websites don’t always have large teams behind them

to distribute the work of optimizing pages. If there is a lack of

resources, you can create your own tools to ease this process.

If these tools do not provide the help you need, then consider a

phased approach. This entails working on one section at a time with

an optimization schedule. This is a day-by-day process and may take

longer, but relying on metrics like organic search traffic will help you

determine what to optimize first.

Users come first: Your website navigation should satisfy users

first. Then, optimize your navigation for SEO performance. Never

compromise the user experience.

Cross-linking between content silos: Content relevancy

between pages is important for ranking, which comes naturally in

a well-categorized, hierarchical site architecture. However, this can

have limitations when it lacks cross-linking between content silos

where some pages are just too deep or too far away from receiving

a good amount of link juice from other sources.

Blogs to products, products to blogs: Create high-quality

content that is helpful and relevant to your target audience. If these

blog posts help in a product buying decision, then link to the blog

post from the specific product page(s).

7 Key Takeaways

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

Tracking parameters: Avoid using them; use the onClick event

handler on links for tracking purposes. It is always safe to have a

self-referencing canonical tag.

JavaScript links: Avoid using JavaScript to write content and

links. If there is no way around it, there are methods to make it

work.

First link priority: Ideally, main content comes first. Next, is the

sidebar, followed by the top bar. Lastly, handle the footer. Further

testing is needed to determine if this is really still a valid concern,

but it doesn’t hurt to stick to this method.

Huge websites: Thousands to millions of pages are hard to do

all of the above. Delegate to a team, automate tasks by using tools,

or handle the issues one at a time.

HOW TO IMPROVE YOUR WEBSITE NAVIGATION: 7 ESSENTIAL BEST PRACTICES

9

HTTP or HTTPS? Why

You Need a Secure

Site

Chapter 10

Written By

President, JLH Marketing

Jenny Halasz

10

When Google first started

encouraging sites to go to

HTTPS in May 2010, many

webmasters scoffed at the

idea.

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

After all, HTTPS was only for sites that have transactions or which collect

personal information, right?

Then on August 6, 2014, Google announced that they would be

showing a preference for HTTPS sites in search results. This led

SEOs all over the world to declare that HTTPS was now mandatory, and a

ranking factor.

10

Finally, Google amended its advice on May 13, 2015. They stated

that HTTPS was not actually a ranking factor, just that when

it came to certain types of queries, they’d show a preference for

it. HTTPS was a “tiebreaker”. Google doubled down on this on

September 15 of that year.

Webmasters breathed a collective sigh of relief, as their SEOs and

marketing directors stopped pushing HTTPS so hard. After all,

migrating to HTTPS is a lot of work!

It requires that all of the former pages be redirected, that all images

and other linked file types be secure, and back then, it could even

slow down the server response time a bit as that “handshake”

verification took place (this is no longer true).

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

10

Many SEOs Didn’t Believe in

HTTPS at First

Public Wi-Fi networks can insert advertising on

your site if your site is not HTTPS.

As all this was happening, I went on speaking about SEO, always

indicating that I felt HTTPS was not that important unless you were

collecting personal information or credit card numbers through your

site.

But in 2012, I attended a conference where I learned something that

would change the way I felt about HTTPS forever.

Every time I tell people this tidbit, they are surprised to learn about it.

Are you ready for it?

Still not sure why that’s a

big deal? Here’s what my

website looked like back

in 2012. It was not secure:

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

10

Here’s what my

website looked like

that day my opinion

on HTTPS changed

forever.

You see, I was connected to the free Wi-Fi network provided by my

hotel. I saw these ads show up on my website and immediately

went into a tailspin… I could not understand how I could have ads on

my site!

I didn’t use AdSense; I had never added any ad code on my site.

But there it was, right there in the HTML! I dug around in the code,

thinking for sure that I’d been hacked.

Hint: The difference is

the AdSense block in

the lower left corner

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

10

Finally, I called the tech support number on the notepad

by the phone:

“Hello Tech Support for XYZ Hotel Wi-Fi”

“Hi, can you tell me why I’m seeing ads on websites that I

typically never see ads on?”

“Yes ma’am. The hotel uses Google AdSense to defray the cost

of the free Wi-Fi service. The ads are dynamically inserted in

applicable websites.”

I hung up the phone in shock. Really? The network could

change what appeared in the code?

I tested a few other sites. Sure enough, there was my son’s preschool.

With an ad for a Las Vegas hotel in the bottom left corner

– same place the ad on my site had been.

I checked a few others… the local police station… with an ad for a

nearby restaurant.

The nearby mall had an ad for skin care products not sold in any

of the stores at the mall.

That’s when I realized that this had some serious possible

consequences.

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

10

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

10

What if an ad for a steak restaurant started showing up on a site

of vegan recipes? That would be completely off-brand and could

potentially lose them a visitor.

Not to mention the other nefarious things people could potentially

do to an insecure site.

I researched and realized that the protocol was what made this

possible.

Without a public/private key pair (as is provided in HTTPS), an

intermediary could easily intercept and steal or change any

information before it got to its destination (the browser).

You Need to Go to HTTPS

This means that without that secure connection, any network that is

between the source host and the destination host can change what

the destination host gets.

If you don’t understand why that’s important, check out this trace

route (tracert) from my home network to Google.com. Each one of

these entries is a separate “hop” or server.

Without HTTPS, any one of these servers could change what

Google delivered to my browser as a result (Google is HTTPS so that

wouldn’t happen).

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

10

For the safety and security of your visitors, your network, and

yourself, you need to make your site HTTPS.

It really doesn’t matter if your site is just a brochure site. Even if you

don’t collect any emails or have any login screens, you still need to

migrate your site to HTTPS.

TL;DR? More Reasons You Should

Switch to HTTPS

Protect Your Users’ Information. Make sure their data is

protected as it passes through all of those hops to get to you.

Get the Lock Icon in the Browser Window. It looks like this:

You Have to Have it to Implement AMP. AMP technology only

works on a secure server. AMP’s creators designed it that way on

purpose.

Protect Your Brand. There’s a lot more that can be inserted in

websites beyond ads. Think pr0n, pills, and gambling.

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

10

Better Analytics Data. HTTPS connections don’t allow data

from HTTP connections to be sent through HTTPS channels.

If your site is not HTTPS, you can lose referrer data and other

information from secure sites that link to your site.

Many Applications, Third Parties, and Browser Service

Workers Will Not Support HTTP Sites. If your site is not

secure, you will have problems installing, creating and even

using many third-party tools and scripts.

It’s a Tie-Breaker for Google Ranking. All things being

equal, Google will choose to rank sites that are HTTPS before

sites that are HTTP.

Ready to Switch to HTTPS?

We highly recommend this guide: HTTP to HTTPS Migration: The

Ultimate Stress-Free Guide.

HTTP OR HTTPS? WHY YOU NEED A SECURE SITE

10

How to Improve Page

Speed for More Traffic

& Conversions

Chapter 11

Written By

CEO, Spartan Media

Jeremy Knauff

11

Page speed is a critical factor

in digital marketing today.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

It has a significant impact on:

How long visitors stay on your site.

How many of them convert into paying customer.

How much you pay on a CPC basis in paid search.

Where you rank in organic search.

Unfortunately, most websites perform poorly when it comes to page

speed, and that has a direct negative impact on their revenue.

There is an almost infinite number of things we can spend our days

doing as digital marketers, and there’s never enough time to do them all.

As a result, some things get pushed to the back burner.

11

One of the things that seem to get pushed back most often is

optimizing page speed. This is easy to understand because

most people don’t truly comprehend the importance of this often

overlooked detail, so they don’t see the value in investing time and

money to improve it by a few seconds or less.

What may seem like an inconsequential amount of time to some

marketers, including those who focus solely on search engine

optimization, has been proven to be monumental by data from

industry giants all the way down to our own analytics data.

I’ll assume that you’re like me and you want to maximize your results,

and of course, your revenue, right? Then let’s get started in making

your website faster than greased snot! (That’s quite a visual, isn’t it?)

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

1. Ditch the Budget Web Hosting

We’re all trying to save money these days, after all, those

subscriptions to Raven, SEMrush, Moz, and all the other tools we

use on a daily basis add up quickly. It’s almost like having an extra

kid.

One way a lot of people try to save money is by choosing the kind

of cheap shared hosting that crams as many websites as they

can fit onto a server, much like a bunch of clowns piling into a

single car. Performance be damned!

Sure, your website will be available most of the time as it would

with most any web host, but it will load so bloody slowly that your

visitors will leave frustrated without ever converting into buyers.

“But it’s barely noticeable!” these bargain shoppers insist.

Here’s the thing — it might be barely noticeable to you

because it’s your baby and you love it.

But everyone else only wants to get in

and get out of your website as quickly

as possible.

People want to be on your site for just

long enough to do what they came

to do, whether that means to get an

answer, buy a product, or some other

specific objective. If you slow them

down even a little bit, they will be likely to

hate their experience and leave without converting.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

Think about it like this:

Most people love their own kids unconditionally. But someone

else’s kid screaming, throwing things, disrupting their night out at a

restaurant? They hate that kid. It’s the same with your website.

How Much of a Difference Does It Really Make?

According to a study conducted by Amazon, a difference of just

100ms — a unit of time that a human can’t even perceive, was

enough to reduce their sales by 1 percent. Walmart found similar

results.

If that tiny unit of time has that much direct impact on sales, what

kind impact do you think an extra second or more will have?

But it doesn’t stop there because how quickly (or slowly) your

website loads also has an impact on organic search ranking and

pay-per-click costs. In other words, if your website loads slowly, you

should expect your competitors who have invested in this critical

area to eat your lunch.

Bottom line: skip the budget web hosting. If they are selling it like a

commodity (based mainly on price) then they’ll treat their customers

like a commodity too.

There are a lot of web hosts that are optimized for speed,

particularly for WordPress websites, and some of them are priced

similarly to the budget options. So ask around, do some testing, and

invest in a web host that will give you the performance to satisfy

both your visitors and Google.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

Every file needed for a webpage to render and function, such as

HTML, CSS, JavaScript, images, and fonts require a separate HTTP

request. The more requests made, the slower that page will load.

Now if you’re anything like most of the people I talk to, you’re

probably thinking “Oh, I don’t need to worry about that, Jeremy. I

know what I’m doing and I don’t add a bunch of bloated garbage

into my website!”

That may be partially true. You may not add a bunch of bloated

garbage to your website, but for 90 percent+ of the websites that I

encounter — it’s still there anyway.

That bloat isn’t there because the Bloat Fairy snuck it in while you

were sleeping. It’s there because a majority of web designers,

regardless of skill or experience, don’t make page speed a priority.

The sad truth is that most don’t even know how.

Here’s where the problem starts:

Most themes load one or more CSS files and several JavaScript files.

Some, such as Jquery or FontAwesome, are usually loaded remotely

from another server, which dramatically increases the time it takes a

page to load.

This becomes even more problematic when you consider the

additional CSS and JavaScript files added by plugins. It’s easy to

end up with half a dozen or more HTTP requests just from CSS and

JavaScript files alone.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

2. Reduce HTTP Calls

When you factor in all of the images on a page, which

each require a separate HTTP request, it quickly gets out

of hand.

Merge JavaScript files into one file.

Merge CSS files into one file.

Reduce or eliminate plugins that load their own JavaScript and/

or CSS files. In some cases, as with Gravity Forms, you have the

option to disable them from being loaded.

Use sprites for frequently used images.

Use a font like FontAwesome or Ionic Icons instead of image

files wherever possible because then only one file needs to be

loaded.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

3. Include the Trailing Slash

Omitting the trailing slash on links pointing to your website, whether

from external sources (link building efforts) or from within your own

website, has an adverse impact on speed.

Here’s how:

When you visit a URL without the trailing slash, the web server will

look for a file with that name. If it doesn’t find a file with that name,

it will then treat it as a directory and look for the default file in that

directory.

In other words, by omitting the trailing slash, you’re forcing the

server to execute an unnecessary 301 redirect. While it may seem

instantaneous to you, it does take slightly longer, and as we’ve

already established, every little bit adds up.

https://example.com (this is bad)

or

https://example.com/services (this is also bad)

vs

https://example.com/ (this is good)

or

https://example.com/services/ (this is also good)

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

Enabling GZIP compression can significantly reduce the amount of

time it takes to download your HTML, CSS, JavaScript files because

they are downloaded as much smaller compressed files, which are

then decompressed once they get to the browser.

Don’t worry — your visitors won’t have to do anything extra because

all modern browsers support GZIP and automatically process it for

all HTTP requests already.

With browser caching enabled, the elements of a webpage are

stored in your visitors’ browser so the next time they visit your site,

or when they visit another page, their browser can load the page

without having to send another HTTP request to the server for any of

the cached elements.

Once the first page has been loaded and its elements are stored

in the user’s cache, only new elements need to be downloaded on

subsequent pages. This can drastically reduce the number of files

that need to be downloaded during a typical browsing session.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

4. Enable Compression

5. Enable Browser Caching

Minifying your CSS and JavaScript files removes unnecessary

white space and comments to reduce the file size, and as a

result, the time it takes to download them.

Fortunately, this doesn’t have to be a manual process because

there are several tools available online to convert a file into a

smaller, minified version of itself.

There are also several plugins available for WordPress that will

replace the links in your website head for your regular CSS and

JavaScript files with a minified version of them without modifying

your original files, including popular caching plugins such

as:

W3 Total Cache

WP Super Cache

WP Rocket

It may take a bit of effort to get the settings just right because

minification can often break CSS and JavaScript, so once you’ve

minified everything, be sure to test your website thoroughly.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

6. Minify Resources

Your website can appear to the visitor to load more quickly if it’s

coded to prioritize above-the-fold content — in other words, the

content that is visible before a visitor scrolls.

This means ensuring that any elements that appear above the fold

are also as near the beginning of the HTML code so the browser

can download and render them first.

It’s also critical to include any CSS and JavaScript that are required to

render that area inline rather than in an external CSS file.

Because mobile devices with high-quality cameras are common and

modern content management systems such as WordPress makes

it convenient to upload images, many people simply shoot a photo

and upload it without realizing that, often, the image is at least four

times bigger than is necessary.

This slows your website down considerably — especially for mobile

users.

Optimizing the media files on your website has the potential to

improve your page speed tremendously, and doing so is relatively

easy, so it’s a good investment of your time.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

7. Prioritize Above-the-Fold

Content

8. Optimize Media Files

Optimizing Images

Opt for the ideal format. JPG is perfect for photographic images,

while GIF or PNG are best for images with large areas of solid color.

8-bit PNG files are for images without an alpha channel (transparent

background) and 24-bit files are for images with an alpha channel.

Ensure images are properly sized. If an image is displayed at 800

pixels wide on your website, there is no benefit to using a 1600

pixels wide image.

Compress the image file. Aside from being the top image editing

program, Adobe Photoshop has awesome image compression

capabilities and starts at $9.99/month. You can also use free

WordPress plugins – such as WWW Image Optimizer, Imsanity,

and TinyJPG – that automatically compress uploaded images.

Optimizing Video

Choose the ideal format. MP4 is best in most cases because it

produces the smallest file size.

Serve the optimal size (dimensions) based on visitors’ screen size.

Eliminate the audio track if the video is used in the background as a

design element.

Compress the video file. I use Adobe Premiere most of the time, but

Camtasia is a solid choice too.

Reduce the video length.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

Consider uploading videos to YouTube or Vimeo instead of

serving them locally and use their iframe embedding code.

You shouldn’t stop there though because that only scratches the

surface.

To truly optimize the media on your website, you need to serve the

appropriately-sized images based on the screen size rather than

simply resizing them.

There are two ways to handle this, based on the implementation

of an image.

Images within the HTML of your website can be served using src

set, which enables the browser to select, download, and display

the appropriate image based on the screen size of the device a

visitor is using.

Images placed via CSS – typically as background images, can be

served using media queries to select the appropriate image based

on screen size of the device a visitor is using.

Caching enables your web server to store a static copy of your

webpages so they can be delivered more quickly to a visitor’s

browser, while a CDN allows those copies to be distributed to

servers all over the world so that a visitor’s browser can download

them from the server closest to their location. This improves page

speed dramatically.

HOW TO IMPROVE PAGE SPEED FOR MORE TRAFFIC & CONVERSIONS

11

9. Utilize Caching & CDNs

7 Ways a Mobile-First

Index Impacts SEO

Written By

Chapter 12

Search Engine Journal

Roger Montti

12

If you don’t like change, then

the Internet is not for you.

7 WAYS A MOBILE-FIRST INDEX IMPACTS SEO

Google is constantly changing how they’re indexing and ranking sites. It’s

realistic to expect more changes on the way.

I’ve identified seven insights about a mobile-first index and how that

may influence rankings and SEO.

12

7 WAYS A MOBILE-FIRST INDEX IMPACTS SEO

12

It may be inappropriate to generalize what kind of content is best

for a mobile-first index. Every search query is different and how it is

ranked in Google can be different.

Here is a sample of a few kinds of queries:

Long tail queries

Informational queries (what actor starred in...)

Local search queries

Transactional queries

Research queries

“How do I” queries?

Conversational Search

Personal Search

Personal Search & Conversational Search in Mobile

Personal Search and Conversational Search are the latest evolution

in how people search. It is driven by mobile searches.

The way people search has changed because they are searching

on phones. This must be taken into consideration when creating

your search strategy.

1. Mobile-First Informational

Needs Are Changing

7 WAYS A MOBILE-FIRST INDEX IMPACTS SE0

12

Personal Search

According to Google’s page on Personal Searches:

“Over the past two years, we’ve seen an increase in searches

that include highly personal and conversational language—using

words like “me,” “my,” and “I.”

60% + Growth in mobile searches for “__ for me” in the past two

years.

80% + Growth in mobile searches for “__ should I __” in the past

two years.”

According to Google, Personal Searches fall into three

categories:

Solving a problem

Getting things done

Exploring around me

Conversational Search

Conversational search is a reference to the use of natural language

in search queries. This means that users are literally speaking to

their devices and expecting a natural response.

This is another change in how people search that is changing how

we must think of content when creating content.

7 WAYS A MOBILE-FIRST INDEX IMPACTS SEO

12

Many publishers, including Search Engine Journal, have

experienced an increase in traffic by refashioning existing content

to better meet the needs of mobile users.

According to Google’s web page on Conversational Search:

1. Mobile searches for “do I need” have grown over 65%.

For example, “how much do I need to retire,” “what size generator

do I need,” and “how much paint do I need.”

2. Mobile searches for “should I” have grown over 65%.

For example, “what laptop should I buy,” “should I buy a house,”

“what SPF should I use,” and “what should I have for dinner.”

3. Mobile searches starting with “can I” have grown over

85%.

For example, “can I use paypal on amazon,” “can I buy stamps at

walmart,” and “can I buy a seat for my dog on an airplane.”

Mobile Search Trends Drive Content Relevance Trends

The above kinds of queries for both personal and conversational

search are trending upwards and represent a meaningful change

in what people are looking for. Content should adapt to that.

Each kind search query can be answered by a different kind of

web page, with different content length, with different needs for

diagrams, maps, depth, and so on.

7 WAYS A MOBILE-FIRST INDEX IMPACTS SE0

12

One simply cannot generalize and say that Google prefers

short form content because that’s not always what mobile

users prefer.

Thinking in terms of what most mobile users might prefer for a

specific query is a great start.

But the next step involves thinking about the problem that a

specific search query is trying to solve and what the best solution

for most users is going to be.

Then crafting a content-based response that is appropriate for that

situation.

And as you’ll read below, for some queries the most popular

answer might vary according to time. For some queries, a desktop

optimal content might be appropriate.

7 WAYS A MOBILE-FIRST INDEX IMPACTS SEO

12

Identifying the problem users are trying to solve can lead to multiple

answers.

If you look at the SERPs you will see there are different kinds of sites.

Some might be review sites, some might be informational, some

might be educational.

Those differences are indications that there multiple problems

users are trying to solve. What’s helpful is that Google is highly likely

to order the SERPs according to the most popular user intent, the

answer that satisfies the most users.

So if you want to know which kind of answer to give on a page, take

a look at the SERPs and let the SERPs guide you.

Sometimes this means that most users tend to be on mobile and

short-form content works best.

Sometimes it’s fifty/fifty and most users prefer in-depth content or

multiple product choices or fewer product choices.

Don’t be afraid of the mobile index. It’s not changing much.

It’s simply adding an additional layer, to understand which kind

of content satisfies the typical user (mobile, laptop, desktop,

combination) and the user intent.

It’s just an extra step to understanding who the most users are and

from there asking how to satisfy them, that’s all.

2. Satisfy the Most Users

7 WAYS A MOBILE-FIRST INDEX IMPACTS SE0

12

Every search query demands a specific kind of result because the

user intent behind each query is different. Mobile adds an additional

layer of intent to search queries.

In a Think with Google publication about how people use their

devices (PDF), Google stated this:

“The proliferation of devices has changed the way people interact

with the world around them. With more touchpoints than ever

before, it’s critical that marketers have a full understanding of how

people use devices so that they can be here and be useful for their

customers in the moments that matter.”

3. Time Influences Observed

User Intent

7 WAYS A MOBILE-FIRST INDEX IMPACTS SEO

12

Time plays a role in how the user intent changes.

The time of day that a query is made can influence what device that

user is using, which in turn says something about that users needs

in terms of speed, convenience, and information needs.

Google’s research from the above-cited document states this:

“Mobile leads in the morning, but computers become dominant

around 8 a.m. when people might start their workday. Mobile takes

the lead again in the late afternoon when people might be on the

go, and continues to increase into the evening, spiking around

primetime viewing hours.”

This is what I mean when I say that Google’s mobile index is

introducing a new layer of what it means to be relevant. It’s not about

your on-page keywords being relevant to what a user is typing.

A new consideration is about how your web page is relevant to

someone at a certain time of day on a certain device and how you’re

going to solve the most popular information need at that time of day.

Google’s March 2018 official mobile-first announcement

stated it like this:

“We may show content to users that’s not mobile-friendly or that

is slow loading if our many other signals determine it is the most

relevant content to show.”

What signals is Google looking at? Obviously, the device itself could

be a signal.

But also, according to Google, time of day might be a signal because

not only does device usage fluctuate during the day but the intent

does too.

7 WAYS A MOBILE-FIRST INDEX IMPACTS SE0

12

Google’s focus on user intent 100 percent changes what the phrase

“relevant content” means, especially in a mobile-first index.

People on different devices search for different things. It’s not that

the mobile index itself is changing what is going to be ranked.

The user intent for search queries is constantly changing,

sometimes in response to Google’s ability to better understand

what that intent is.

Some of those core algorithm updates could be changes related to

how Google understands what satisfies users.

You know how SEOs are worrying about click-through data? They

are missing an important metric. CTR is not the only measurement

tool search engines have.

Do you think CTR 100 percent tells what’s going on in a mobilefirst

index? How can Google understand if a SERP solved a user’s

problem if the user does not even click through?

That’s where a metric similar to Viewport Time comes in. Search

engines have been using variations of Viewport Time to understand

mobile users.

4. Defining Relevance in a

Mobile-First Index

Yet the SEO industry is still wringing its hands about CTR. Ever feel

like a piece of the ranking puzzle is missing? This is one of

those pieces.

Google’s understanding of what satisfies users is constantly

improving. And that impacts the rankings. How we provide the best

experience for those queries should change, too.

An important way those solutions have changed involves

understanding the demographics of who is using a specific kind of

device.

What does it mean when someone asks a question on one device

versus another device?

One answer is that the age group might influence who is asking a

certain question on a certain device.

For example, Google shared the following insights about mobile

and desktop users (PDF). Searchers in the Beauty and Health

niche search for different kinds of things according to device.

Examples of top beauty and health queries on mobile devices are

for topics related to tattoos and nail salons.

Examples of Beauty and Health desktop queries indicate an older

user because they’re searching for stores like Saks and beauty

products such as anti-aging creams.

7 WAYS A MOBILE-FIRST INDEX IMPACTS SEO

12

7 WAYS A MOBILE-FIRST INDEX IMPACTS SE0

12

It’s naïve to worry about whether you have enough synonyms on

your page. That’s not what relevance is about.

Relevance is not about keyword synonyms. Relevance is often

about problem-solving at certain times of day and within specific

devices to specific age groups.

You can’t solve that by salting your web page with synonyms.

7 WAYS A MOBILE-FIRST INDEX IMPACTS SEO

12

An important quality of the mobile-first index is

convenience when satisfying a user intent.

Does the user intent behind the search query demand a quick

answer or a shorter answer?

Does the web page make it hard to find the answer?

Does the page enable comparison between different

products?

Now answer those questions by adding the phrase, on mobile,

on a tablet, on a desktop and so on.

5. Mobile First Is Not About

User-Friendliness

7 WAYS A MOBILE-FIRST INDEX IMPACTS SE0

12

Google can know if a user understands your content. Users vote

with their click and viewport time data and quality raters create

another layer of data about certain queries.

With enough data Google can predict it what a user might find

useful. This is where machine learning comes in.

Here’s what Google says about machine learning in the

context of User Experience (UX):

“Machine learning is the science of making predictions based

on patterns and relationships that’ve been automatically

discovered in data.”

If content that is difficult to read is a turn-off, that may be

reflected in what sites are ranked and what sites are not.

If the topic is complex and a complex answer solves the problem

then that might be judged the best answer.

I know we’re talking about Google but it’s useful to understand

the state of the art of search in general.

6. Would a Visitor Understand

Your Content?

7 WAYS A MOBILE-FIRST INDEX IMPACTS SEO

12

Microsoft published a fascinating study about teaching a machine

to predict what a user will find interesting. The paper is titled,

Predicting Interesting Things in Text.

This research focused on understanding what made content

interesting and what caused users to keep clicking to another page.

In other words, it was about training a machine to understand what

satisfies users.

Here’s a synopsis:

“We propose models of “interestingness”, which aim to predict the

level of interest a user has in the various text spans in a document.

We obtain naturally occurring interest signals by observing user

browsing behavior in clicks from one page to another. We cast the

problem of predicting interestingness as a discriminative learning

problem over this data.

We train and test our models on millions of real world transitions

between Wikipedia documents as observed from web browser

session logs. On the task of predicting which spans are of most

interest to users, we show significant improvement over various

baselines and highlight the value of our latent semantic model.”

In general, I find good results with content that can be appreciated

by the widest variety of people.

7 WAYS A MOBILE-FIRST INDEX IMPACTS SE0

12

This isn’t strictly a mobile-first consideration but it is increasingly

important in an Internet where so people of diverse backgrounds

are accessing a site with multiple intents multiple kinds of devices.

Achieving universal popularity becomes increasingly difficult so it

may be advantageous to appeal to the broadest array of people in

a mobile-first index.

Looked at a certain way, it could be said that Google’s desire to

show users what they want to see has remained consistent.

What has changed is the users’ age, what they desire, when they

desire it and what device they desire it on. So the intent of Google’s

algorithm likely remains the same.

The mobile-first index can be seen as a logical response to how

users have changed. It’s backwards to think of it as Google forcing

web publishers to adapt to Google.

What’s really happening is that web publishers must adapt to how

their users have changed.

Ultimately that is the best way to think of the mobile-first index. Not

as a response to what Google wants but to approach the problem

as a response to the evolving needs of the user.

7. Google’s Algo Intent Hasn’t

Changed

The Complete Guide

to Mastering Duplicate

Content Issues

Written By

Chapter 13

VP Search and Advertising, The Karcher Group

Stoney G deGeyter

13

In the SEO arena of website

architecture, there is little

doubt that eliminating

duplicate content can be one

of the hardest fought battles.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

Too many content management systems and piss-poor developers build

sites that work great for displaying content but have little consideration

for how that content functions from a search-engine-friendly perspective.

And that often leaves damaging duplicate content dilemmas for the SEO

to deal with.

13

There are two kinds of duplicate content, and both can

be a problem:

Onsite duplication is when the same content is duplicated on two

or more unique URLs of your site. Typically, this is something that

can be controlled by the site admin and web development team.

Offsite duplication is when two or more websites publish the

exact same pieces of content. This is something that often

cannot be controlled directly but relies on working with thirdparties

and the owners of the offending websites.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

Why Is Duplicate Content a

Problem?

The best way to explain why duplicate content is bad is to first tell

you why unique content is good.

Unique content is one of the best ways to set yourself apart from

other websites. When the content on your website is yours and

yours alone, you stand out. You have something no one else has.

On the other hand, when you use the same content to describe your

products or services or have content republished on other sites, you

lose the advantage of being unique.

Or, in the case of onsite duplicate content, individual pages lose the

advantage of being unique.

Look at the illustration below. If A represents content

that is duplicated on two pages, and B

through Q represents pages linking to that

content, the duplication causes a split

the link value being passed.

Now imagine if pages B-Q all linked to

only on page A. Instead of splitting the

value each link provides, all the value

would go to a single URL instead, which

increases the chances of that content

ranking in search.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

Whether onsite or offsite, all duplicate content competes against

itself. Each version may attract eyeballs and links, but none will

receive the full value it would get if it were the sole and unique

version.

However, when valuable and unique content can be found on no

more than a single URL anywhere on the web, that URL has the

best chance of being found based on it being the sole collector of

authority signals for that content.

Now, having that understanding, let’s look at the problems and

solutions for duplicate content.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

Offsite Duplicate Content

Offsite duplication has three primary sources:

Third-party content you have republished on your own site.

Typically, this is in the form of generic product descriptions

provided by the manufacturer.

Your content that has been republished on third-party sites with

your approval. This is usually in the form of article distribution or

perhaps reverse article distribution.

Content that someone has stolen from your site and republished

without your approval. This is where the content scrapers and

thieves become a nuisance.

Let’s look at each.

Content Scrapers & Thieves

Content scrapers are one of the biggest offenders in duplicate

content creation. Spammers and other nefarious perpetrators

build tools that grab content from other websites and then publish

it on their own.

For the most part, these sites are trying to use your content to

generate traffic to their own site in order to get people to click their

ads. (Yeah, I’m looking at you, Google!)

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

Unfortunately, there isn’t much you can do about this other than to

submit a copyright infringement report to Google in hopes that

it will be removed from their search index. Though, in some cases,

submitting these reports can be a full-time job.

Another way of dealing with this content is to ignore it, hoping

Google can tell the difference between a quality site (yours) and

the site the scraped content is on. This is hit and miss as I’ve seen

scraped content rank higher than the originating source.

What you can do to combat the effects of scraped content is to

utilize absolute links (full URL) within the content for any links

pointing back to your site. Those stealing content generally aren’t

in the business of cleaning it up so, at the very least, visitors can

follow that back to you.

You can also try adding a canonical tag back to the source page (a

good practice regardless). If the scrapers grab any of this code, the

canonical tag will at least provide a signal for Google to recognize

you as the originator.

Article Distribution

Several years ago, it seemed like every SEO was republishing their

content on “ezines” as a link building tactic. When Google cracked

down on content quality and link schemes, republishing fell by the

wayside.

But with the right focus, it can be a solid marketing strategy. Notice,

I said “marketing” rather than “SEO” strategy.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

For the most part, any time you’re publishing content on other

websites, they want the unique rights to that content.

Why? Because they don’t want multiple versions of that content on

the web devaluing what the publisher has to offer.

But as Google has gotten better about assigning rights to the

content originator (better, but not perfect), many publishers are

allowing content to be reused on the author’s personal sites as well.

Does this create a duplicate content problem? In a small way, it can,

because there are still two versions of the content out there, each

potentially generating links.

But in the end, if the number of duplicate versions is limited and

controlled, the impact will be limited as well. In fact, the primary

downside lands on the author rather than the secondary publisher.

The first published version of the content will generally be credited

as the canonical version. In all but a few cases, these publishers

will get more value from the content over the author’s website that

republishes it.

Generic Product Descriptions

Some of the most common forms of duplicated content comes from

product descriptions that are reused by each (and almost every)

seller.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

A lot of online retailers sell the exact same products as thousands

of other stores. In most cases, the product descriptions are

provided by the manufacturer, which is then uploaded into each

site’s database and presented on their product pages.

While the layout of the pages will be different, the bulk of the

product page content (product descriptions) will be identical.

Now multiply that across millions of different products and

hundreds of thousands of websites selling those products, and

you can wind up with a lot of content that is, to put it mildly, not

unique.

How does a search engine differentiate between one or another

when a search is performed?

On a purely content-analysis level, it can’t. Which means the

search engine must look at other signals to decide which one

should rank.

One of these signals is links. Get more links and you can win the

bland content sweepstakes.

But if you’re up against a more powerful competitor, you may

have a long battle to fight before you can catch them in the link

building department. Which brings you back to looking for another

competitive advantage.

The best way to achieve that is by taking the extra effort to write

unique descriptions for each product. Depending on the number

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

Any page with unique content is going to automatically have an

inherent advantage over similar but duplicate content. That may or

may not be enough to outrank your competition, but it surely is the

baseline for standing out to not just Google, but your customers as

well.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

of products you offer, this could end up being quite a challenge, but

in the end, it’ll be well worth it.

Take a look at the illustration below. If all the gray pages represent

the same product with the same product descriptions, the yellow

represents the same product with a unique description.

If you were Google, which one would you want to rank higher?

Onsite Duplicate Content

Technically, Google treats all duplicate content the same, so onsite

duplicate content is really no different than offsite.

But onsite is less forgivable because this is one type of duplication

that you can actually control. It’s shooting your SEO efforts in the

proverbial foot.

Onsite duplicate content generally stems from bad site architecture.

Or, more likely, bad website development!

A strong site architecture is the foundation for a strong website.

When developers don’t follow search-friendly best practices, you can

wind up losing valuable opportunity to get your content to rank due

to this self-competition.

There are some who argue against the need for good architecture,

citing Google propaganda about how Google can “figure it out.” The

problem with that is that it relies on Google figuring things out.

Yes, Google can determine that some duplicate content should be

considered one and the same, and the algorithms can take this into

account when analyzing your site, but that’s no guarantee they will.

Or another way to look at it is that just because you know someone

smart doesn’t necessarily mean they’ll be able to protect you from

your own stupidity! If you leave things to Google and Google fails,

you’re screwed.

Now, let’s dive into some common onsite duplicate content

problems and solutions.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

The Problem: Product Categorization Duplication

Far too many ecommerce sites suffer from this kind of

duplication. This is frequently caused by content management

systems that allow you to organize products by category, where a

single product can be tagged in multiple categories.

That in itself isn’t bad (and can be great for the visitor), however

in doing so, the system generates a unique URL for each

category in which a single product shows up in.

Let’s say you’re on a home repair site and you’re looking for a

book on installing bathroom flooring.

You might find the book you’re looking for by following

any of these navigation paths:

Home > flooring > bathroom > books

Home > bathroom > books > flooring

Home > books > flooring > bathroom

Each of these is a viable navigation path, but the problem

arises when a unique URL is generated for each path:

https://www.myfakesite.com/flooring/bathroom/books/fakebook-

by-fake-author

https://www.myfakesite.com/bathroom/books/flooring/fakebook-

by-fake-author

https://www.myfakesite.com/books/flooring/bathroom/fakebook-

by-fake-author

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

I’ve seen sites like this create up to ten URLs for every single

product turning a 5k product website into a site with 45k duplicate

pages. That is a problem.

If our example product above generated ten links, those links would

end up being split three ways. Whereas, if a competitor’s page for

the same product got the same ten links, but to only a single URL,

which URL is likely to perform better in search? The competitor’s!

Not only that, but search engines limit their crawl bandwidth so they

can spend it on indexing unique and valuable content.

When your site has that many duplicate pages, there is a strong

chance the engine will stop crawling before it even gets a fraction of

your unique content indexed.

This means hundreds of valuable pages won’t be available in search

results and those that are indexed are duplicates competing against

each other.

The Solution: Master URL Categorizations

One fix to this problem is to only tag products for a single category

rather than multiples. That solves the duplication issue, but it’s not

necessarily the best solution for the shoppers since it eliminates

the other navigation options for finding the product(s) they want. So,

scratch that one off the list.

Another option is to remove any type of categorization from the

URLs altogether.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

This way, no matter the navigation path used to find the

product, the product URL itself is always the same, and

might look something like this:

https://www.myfakesite.com/products/fake-book-by-fake-author

This fixes the duplication without changing how the visitor is able to

navigate to the products. The downside to this method is that you

lose the category keywords in the URL. While this provides a small

benefit to the totality of SEO, every little bit can help.

If you want to take your solution to the next level, getting the most

optimization value possible while keeping the user experience at the

same time, build an option that allows each product to be assigned

to a “master” category, in addition to others.

When a master category is in play, the product can continue to be

found through the multiple navigation paths, but the product page is

accessed by a single URL that utilizes the master category.

That might make the URL look something like this:

https://www.myfakesite.com/flooring/fake-book-by-fake-author OR

https://www.myfakesite.com/bathroom/fake-book-by-fake-author OR

https://www.myfakesite.com/books/fake-book-by-fake-author

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

This latter solution is the best overall, though it does take some

additional programming. However, there is one more relatively

easy “solution” to implement, but I only consider it a band-aid until

a real solution can be implemented.

Band-Aid Solution: Canonical Tags

Because the master-categorization option isn’t always available

to out of the box CMS or ecommerce solutions, there is an

alternative option that will “help” solve the duplicate content

problem.

This involves preventing search engines from indexing all noncanonical

URLs. While this can keep duplicate pages out of

the search index, it doesn’t fix the issue of splitting the page’s

authority. Any link value sent to a non-indexable URL will be lost.

The better band-aid solution is to utilize canonical tags. This is

similar to selecting a master category but generally requires little, if

any, additional programming.

You simply add a field for each product that allows you to assign a

canonical URL, which is just a fancy way of saying, “the URL you

want to show up in search.”

The canonical tag looks like this:

<link rel=“canonical” href=“https://www.myfakesite.com/books/

fake-book-by-fake-author” />

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

Despite the URL the visitor is on, the behind-the-scenes canonical

tag on each duplicate URL would point to a single URL.

In theory, this tells the search engines not to index the non-canonical

URLs and to assign all other value metrics over to the canonical

version. This works most of the time, but in reality, the search

engines only use the canonical tag as a “signal.” They will then

choose to apply or ignore it as they see fit.

You may or may not get all link authority passed to the correct page,

and you may or may not keep non-canonical pages out of the index.

I always recommend implementing a canonical tag, but because it’s

unreliable, consider it a placeholder until a more official solution can

be implemented.

The Problem: Redundant URL Duplication

One of the most basic website architectural issues revolves around

how pages are accessed in the browser.

By default, almost every

page of your site can

be accessed using a

slightly different URL.

If left unchecked,

each URL leads to

the exact same page

with the exact same

content.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

Considering the home page alone, it can likely be accessed

using four different URLs:

http://site.com

http://www.site.com

https://site.com

https://www.site.com

And when dealing with internal pages, you can get an

additional version of each URL by adding a trailing slash:

http://site.com/page

http://site.com/page/

http://www.site.com/page

http://www.site.com/page/

Etc.

That’s up to eight alternate URLs for each page! Of course, Google

should know that all these URLs should be treated as one, but

which one?

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

The Solution: 301 Redirects & Internal Link Consistency

Aside from the canonical tag, which I addressed above, the

solution here is to ensure you have all alternate versions of the

URLs redirecting to the canonical URL.

Keep in mind, this isn’t just a home page issue. The same issue

applies to every one of your site URLs. Therefore, the redirects

implemented should be global.

Be sure to force each redirect to the canonical version. For

instance, if the canonical URL is https://www.site.com, each

redirect should point there.

Many make the mistake of adding additional redirect

hops that might look like this:

Site.com > https://site.com > https://www.site.com

Site.com > www.site.com > https://www.site.com

Instead, the redirects should look like this:

http://site.com > https://www.site.com/

http://www.site.com > https://www.site.com/

https://site.com > https://www.site.com/

https://www.site.com > https://www.site.com/

http://site.com/ > https://www.site.com/

http://www.site.com/ > https://www.site.com/

https://site.com/ > https://www.site.com/

By reducing the number of redirect hops you speed up page load,

reduce server bandwidth, and have less that can go wrong along

the way.

Finally, you’ll need to make sure all internal links in the site point to

the canonical version as well.

While the redirect should solve the duplicate problem, redirects can

fail if something goes wrong on the server or implementation side

of things.

If that happens, even temporarily, having only the canonical pages

linked internally can help prevent a sudden surge of duplicate

content issues from popping up.

The Problem: URL Parameters & Query Strings

Years ago, the usage of session IDs created a major duplicate

content problem for SEOs.

Today’s technology, however, has made session IDs all but

obsolete, but another problem has arisen that is just as bad, if not

worse: URL parameters.

Parameters are used to pull fresh content from the server, usually

based on one or more filter or selections being made.

The two examples below show alternate URLs for a single URL:

site.com/shirts/.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

The first shows the shirts filtered by color, size, and style,

the second URL shows shirts sorted by price, then a

certain number of products to show per page:

Site.com/shirts/?color=red&size=small&style=long_sleeve

Site.com/shirts/?sort=price&display=12

Based on these filters alone, there are three viable URLs that search

engines can find.

But the order of these parameters can change based on the

order in which they were chosen, which means you might

get several more accessible URLs like this:

Site.com/shirts/?size=small&color=red&style=long_sleeve

Site.com/shirts/?size=small&style=long_sleeve&color=red

Site.com/shirts/?display=12&sort=price

And this:

Site.com/shirts/?size=small&color=red&style=long_

sleeve&display=12&sort=price

Site.com/shirts/?display=12&size=small&color=red&sort=price

Site.com/

shirts/?size=small&display=12&sort=price&color=red&style=long_

sleeve

Etc.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

You can see that this can produce a lot of URLs, most of which

will not pull any type of unique content. Of the parameters above,

the only one you might want to write sales content for is the style.

The rest, not so much.

The Solution: Parameters for Filters, Not Legitimate

Landing Pages

Strategically planning your navigation and URL structure is critical

for getting out ahead of the duplicate content problems.

Part of that process includes understanding the difference

between having a legitimate landing page and a page that allows

visitors to filter results. And then be sure to treat these accordingly

when developing the URLs for them.

Landing page (and canonical) URLs should look like this:

Site.com/shirts/long-sleeve/

Site.com/shirts/v-neck/

Site.com/shirts/collared/

And the filtered results URLs would look something like

this:

Site.com/shirts/long-sleeve/?size=small&color=red&display=12&

sort=price

Site.com/shirts/v-neck/?color=red

Site.com/shirts/

collared/?size=small&display=12&sort=price&color=red

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

With your URLs built correctly, you can do two things:

Add the correct canonical tag (everything before the “?” in the

URL).

Go into Google Search Console and tell Google to ignore all such

parameters.

If you consistently use parameters only for filtering and sorting

content, you won’t have to worry about accidentally telling Google

not to crawl a valuable parameter… because none of them are.

But because the canonical tag is only a signal, you must complete

step two for best results. And remember this only affects Google.

You have to do the same with Bing.

Pro Developer Tip: Search engines typically ignore everything to

the right of a pound “#” symbol in the URL.

If you program that into every URL prior to any parameter,

you won’t have to worry about the canonical being only a

band-aid solution:

Site.com/shirts/long-sleeve/#?size=small&color=red&display=12

&sort=price

Site.com/shirts/v-neck/#?color=red

Site.com/shirts/

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

collared/#?size=small&display=12&sort=price&color=red

If any search engine were to access the URLs above, they would

only index the canonical part of the URL and ignore the rest.

The Problem: Ad Landing Page & A/B Test Duplication

It’s not uncommon for marketers to develop numerous versions of

similar content, either as a landing page for ads, or A/B/multivariate

testing purposes.

This can often get you some great data and feedback, but if those

pages are open for search engines to spider and index, it can

create duplicate content problems.

The Solution: NoIndex

Rather than use a canonical tag to point back to the master page,

the better solution here is to add a noindex meta tag to each page

to keep them out of the search engines’ index altogether.

Generally, these pages tend to be orphans, not having any direct

links to them from inside the site. But that won’t always keep search

engines from finding them.

The canonical tag is designed to transfer page value and authority

to the primary page, but since these pages should not be collecting

any value, keeping them out of the index is preferred.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

When Duplicate Content Isn’t

(Much Of) a Problem

One of the most common SEO myths is that there is a duplicate

content penalty.

There isn’t. At least no more than there is a penalty for not putting

gas in your car and letting it run empty.

Google may not be actively penalizing duplicate content, but that

doesn’t mean there are not natural consequences that occur

because of it.

Without the threat of penalty, that gives marketers a little more

flexibility in deciding which consequences they are willing to live

with.

While I would argue that you should aggressively eliminate (not just

band-aid over) all on-site duplicate content, offsite duplication may

actually create more value than consequences.

Getting valuable content republished off-site can help you build

brand recognition in a way that publishing it on your own can’t.

That’s because many offsite publishers have a bigger audience and

a vastly larger social reach.

Your content, published on your own site may reach thousands of

eyeballs, but published offsite it might reach hundreds of thousands.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

Many publishers do expect to maintain exclusive rights to the

content they publish, but some allow you to repurpose it on your

own site after a short waiting period. This allows you to get the

additional exposure while also having the opportunity to build up

your own audience by republishing your content on your site at a

later date.

But this type of article distribution needs to be limited in order

to be effective for anyone. If you’re shooting your content out to

hundreds of other sites to be republished, the value of that content

diminishes exponentially.

And typically, it does little to reinforce your brand because the

sites willing to publish mass duplicated content are of little value to

begin with.

In any case, weigh the pros and cons of your content being

published in multiple places.

If duplication with a lot of branding outweighs the smaller authority

value you’d get with unique content on your own site, then, by all

means, pursue a measured republishing strategy.

But the keyword there is measured. What you don’t want to be is

the site that only has duplicate content. At that point, you begin to

undercut the value you’re trying to create for your brand.

THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

By understanding the problems, solutions and, in some cases,

value, of duplicate content, you can begin the process of

eliminating the duplication you don’t want and pursuing the

duplication you do.

In the end, you want to build a site that is known for strong,

unique content, and then use that content to get the highest value

possible. THE COMPLETE GUIDE TO MASTERING DUPLICATE CONTENT ISSUES

13

A Technical SEO Guide

to Redirects

Chapter 14

Written By

Lead Developer, Search Engine Journal

Vahan Petrosyan

14

Websites change structure,

delete pages and often move

from one domain to another. A

TECHNICAL SEO GUIDE TO REDIRECTS

Handling redirects correctly is crucial in order to avoid losing rankings

and help search engines understand the changes you have done.

Redirects have a status code starting with number three (i.e., 3XX).

There are 100 different possible status codes but only a few are

implemented to carry certain information.

In this guide, we will cover 3XX redirects relevant to SEO.

14

301: Moved Permanently

307: Temporary Redirect

This well-known redirect indicates to a client* that the resource was

changed to another location and that it should use the new URL for

future requests. When search engines see a 301 redirect, they pass

the old page’s ranking to the new one.

Before making a change, you need to be careful when deciding to

use a 301 redirect. This is because if you change your mind later

and decide to remove the 301 redirect, your old URL may not rank

anymore.

Even if you swap the redirects, it will not help you get the old page

back to its previous ranking position. So the main thing to remember

is that there’s no way to undo a 301 redirect.

(*For beginners who may get confused with generic name client is

used instead of browser since not only browsers are able to browse

URLs but also search engine bots which are not browsers.)

In HTTP 1.1, a 301 redirect means the resource is temporarily moved

and the client should use the original resource’s URL for future

requests.

For SEO, this means the client should follow a redirect but search

engines should not update their links in the SERPs to the new,

temporary page.

In a 307 redirect, PageRank is not passed from the original resource

to the new one – contrary to a 301 redirect.

A TECHNICAL SEO GUIDE TO REDIRECT

14

302: Found

302 vs. 307

This means that the resource a client is looking for was found on

another URL in the HTTP 1.1 version but was temporarily moved in

HTTP 1.0.

In almost all cases, 302 and 307 redirects will be treated the same.

But a 302 status code doesn’t necessarily mean the client must

follow a redirect and it is not considered an error if it decides to stay

there.

Modern clients will most likely follow the new destination but some

old clients may incorrectly stay on the same URL.

Contrary to a 302 status code, the 307 status code guarantees that

the request method will not be changed. For instance, the GET

request must continue to GET and POST to POST.

With a 302 status code, some old or buggy clients may change the

method which may cause unexpected behavior.

For temporary redirects, you can use either 302 or 307 – but I do

prefer 307.

For routine redirect tasks, 301 (permanent redirect) and 307

(temporarily redirect) status codes should be used depending on

what type of change you are implementing on your website. On both

cases, the syntax of redirects doesn’t change.

A TECHNICAL SEO GUIDE TO REDIRECT

14

You may handle redirect via server config files .htaccess on Apache,

example.conf file on Ngix, or via plugins if you are using WordPress.

In all instances, they have the same syntax for writing redirect rules.

They differ only with commands used in configuration files.

For example, redirect on Apache will look like this:

Options +FollowSymlinks

RewriteEngine on

RedirectMatch 301 ^/oldfolder/ /newfolder/

( you can read about symlinks here ). and on Ngix servers like

rewrite ^/oldfolder/ /newfolder/ permanent;

The commands used to tell servers status code of redirect

and the action command differ. For instance:

Servers status code of redirect: “301” vs. “permanent”

Action command: “RedirectMatch” vs. “rewrite”.

But the syntax of the redirect ( ^/oldfolder/ /newfolder/ ) is the same

for both.

On Apache, make sure on your server mod_rewrite and mod_alias

modules (which are responsible for handling redirects) are enabled.

Since the most widely spread server types is Apache, here are

examples for .htaccess apache files. Make sure that the .htaccess

file has these two lines

Options +FollowSymlinks

RewriteEngine on

A TECHNICAL SEO GUIDE TO REDIRECT

14

above the redirect rules and put the rules below them.

For understanding the examples below you may refer table below

on RegExp basics.

Redirect Single URL

The most common and widely used type of redirect that is used

when deleting pages or changing page URLs. For instance, say

you changed URL from /old-page/ to /new-page/. The redirect rule

would be:

RewriteRule ^old-page(/?|/.*)$ /new-page/ [R=301,L]

OR

RedirectMatch 301 ^/old-page(/?|/.*)$ /new-page/

A TECHNICAL SEO GUIDE TO REDIRECT

14

The only difference between the two methods is that the first one

uses Apache mod_rewrite module and the second one uses mod_

alias. It can be done using both methods.

Regular expression “^” means URL must start with “/old-page” while

(/?|/.*)$ indicates that anything that follows “/old-page/” with slash

“/” or without exact match must be redirected to /new-page/.

We could also use (.*) ie. ^/old-page(.*), but the problem is, if you

have another page with a similar URL like /old-page-other/, it will

also be redirected when we only want to redirect /old-page/.

The following URLs will match and directed to new page

It will redirect any variation of page URL to new one.

A TECHNICAL SEO GUIDE TO REDIRECT

14

Redirect All Except

Directory Change

Let’s say we have bunch of URLs like /category/old-subcategory-1/, /

category/old-subcategory-2/, /category/final-subcategory/ and want

to merge all subcategories into /category/final-subcategory/. We

need here “all except” rule

RewriteCond %{REQUEST_URI} !/category/final-subcategory/

RewriteCond %{REQUEST_FILENAME} !-f

RewriteRule ^(category/). /category/final-subcategory/ [R=301,L]

Here, we want to redirect all under /category/ on the fifth line except

if it is /category/final-subcategory/ on the fourth line. We also have

“!-f” rule on the fourth line which means to ignore any file like images,

CSS or javascript files.

Otherwise, if we have some assets like “/category/image.jpg” it will

be also redirected to “/final-subcategory/” and cause a page break.

In case you did a category restructuring and want to move

everything under the old directory to the new one, you can use the

rule below.

RewriteRule ^old-directory$ /new-directory/ [R=301,NC,L]

RewriteRule ^old-directory/(.*)$ /new-directory/$1 [R=301,NC,L]

I used $1 in the target to tell the server that it should remember

everything in the URL that follows /old-directory/ (i.e., /old-directory/

subdirectory/) and pass it (i.e., “/subdirectory/” ) onto the destination.

As a result, it will be redirected to /new-directory/subdirectory/.

A TECHNICAL SEO GUIDE TO REDIRECT

14

I used two rules: one case with no trailing slash at the end and the

other one with a trailing slash.

I could combine them into one rule using (/?|.*)$ RegExp at the end,

but it would cause problems and add “//” slash to the end of URL

when the requested URL with no trailing slash has a query string

(i.e., “/old-directory?utm_source=facebook” would be redirected to

“/new-directory//?utm_source=facebook”).

Remove a Word from URL

Let’s say you have 100 URLs in your website with city name

“chicago” and want to remove it.

Example, for the URL http://yourwebiste.com/example-chicagoevent/,

the redirect rule would be:

RewriteRule ^(.*)-chicago-(.*) http://%{SERVER_NAME}/$1-$2

[NC,R=301,L]

If the example URL is in the form http:// yourwebiste.com/example/

chicago/event/, then redirect will be:

RewriteRule ^(.*)/chicago/(.*) http://%{SERVER_NAME}/$1/$2

[NC,R=301,L]

A TECHNICAL SEO GUIDE TO REDIRECT

14

Canonicalization

Having canonical URLs is the most important part of SEO.

If it is missing, you might endanger your website with duplicate

content issues because search engines treat URLs with “www” and

“non-www” versions as different pages with the same content.

Therefore, it is mandatory to make sure you run website only with

only one version you choose.

If you want to run your website with “www” version, use this rule:

RewriteCond %{HTTP_HOST} ^yourwebsite\.com [NC]

RewriteRule ^(.*)$ http://www.yourwebsite.com/$1 [L,R=301]

For a “non-www” version:

RewriteCond %{HTTP_HOST} ^www\.yourwebsite\.com [NC]

RewriteRule ^(.*)$ http://yourwebsite.com/$1 [L,R=301]

Trailing slash is also part of canonicalization since URLs with a

slash at the end or without are also treated differently.

RewriteCond %{REQUEST_FILENAME} !-f

RewriteRule ^(.*[^/])$ /$1/ [L,R=301]

This will make sure /example-page is redirected to /example-page/.

You may choose to remove the slash instead of adding then you will

need the other rule below:

RewriteCond %{REQUEST_FILENAME} !-d

RewriteRule ^(.*)/$ /$1 [L,R=301]

A TECHNICAL SEO GUIDE TO REDIRECT

14

HTTP to HTTPS Redirect

Redirect from Old Domain to New

After Google’s initiative to encourage website owners to use SSL,

migrating to HTTPS is one of the commonly used redirects that

almost every website has.

The rewrite rule below can be used to force HTTPS on every

website.

RewriteCond %{HTTP_HOST} ^yourwebsite\.com [NC,OR]

RewriteCond %{HTTP_HOST} ^www\.yourwebsite\.com [NC]

RewriteRule ^(.*)$ https://www.yourwebsite.com/$1 [L,R=301,NC]

Basically, you can combine www or non-www version redirect into

one HTTPS redirect rule using this.

This is also one of the most used redirects when you decide to

do rebranding and you need to change domain. The rule below

redirects old-domain.com to new-domain.com

RewriteCond %{HTTP_HOST} ^old-domain.com$ [OR]

RewriteCond %{HTTP_HOST} ^www.old-domain.com$

RewriteRule (.*)$ http://www.new-domain.com/$1 [R=301,L]

It uses two cases: one with “www” version of URLs and another

“non-www” because any page for historical reasons may have

incoming links to both versions.

A TECHNICAL SEO GUIDE TO REDIRECT

14

Most site owners use WordPress and may not need to use

.htaccess file for redirects but use plugin instead. Handling redirects

by using plugins may be a little different from what we discussed

above and you may need to read their documentation in order to be

able to handle RegExp correctly for specific plugin.

From existing ones I would recommend free plugin called

Redirection which has many parameters to control redirect rules

and many useful docs.

Redirect Bad Practices

1. Redirecting All 404 Broken URLs to the Home Page

This case often happens when you are lazy to investigate all of your

404 URLs and map them to the appropriate landing page.

According to Google, they are still all treated as 404s.

If you have too many pages like this, you should consider creating

beautiful 404 pages and engage users to browse further or find

something other than what they were looking for by displaying a

search option.

A TECHNICAL SEO GUIDE TO REDIRECT

14

It is strongly recommended by Google that redirected page content

should be equivalent to the old page. Otherwise, such redirect may

be considered as soft 404 and you will lose the rank of that page.

2. Wrong Mobile Page Specific Redirects

If you have different URLs for desktop and mobile websites (i.e.,

“yoursite.com” for desktop and “m.yoursite.com” for mobile), you

should make sure to redirect users to the appropriate page of the

mobile version.

Correct: “yoursite.com/sport/” to “m.yoursite.com/sport/”

Wrong: “yoursite.com/sport/” to “m.yoursite.com”

Also, you have to make sure that if one page is a 404 on desktop, it

should also be a 404 on mobile.

If you have no mobile version for a page, you can avoid redirecting

to mobile version and keep them on the desktop page.

3. Using Meta Refresh

It is possible to do redirect using meta refresh tag like example

below:

<meta http-equiv=”refresh” content=”0;url=http://yoursite.com/newpage/”

/>

If you insert this tag in /old-page/ it will redirect the user

immediately to /new-page/. This redirect is not prohibited by

Google but they clearly don’t recommend using it.

A TECHNICAL SEO GUIDE TO REDIRECT

14

According to John Mueller, search engines may not be able to

recognize that type of redirect properly. The same is also true about

JavaScript redirects.

4. Too Many Redirects

This message displays when you have a wrong regular expression

setup and it ends up in an infinite loop.

A TECHNICAL SEO GUIDE TO REDIRECT

14

Usually, this happens when you have a redirects chain.

Let’s say you redirected page1 to page2 a long time ago. Now you

might have forgotten that page1 is redirected and decide to redirect

page2 to page1 again.

As a result, you will end up with a rule like this:

RewriteRule ^page1 /page2 [R=301,NC,L]

RewriteRule ^page2 /page1 [R=301,NC,L]

This will create an infinite loop and produce the error shown above.

A TECHNICAL SEO GUIDE TO REDIRECT

14

SEO-Friendly

Pagination: A Complete

Best Practices Guide

Chapter 15

Written By

International Digital Director, Ringier

Jes Scholz

15

Site pagination is a wily

shapeshifter.

SEO FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

It’s used in contexts ranging from displaying items on category pages, to

article archives, to gallery slideshows and forum threads.

For SEO professionals, it isn’t a question of if you’ll have to deal with

pagination, it’s a question of when.

At a certain point of growth, websites need to split content across a

series of component pages for user experience (UX).

15

Our job is to help search engines crawl and understand the

relationship between these URLs so they index the most

relevant page.

Over time, the SEO best practices of pagination handling have

evolved. Along the way, many myths have presented themselves as

facts. But no longer.

This article will:

Debunk the myths around how pagination hurts SEO.

Present the optimal way to manage pagination.

Review misunderstood or subpar methods of pagination

handling.

Investigate how to track the KPI impact of pagination.

But, before I dig into these details. It’s important to note that

pagination isn’t for ranking purposes, but it still has value.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

How Pagination Can Hurt SEO

You’ve probably read that pagination is bad for SEO.

However, in most cases, this is due to a lack of correct pagination

handling, rather than the existence of pagination itself.

Let’s look at the supposed evils of pagination and how to overcome

the SEO issues it could cause.

Pagination Causes Duplicate Content

Correct if pagination has been improperly implemented, such

as having both a “View All” page and paginated pages without a

correct rel=canonical or if you have created a page=1 in addition to

your root page.

Incorrect when you have SEO friendly pagination. Even if your H1

and meta tags are the same, the actual page content differs. So it’s

not duplication.

Joost de Valk

@jdevalk

John

@JohnMu

@JohnMu do you agree that people can safely ignore the

duplicate meta description warning in Google Search

Console for paginated archive URLs?

Yep, that’s fine. It’s useful to get feedback on duplicate

titles & descriptions if you accidentally use them on

totally separate pages, but for paginated series, it’s kinda

normal & expected to use the same.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Pagination Creates Thin Content

Correct if you have split an article or photo gallery across multiple

pages (in order to drive ad revenue by increasing pageviews), leaving

too little content on each page.

Incorrect when you put the desires of the user to easily consume

your content above that of banner ad revenues or artificially inflated

pageviews. Put a UX-friendly amount of content on each page.

Pagination Dilutes Ranking Signals

Correct if pagination isn’t handled well as it can cause internal

link equity and other ranking signals, such as backlinks and social

shares, to be split across pages.

Incorrect when rel=”prev” and rel=”next” link attributes are used on

paginated pages, so that Google knows to consolidate the ranking

signals.

Pagination Uses Crawl Budget

Correct if you’re allowing Google to crawl paginated pages. And

there are some instances where you would want to use that budget.

For example, for Googlebot to travel through paginated URLs to

consolidate ranking signals and to reach deeper content pages.

Often incorrect when you set Google Search Console pagination

parameter handling to “Do not crawl” or set a robots.txt disallow, in

the case where you wish to conserve your crawl budget for more

important pages.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Managing Pagination According to

SEO Best Practices

Use rel=”next” & rel=”prev” Link Attributes

You should indicate the relationship between component URLs in a

paginated series with rel=”next” and rel=”prev” attributes.

Google recommends this option, noting they take this markup as

“a strong hint” that you would like the pages to be treated “as a logical

sequence, thus consolidating their linking properties and usually

sending searchers to the first page.”

Practically, this means rel=”next” / “prev” are treated as signals rather

than directives. They won’t always prevent paginated pages from

being displayed in search results. But such an occurrence would be

rare.

Complement the rel=”next” / “prev” with a self-referencing

rel=”canonical” link. So /category?page=4 should rel=”canonical” to /

category?page=4.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

This is the recommended approach by Google, as pagination

changes the page content and so is the master copy of that page.

If the URL has additional parameters, include these in

the rel=”prev” / “next” links, but don’t include them in the

rel=”canonical”.

For example:

<link rel=”next” href=”https://www.example.com/

category?page=2&order=newest” />

<link rel=”canonical” href=”https://www.example.com/

category?page=2” />

Doing so will indicate a clear relationship between the pages,

without sending ranking signals to non-SEO relevant parameterbased

URLs and preventing the potential of duplicate content.

Common errors to avoid:

Placing the link attributes in the <body> content. They’re only

supported by search engines within the <head> section of your

HTML.

Adding a rel=”prev” link to the first page (a.k.a. the root page) in

the series or a rel=”next” link to the last. For all other pages in the

chain, both link attributes should be present.

Beware of your root page canonical URL. Chances are on

?page=2, rel=prev should link to the canonical, not a ?page=1.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

The <head> code of a four-page series will look

something like this:

One pagination tag on the root page, pointing to the next page

in series.

<link rel=”next” href=”https://www.example.com/

category?page=2″>

<link rel=”canonical” href=”https://www.example.com/

category”>

Two pagination tags on page 2.

<link rel=”prev” href=”https://www.example.com/

category”>

<link rel=”next” href=”https://www.example.com/

category?page=3″>

<link rel=”canonical” href=”https://www.example.com/

category?page=2”>

Two pagination tags on page 3.

<link rel=”prev” href=”https://www.example.com/

category?page=2″>

<link rel=”next” href=”https://www.example.com/

category?page=4″>

<link rel=”canonical” href=”https://www.example.com/

category?page=3”>

One pagination tag on page 4, the last page in the paginated

series.

<link rel=”prev” href=”https://www.example.com/

category?page=3”>

<link rel=”canonical” href=”https://www.example.com/

category?page=4”>

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Modify Paginated Pages Titles & Meta Descriptions

Although the rel=”next” and rel=”prev” attributes should, in most

cases, cause Google to return the root page in the SERPs, you can

further encourage this and prevent “Duplicate meta descriptions”

or “Duplicate title tags” warnings in Google Search Console with an

easy modification to your code.

If the root page has the formula:

These paginated URL page titles and meta description are

purposefully suboptimal to dissuade Google from displaying these

results, rather than the root page.

Don’t Include Paginated Pages in XML Sitemaps

While rel=”next” / “prev” pagination URLs are technically indexable,

they aren’t an SEO priority to spend crawl budget on.

As such, they don’t belong in your XML sitemap.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Handle Pagination Parameters in Google Search Console

If you have a choice, run pagination via a parameter rather than a

static URL. For example:

example.com/category?page=2 over example.com/category/page-

2

You can then configure the parameter in Google Search Console

to “Paginates” and at any time change the signal to Google to crawl

“Every URL” or “No URLs”, based on how you wish to use your crawl

budget. No developer needed!

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Misunderstood, Outdated or Plain

Wrong SEO Solutions to Paginated

Content

Do Nothing

Google says they do “a good job returning the most relevant results

to users, regardless of whether content is divided into multiple

pages” and recommends you can handle pagination by doing

nothing.

While there is a core of truth to this statement, by doing nothing

you’re gambling with your SEO.

There’s always value in giving clear guidance to crawlers how you

want them to index and display your content.

Canonicalize to a View All Page

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

The last option recommended by Google is a View All page. This

version should contain all the component page content on a single

URL.

Additionally, the paginated pages should all rel=”canonical” to the

View All page to consolidate ranking signals.

The argument here is that searchers prefer to view a whole article

or list of categories items on a single page, as long as it’s fast

loading and easy to navigate.

So if your paginated series has an alternative View All version that

offers the better user experience, Google will favor this page for

inclusion in the search results as opposed to a relevant segment

page of the pagination chain.

Which raises the question – why do you have paginated pages in

the first place?

Let’s make this simple.

If you can provide your content on a single URL while offering a

good user experience, there is no need for pagination or a View All

version.

If you can’t, for example, a category page with thousands of

products would be ridiculously large and take too long to load, then

paginate with rel=”next” / “prev”. View All is not the best option as it

would not offer a good user experience.

Using both rel=”next” / “prev” and a View All version gives no clear

mandate to Google and will result in confused crawlers.

Don’t do it.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Canonicalize to the First Page

A common mistake is to point the rel=”canonical” from all paginated

results to the root page of the series.

Some ill-informed SEO people suggest this as a way to consolidate

authority across the set of pages to the root page, but this is

unnecessary when you have rel=”next” and rel=”prev” attributes.

Incorrect canonicalization to the root page runs the risk of

misdirecting search engines into thinking you have only a single

page of results.

Googlebot then won’t index pages that appear further along the

chain, nor acknowledge the signals to the content linked from those

pages.

You don’t want your detailed content pages dropping out of the

index because of poor pagination handling.

Google is clear on the requirement. Each page within a paginated

series should have a self-referencing canonical, unless you use a

View All page.

Use the rel=canonical incorrectly and chances are Googlebot will

just ignore your signal.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Noindex Paginated Pages

A classic method to solve pagination issues was a robots noindex

tag to prevent paginated content from being indexed by search

engines.

Relying solely on the noindex tag for pagination handling will

result in ranking signals from your component pages not being

consolidated. Clearly inferior SEO to using rel=”next” / “prev”.

But as the rel=”next” / “prev” method allows search engines to index

pagination pages, I’ve also seen some SEO folks advising to add

“extra security” with a noindex tag.

This is unnecessary. Only in rare circumstances would Google

choose to return a paginated page in the SERPs. The benefits are, at

best, theoretical.

But what you may not be aware of is that a long-term noindex on

a page will eventually lead Google to nofollow the links on that

page. So, again, it could potentially cause content linked from the

paginated pages to be removed from the index.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Pagination & Infinite Scrolling

A newer form of pagination handling is by infinite scroll, where

content is pre-fetched and added directly to the user’s current page

as they scroll down.

Users may appreciate this, but Googlebot? Not so much.

Googlebot doesn’t emulate behavior like scrolling to the bottom of a

page or clicking to load more. Meaning without help, search engines

can’t effectively crawl all of your content.

To be SEO-friendly, convert your infinite scroll page to an equivalent

paginated series that is accessible even with JavaScript disabled.

As the user scrolls, use JavaScript to adapt the URL in the address

bar to the component paginated page.

Additionally, implement a pushState for any user action that

resembles a click or actively turning a page. You can check out this

functionality in the demo created by John Mueller.

Essentially, you’re still implementing the SEO best practice

recommended above, you are just adding additional user experience

functionality on top.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Discourage or Block Pagination Crawling

Some SEO pros recommend avoiding the issue of pagination

handling altogether by simply blocking Google from crawling

paginated URLs.

In such a case, you would want to have well-optimized XML

sitemaps to ensure pages linked via pagination have a chance to be

indexed.

There are three ways to do this:

The messy way: Add nofollow to all links that point towards

paginated pages.

The cleaner way: Use a robots.txt disallow.

The no dev needed way: Set paginated page parameter to

“Paginates” and for Google to crawl “No URLs” in Google Search

Console.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

By using one of these methods to discourage search

engines from crawling paginated URLs you:

Stop search engines from consolidating ranking signals of

paginated pages.

Prevent the passing of internal link equity from paginated pages

down to the destination content pages.

Hinder Google’s ability to discover your destination content pages.

The obvious upside is that you save on crawl budget.

There is no clear right or wrong here. You need to decide what is

the priority for your website.

Personally, if I were to prioritize crawl budget, I would do so by

using pagination handling in Google Search Console as it has the

optimum flexibility to change your mind.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

Tracking the KPI Impact of

Pagination

So now you know what to do, how do you track the effect of

optimization pagination handling?

Firstly, gather benchmark data to understand how your current

pagination handing is impacting SEO.

Sources for KPIs can include:

Server log files for the number of paginated page crawls.

Site: search operator (for example site:example.com inurl:page) to

understand how many paginated pages Google has indexed.

Google Search Console Search Analytics Report filtered by pages

containing pagination to understand the number of impressions.

Google Analytics landing page report filtered by paginated URLs to

understand on-site behavior.

If you see an issue getting search engines to crawl your site

pagination to reach your content, you may want to change the

pagination links.

Once you have launched your best practice pagination handling,

revisit these data sources to measure the success of your efforts.

SEO-FRIENDLY PAGINATION: A COMPLETE BEST PRACTICES GUIDE

15

What is Schema

Markup & Why It’s

Important for SEO

Chapter 16

Written By

Founder, Measurable SEO

Chuck Price

16

What is Schema Markup?

Schema markup, found at Schema.org, is a form of microdata.

Once added to a webpage, schema markup creates an

enhanced description (commonly known as a rich snippet),

which appears in search results.

Top search engines – including Google, Yahoo, Bing, and Yandex

– first started collaborating to create Schema.org, back in 2011.

Schema markup is especially important in the age of

Hummingbird and RankBrain. How a search engine interprets

the context of a query will determine the quality of a search

result.

Schema can provide context to an otherwise ambiguous

webpage.

Via Schema.org:

“Most webmasters are familiar with HTML tags on their

pages. Usually, HTML tags tell the browser how to display the

information included in the tag. For example, <h1>Avatar</h1>

tells the browser to display the text string “Avatar” in a heading

1 format. However, the HTML tag doesn’t give any information

about what that text string means—”Avatar” could refer to the

hugely successful 3D movie, or it could refer to a type of profile

picture—and this can make it more difficult for search engines to

intelligently display relevant content to a user.”

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

Does Schema Improve Your

Search Rankings?

There is no evidence that microdata has a direct affect on organic

search rankings.

Nonetheless, rich snippets do make your webpages appear more

prominently in SERPs. This improved visibility has been shown to

improve click-through rates.

According to a study by acmque, less than one-third of Google’s

search results include a rich snippet with Schema.org markup. This

exposes a huge opportunity for the rest. Very few things in SEO,

today, can move the dial quickly. This can.

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

What Is Schema Used For?

Businesses and organizations

Events

People

Products

Recipes

Reviews

Videos

Above are some of the most popular uses of schema. However,

there’s a good chance that if you have any sort of data on your

website, it’s going to have an associated itemscope, itemtype and

itemprop.

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

Adding Schema to Your

Webpages

Using Microdata

Microdata is a set of tags that aims to make annotating HTML

elements with machine-readable tags much easier. Microdata is a

great place for beginners to start because it’s so easy to use.

However, the one downside to using microdata is that you have to

mark every individual item within the body of your webpage. As you

can imagine, this can quickly get messy.

Before you begin to add schema to your webpages, you need to

figure out the ‘item type’ of the content on your webpage.

For example, does your web content focus on food? Music? Tech?

Once you’ve figured out the item type, you can now determine how

you can tag it up.

Let’s look at an example. Let’s say that you own a store that sells

high-quality routers. If you were to look at the source code of your

homepage you would likely see something akin to this:

<div>

<h1>TechHaven</h1>

<h2>The best routers you’ll find online!</h2>

<p>Address:</p>

<p>459 Humpback Road</p>

<p>Rialto, Ca</p>

<p>Tel: 909 574 3903</p>

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

<p><a href=”http://www.techhaven.com/menu”>Click here to

view our best routers!</a></p>

<p>We’re open: </p>

<p>Mon-Sat 8am – 10:30pm</p>

<p>Sun: 2pm – 8pm</p>

</div>

Once you dive into the code, you’ll want to find the portion of your

webpage that talks about what your business has to offer. In this

example, that data can be found between the two <div> tags.

Now, at the top you can add in:

<div itemscope>

By adding this tag, we’re stating that the HTML code contained

between the <div> blocks are identifying a specific item.

Next, we have to identify what that item is by using the ‘itemtype’

attribute to identify the type of item our webpage is about (tech).

<div itemscope itemtype=”http://schema.org/tech”>

An item type comes in the form of a URL (such as http://schema.

org/tech). Let’s say, for example, that your site revolved around

beauty products instead of technology. Your item type URL might

look like this:

http://schema.org/beauty.

To make things easier you can browse a list of item types here,

plus you can view extensions to identify the specific entity that

you’re looking for. Keep in mind that this list is not all encompassing,

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

so there is a possibility that you won’t find the item type for your

specific niche.

Tracking back to the tech page, you want to tag the part of the

webpage that contains the name of the business. You can do this

between the <h1> tags.

Now, we’ll be using the ‘itemprop’ tag, which labels the properties of

an item:

<h1 itemprop=”name”>Tech Haven</h1>

You can apply these tags to the rest of the page now. When using

tags to identify item properties, it’s not necessary to tag the entire

line, just the one portion the property is making reference to.

For example, if you have a line that says Address: 1234 w sunshine

blvd, then you only need to apply the tags around the address itself

and nothing else.

<h2 itemprop=”description”>The best routers you’ll find

online!</h2>

<p>Address:</p>

<span itemprop=”address” itemscope itemtype=”http://

schema.org/PostalAddress”>

<p itemprop=”streetAddress”>459 Humpback Road </p>

<p itemprop=”addressLocality”>Rialto, Ca</p></span>

<p>Tel: <span itemprop=”telephone”>909 574 3903</

span></p>

<p><a itemprop=”menu” href=”http:// http://www.techhaven.

com/menu “>Click here to view our tasty range of dishes!</

a></p>

<p>We’re open:</p>

<p itemprop=”openingHours”>Mon-Sat 8am – 10:30pm</p>

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

<p itemprop=”openingHours”>Sun: 2pm – 8pm</p>

</div>

This code may look complicated, but schema.org provides examples

on how to use the different item types, so you can actually see what

the code is supposed to do. Don’t worry, you won’t be left out in the

cold trying to figure this out on your own!

If you’re still feeling a little intimidated by the code, Google’s

Structured Data Markup Helper makes it super easy to tag your

webpages.

To use this amazing tool, just select your item type, paste in the

URL of the target page or the content you want to target, and then

highlight the different elements so that you can tag them.

Using RDFa

RDFa is an acronym for Resource Description Framework in

Attributes. Essentially, RDFa is an extension to HTML5 and it was

designed to aid users in marking up structured data.

RDFa is considered to be a W3C recommendation, meaning that

it is a web standard, and it can be used to chain structured data

vocabularies together. This is especially useful if you want to add

structured data that stretches beyond the limits of Schema.org.

You can breathe a sigh of relief. RDFa isn’t much different from

Microdata.

Similar to microdata, RDFa tags incorporate with the preexisting

HTML code in the body of your webpage. For the sake of familiarity,

we’ll look at the tech website once again as an example.

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

The HTML for your tech site would likely look like this before it was

modified:

<div>

<h1>Tech Haven</h1>

<h2>The best routers online!</h2>

<p>Address:</p>

<p>459 Humpback Road </p>

<p>Rialto, Ca</p>

<p>Tel: 909 574 3903</p>

<p><a href=”http://www.techhaven.com/menu”>Click here

to view our best routers!</a></p>

<p>We’re open:</p>

<p>Mon-Sat 8:00am – 10:30pm</p>

<p>Sun: 2pm – 8pm</p>

</div>

To begin, you want to ensure that the vocabulary that you’re using is

Schema.org and that the webpage in question is making reference

to a technology page.

For this example, you can search for “technology” on Schema.org

to learn how to tag different elements. Typically, you’ll find examples

near the bottom of the page that will show you how to use them in

practice.

Simply click on the RDFa tab to view specific RDFa examples.

Next, you need to use the vocab tag combined with the URL http://

schema.org to identify the vocabulary for the markup. To identify the

page type, use the typeoftag. Unlike microdata, which uses a URL to

identify types, RDFa uses one or more words to classify types.

<div vocab=”http://schema.org/” typeof=”technology”>

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

If you wish to identify a property further than you should use the

typeof attribute.

For example, if you wish to further expand upon an address

property you can use “PostalAddress” like so:

<div property=”address” typeof=”PostalAddress”>

Comparing microdata and RDFa side by side, the typeof attribute

is the equivalent of the itemtype attribute found in Microdata.

Furthermore, the propertyattribute would be the equivalent to the

itemprop attribute.

For further explanation, you can visit Schema.org to check lists and

view examples. You can find which kinds of elements are defined as

properties, and which are defined as types.

Going back to our earlier example, the tech page would look like

this after it has been appropriately tagged:

<h2 property=”description”>The best routers on the

internet!</h2>

<p>Address:</p>

<div property=”address” typeof=”PostalAddress”>

<p property=”streetAddress”>459 Humpback Road</p>

<p property=”addressLocality”>Rialto, Ca</p>

</div>

<p>Tel: <span property=”telephone”>909 574 3903</

span></p>

<p><a property=”menu” href=”http://www.techhaven/

menu”>Click here to view our best routers!</a></p>

<p>We’re open:</p>

<p property=”openingHours”>Mon-Sat 8am – 10:30pm</p>

<p property=”openingHours”>Sun: 2pm – 8pm</p>

</div>

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

Conclusion

Hopefully, any fears that you might have had when you heard the

word “Schema” or “Structured Data” have been put to rest.

Schema is much easier to apply than it seems and it’s a best practice

that you need to incorporate into your webpages.

The work may seem tedious. But given time and effort, you’ll be

certain to reap the benefits from your labor.

To aid you, every page on Schema.org provides examples on how

to properly apply tags. Of course, you can also fall back on Google’s

Structured Data Testing Tool.

WHAT IS SCHEMA MARKUP & WHY IT’S IMPORTANT FOR SEO

16

Faceted Navigation:

Best Practices for SEO

Chapter 17

Written By

Digital Marketing Specialist, Forthea

Interactive

Natalie Hoben

17

When it comes to large websites,

such as e-commerce sites with

thousands upon thousands of

page, the importance of things

like crawl budget cannot be

understated.

FACETED NAVIGATION: BEST PRACTICES FOR SEO

Building a website with an organized architecture and smart internal

linking strategy is key for these types of sites.

However, doing that properly oftentimes involves new challenges when

trying to accommodate various attributes that are a common theme with

e-commerce (sizes, colors, price ranges, etc.).

Faceted navigation can help solve these challenges on large websites.

However, faceted navigation must be well thought out and executed

properly so that both users and search engine bots remain happy.

17

What is Faceted Navigation?

To begin, let’s dive into what faceted navigation actually is.

Faceted navigation is, in most cases, located on the sidebars of an

e-commerce website and has multiple categories, files, and facets.

It essentially allows people to customize their search based on what

they are looking for on the site.

For example, a visitor may want a purple cardigan, in a size medium,

with black trim.

Facets are indexed categories that help to narrow down a

production listing and also function as an extension of a site’s main

categories.

Facets, in their best form, should ideally provide a unique value for

each selection and, as

they are indexed, each

one on a site should send

relevancy signals to search

engines by making sure

that all critical attributes

appear within the content

of the page.

FACETED NAVIGATION: BEST PRACTICES FOR SEO

17

Filters are utilized to sort items with a listings page.

While the user can use this to narrow down what they are looking for,

the actual content on the page remains the same.

This can potentially lead to multiple URLs creating duplicate content,

which is a concern for SEO.

There are a few potential issues that faceted navigation can create

that can negatively affect SEO.

FACETED NAVIGATION: BEST PRACTICES FOR SEO

17

The main three issues boil down to:

Duplicate content.

Wasted crawl budget.

Diluted link equity.

As different parameters are created, they can quickly multiply. The

number of incredibly-related pieces of content continues to grow

significantly and different links may be going to all of these different

versions of a page, which can dilute link equity, and thus affect the

page’s ranking ability.

In order to be able to make sure that search engine crawlers aren’t

wasting valuable crawl budget on pages that have little to no value,

you need to take certain steps.

That starts with preventing search engine bots from crawling certain

multi-selected facets, such as “color” or “size”.

When trying to determine how to solve this faceted navigation

conundrum, there are a few solutions that are implementable. Which

one to use, however, will rely heavily on what parts of the site should

be indexed.

Noindex tags can be implemented to inform bots of which pages

not to include in the index. This method will remove pages of the

index, however, there will still be crawl budget spent on them and

link equity that is diluted.

Noindex

FACETED NAVIGATION: BEST PRACTICES FOR SEO

17

A disallow can be implemented for certain sections of a site.

The advantage to this solution is that it’s fast and customizable.

However, the disallow is merely a directive for Google, and they do

not have to abide by it.

In addition, link equity may be hindered from flowing to different

parts of the site.

For example, we could disallow red sweaters under $50 in the

robots file, instructing Google to not visit a page with the >$50

parameter. However, if any follow links pointing to any URL with that

parameter in it existed, Google could still possible index it.

Robots.txt

FACETED NAVIGATION: BEST PRACTICES FOR SEO

17

For example, if you wished to include a page for “red sweaters” in

the index, but did not want “red sweaters under $50” in the index,

then a noindex tag to the second result would exclude it.

Bots would still be able to find and crawl the page, though, and

this causes crawl budget to be wasted.

The pages would also still get wasted link equity.

Canonical tags allow you to instruct Google that a group of alike

pages has a preferred version.

Link equity can be consolidated into the chosen preferred page

utilizing this method. However, crawl budget will still be wasted.

Canonical tags can also be ignored by search engines, so this

solution should be used along with another.

For example, /red-sweaters?under-50/ could have the canonical

URL set as /red-sweaters/. Google would attribute the authority and

link equity to the canonicalized page, but crawl budget would still be

wasted.

When it comes to using AJAX to solve faceted navigation issues, the

main positive benefit is that a different, new URL is not generated

when a visitor visits a page and selects a filter.

JavaScript hosted client-slide takes care of the entire process. No

web server is needed.

In order to ensure that this method is effective, it is necessary that a

crawl path is accessible to the particular pages that are important to

get into rankings.

The pushState method of the HTML5 history API and server

configuration that responds to these requests with HTML rendered

server-side can help ensure that AJAX can fully work its magic and

keep SEO in a healthy state.

Canonicalization

AJAX

FACETED NAVIGATION: BEST PRACTICES FOR SEO

17

This should ideally be a last resort option.

It is a decent temporary solution while adjustments are being made

to the navigation. This is because it only instructs Google on how a

site should be crawled, instead of correcting the issue.

By navigating to the URL parameters tool in Google Search

Console, you can choose what effect each parameter has on the

page and how Google should treat those pages.

Implement pagination with rel=”next” and rel=”prev” in order to

group indexing properties, from pages to a series as a whole.

Each page needs to link to children pages and parent. This can be

done with breadcrumbs.

Use rigorous URL facet ordering so that duplication problems do

not arise.

Prevent clicks when no items are present for the filter.

Only use canonical URLs in sitemaps.

Facets should always be presented in a unified, logical manner

(i.e., alphabetical order).

Google Search Console

Other Ways to Get the Most out of

Faceted Navigation

FACETED NAVIGATION: BEST PRACTICES FOR SEO

17

Although faceted navigation can be great for UX, it can cause a

multitude of problems for SEO.

Duplicate content, wasted crawl budget, and diluted link equity can

all cause severe problems on a site.

It is crucial to carefully plan and implement the necessary methods

available in order to avoid any many issues down the line when it

comes to faceted navigation.

Conclusion

FACETED NAVIGATION: BEST PRACTICES FOR SEO

17

Don’t rely solely on one “fix” if it doesn’t take care of indexing, link

dilution, and crawl. For example, noindex and nofollow tags do

not help with crawl budget. Same with configuring parameters in

Google Search Console.

If a particular combination of facets occurs that receive a good

amount of traffic, consider allowing indexation.

Understanding

JavaScript

Fundamentals: Your

Cheat Sheet

Chapter 18

Written By

Technical SEO & Content Manager,

DeepCrawl

Rachel Costello

18

JavaScript is a complex topic

that can be extremely difficult

to get a handle on. UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

However, it has never been more important to understand it because

JavaScript is becoming increasingly prevalent on the websites that we

manage.

As the modern web continues to evolve, JavaScript usage continues

to rise.

SEO professionals may long for times gone by when websites were static

and coded only in HTML and CSS. However, engaging websites often

require interactivity, which is usually powered by JavaScript.

18

The number of JavaScript bytes across the entire web has

increased by 35 percent on desktop and 62.7 percent on mobile

over the last three years.

As Google Webmaster Trends Analyst John Mueller put it:

JavaScript is “not going away.”

This programming language is all around us, so we should get

better acquainted with it. Let’s be proactive and learn more about

JavaScript rather than fearing it.

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

There is often a misconception that JavaScript is solely for

developers to worry about.

I would argue that this isn’t the case, as it can cause a problem for

anyone who wants customers and search engines to be able to

access their website’s content.

If you aren’t completely familiar with JavaScript, or even have

absolutely no idea what it is or does, don’t worry.

I’ve put together a glossary of the key terms and fundamental

concepts you should know to help you get started on your journey

of discovery.

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

JavaScript is a programming language that allows you to

implement complex features on a website, such as dynamic

elements or interactivity.

JavaScript is executed once the information from the HTML and

CSS in the source code has been parsed and constructed.

The JavaScript will then trigger any events or variables specified

within it, the Document Object Model (DOM) will be updated,

and, finally, the JavaScript will be rendered in the browser.

The HTML and CSS will often form the foundations of a page’s

structure, and any

JavaScript will make

the final tweaks and

alterations.

What Is JavaScript?

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

The Document Object Model (DOM) is created when a page

is loaded, and it is made up of nodes and objects which map out

all of the different elements and attributes on a page.

The page is mapped out in this way so that other programs can

modify and manipulate the page in terms of its structure, content,

and styling.

Altering the elements of a page’s DOM is possible through using a

language like JavaScript.

Document Object Model (DOM)

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

ECMAScript (ES) is a scripting language that was created to

standardize the use of JavaScript code.

Different editions of ECMAScript are released when the language is

updated and tweaked over time, such as ES5 and ES6 (which is also

referred to as ES2015).

A transpiler is a tool that transforms source code into a different

programming language. The concept is a bit like Google Translate,

but for code.

You can convert a particular

source language into a

different target language, for

example, JavaScript to C++ or

Python to Ruby.

With regard to JavaScript rendering

particularly, a transpiler is often

recommended for transforming ES6

into ES5 because Google currently

uses an old version of Chrome for

rendering which doesn’t yet support

ES6.

ECMAScript

Transpiling

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

When rendering pages, Google uses a web rendering service

which is based on Chrome 41. This means that Google’s rendering

engine supports the same features and functionalities of that

particular version of Chrome.

When you consider that the most up-to-date version is Chrome

71, you can see that many versions have been launched since

Chrome 41 went live in 2015, and all of these versions came

with new features. This is why Google’s rendering service currently

supports ES5 rather than the later ES6 version of the language.

A single-page application (SPA) is a website or web app that

dynamically re-writes and re-renders a page as a user interacts

with it, rather than making separate requests to the server for new

HTML and content.

JavaScript frameworks can be used to support the dynamically

changing elements of SPAs.

Chrome 41

Single-page Application (SPA)

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

These are all different types of JavaScript frameworks.

Angular and Polymer were developed by Google.

React was developed by Facebook.

Vue was developed by Evan You, who used to work on Google’s

Angular team.

Each JavaScript framework has its own pros and cons, so

developers will choose to work with the one that best suits them and

the project they’re working on.

If you want to learn more about how the different frameworks

measure up, this guide gives a detailed comparison.

Angular, Polymer, React & Vue

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

JavaScript rendering involves taking the script and the

instructions it contains, processing it all, then running it so that

the required output is shown in the browser. There are many

different methods you can use to control the way in which

JavaScript is rendered.

Requiring JavaScript to be rendered on a page can

negatively impact two key areas:

Site speed

Search engine crawling and indexing

Depending on which rendering method you use, you can reduce

page load speed and make sure content is accessible to search

engines for crawling and indexing.

Pre-rendering

Pre-rendering involves rendering the content on a page before it

is requested by the user or search engine, so that they receive a

static page with all of the content on there ready to go.

By preloading a page in this way, it means that your content will

be accessible rather than a search engine or user’s browser

having to render the page themselves.

JavaScript Rendering

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

Pre-rendering is usually used for search engine bots rather than

humans. This is because a static, pre-rendered page will be less

engaging for users as it will lack any dynamic content or interactivity.

Server-side Rendering

The hosting server does the heavy lifting and renders the page so

that the JavaScript has already been processed and the content

is ready to be handed over to the user’s browser or search engine

crawler when it is requested.

This method helps to reduce any strain on the user’s device that

would have been caused by processing JavaScript, and this can

increase page load speed.

Server-side rendering also ensures the full content can be seen and

indexed by search engines.

Client-side Rendering

During client-side rendering, JavaScript is processed by the user’s

browser or by the search engine that’s requesting a page.

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

It is often advised against to use client-side rendering as there is

a delay between Google crawling pages and then being able to

render them.

Google puts pages that need to be rendered into a queue until

enough resources become available to process them.

If you’re relying on Google to render a page clientside,

this can delay indexing by up to a

week after it is initially crawled.

Dynamic Rendering

Dynamic rendering

involves using different

rendering methods

depending on whether a user’s

browser or a search engine bot

is requesting a page.

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

The server will handle the initial request, but the rest of the work of

processing and rendering a page falls on the user’s device or the

search engine.

Hybrid Rendering

Hybrid rendering involves a combination of both server-side

rendering and client-side rendering.

The core content is pre-rendered server-side and sent to the client,

whether that’s the user’s browser or the search engine crawler that’s

requesting the content.

After the page is initially loaded, additional JavaScript for any

interactivity is then rendered client-side.

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

If your site usually renders client-side, when Googlebot is detected

the page will be pre-rendered using a mini client-side renderer (for

example, Puppeteer or Rendertron), so the content can be seen

and indexed straight away.

Hopefully you found this guide useful, and that it helped you better

understand the basics of JavaScript and how it impacts websites.

Now that you’ve brushed up on the key terms, you should be better

equipped to hold your own in conversations with the developers!

Conclusion

UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

18

An SEO Guide to URL

Parameter Handling

Chapter 19

Written By

International Digital Director, Ringier

Jes Scholz

19

While parameters are loved

by developers and analytics

aficionados, they are often an

SEO nightmare. UNDERSTANDING JAVASCRIPT FUNDAMENTALS: YOUR CHEAT SHEET

Endless combinations of parameters can create thousands of URL

variations out of the same content.

The problem is we can’t simply wish parameters away. They play an

important role in a website’s user experience. So we need to understand

how to handle them in an SEO-friendly way.

To do so we explore:

The basics of URL parameters

SEO issues caused by parameters

Assessing the extent of your parameter problem

SEO solutions to tame parameter

Best practice URL parameter handling

18

Also known by the aliases of query strings or URL variables,

parameters are the portion of a URL that follows a question mark.

They are comprised of a key and a value pair, separated by an equal

sign. Multiple parameters can be added to a single page by using an

ampersand.

The most common use cases for parameters are:

Tracking – For example ?utm_medium=social, ?sessionid=123 or

?affiliateid=abc

Reordering – For example ?sort=lowest-price, ?order=highest-rated

or ?so=newest

Filtering – For example ?type=widget, colour=blue or ?pricerange=

20-50

Identifying – For example ?product=small-blue-widget,

categoryid=124 or itemid=24AU

Paginating – For example, ?page=2, ?p=2 or viewItems=10-30

Searching – For example, ?query=users-query, ?q=users-query or

?search=drop-down-option

Translating – For example, ?lang=fr, ?language=de or

What Are URL Parameters?

AN SEO GUIDE TO URL PARAMETER HANDLING

19

1. Parameters Create Duplicate Content

Often, URL parameters make no significant change to the content

of a page. A re-ordered version of the page is often not so different

from the original. A page URL with tracking tags or a session ID is

identical to the original.

For example, the following URLs would all return collection

of widgets.

Static URL: https://www.example.com/widgets

Tracking parameter: https://www.example.com/

widgets?sessionID=32764

Reordering parameter: https://www.example.com/

widgets?sort=newest

Identifying parameter: https://www.example.

com?category=widgets

Searching parameter: https://www.example.com/

products?search=widget

That’s quite a few URLs for what is effectively the same content –

now imagine this over every category on your site. It can really add

up.

The challenge is that search engines treat every parameter based

URL is a new page. So they see multiple variations of the same

SEO Issues with URL

Parameters

AN SEO GUIDE TO URL PARAMETER HANDLING

19

page. All serving duplicate content and all targeting the same

keyword phrase or semantic topic.

While such duplication is unlikely to cause you to be completely

filtered out of the search results, it does lead to keyword

cannibalization and could downgrade Google’s view of your

overall site quality as these additional URLs add no real value.

2. Parameters Waste Crawl Budget

Crawling redundant parameter pages drains crawl budget, reducing

your site’s ability to index SEO relevant pages and increasing server

load.

Google sums up this point perfectly.

“Overly complex URLs, especially those containing multiple

parameters, can cause a problems for crawlers by creating

unnecessarily high numbers of URLs that point to identical or similar

content on your site. As a result, Googlebot may consume much

more bandwidth than necessary, or may be unable to completely

index all the content on your site.”

3. Parameters Split Page Ranking Signals

If you have multiple permutations of the same page content, links

and social shares may be coming in on various versions.

This dilutes your ranking signals. When you confuse a crawler, it

becomes unsure which of the competing pages to index for the

search query.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

4. Parameters Make URLs Less Clickable

Let’s face it. Parameter URLs are unsightly. They’re hard to read.

They don’t seem as trustworthy. As such, they are less likely to be

clicked.

This will impact page performance. Not only because CTR can

influence rankings, but also because it’s less clickable on social

media, in emails, when copy pasted into forums or anywhere else

the full URL may be displayed.

While this may only have a fractional impact on a single page’s

amplification, every tweet, like, share, email, link, and mention

matters for the domain.

Poor URL readability could contribute to a decrease in brand

engagement.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

It’s important to know every parameter used on your website. But

chances are your developers don’t keep an up to date list.

So how do you find all the parameter that need handling? Or

understand how search engines crawl and index such pages?

Know the value they bring to users?

Follow these five steps:

Run a crawler: With a tool like Screaming Frog you can search for

“?” in the URL.

Look in Google Search Console URL Parameters Tool: Google

auto-adds the query strings it finds.

Review your log files: See if Googlebot is crawling parameter

based URLs.

Search with site: inurl: advanced operators: Know how Google

is indexing the parameters you found by putting the key in a

site:example.com inurl:key combination query.

Assess the Extent of Your

Parameter Problem

AN SEO GUIDE TO URL PARAMETER HANDLING

19

AN SEO GUIDE TO URL PARAMETER HANDLING

19

Look in Google Analytics All Pages report: Search for “?” to

see how each of the parameters you found are used by users.

Be sure to check that URL query parameters have not been

excluded in the view setting.

Armed with this data, you can now decide how to best handle

each of your website’s parameters.

You have six tools in your SEO arsenal to deal with URL parameters

on a strategic level.

Limit Parameter-Based URLs

A simple review of how and why parameters are generated can

provide an SEO quick win. You will often find ways to reduce the

number of parameter URLs and so minimize the negative SEO

impact. There are four common issues to begin your review.

1. Eliminate Unnecessary Parameters

Ask you developer for a list of every website parameters and its

function. Chances are, you will discover parameters that no longer

perform a valuable function.

For example, users can be better identified by cookies than

sessionIDs. Yet the sessionID parameter may still exist on your

website as it was used historically.

SEO Solutions to Tame URL

Parameters

AN SEO GUIDE TO URL PARAMETER HANDLING

19

URL parameters should be added to a URL only when they have

a function. Don’t permit parameter keys to be added if the value is

blank.

In the above example, key2 and key3 add no value both literally and

figuratively.

3. Use Keys Only Once

Avoid applying multiple parameters with the same parameter name

and a different value.

For multi-select option, it is better to combine the values together

after a single key.

Or you may discover that a filter in your faceted navigation is rarely

applied by your users.

Any parameters caused by technical debt should be immediately

eliminated.

2. Prevent Empty Values

AN SEO GUIDE TO URL PARAMETER HANDLING

19

4. Order URL Parameters

If the same URL parameter are rearranged, the pages are

interpreted by search engines as equal. As such, parameter order

doesn’t matter from a duplicate content perspective. But each of

those combinations burn crawl budget and split ranking signals.

Avoid these issues by asking your developer to write a script to

always place parameters in a consistent order, regardless of how

the user selected them.

In my opinion, you should start with any translating parameters,

followed by identifying, then pagination, then layering on filtering

and reordering or search parameters and finally tracking.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

Pros:

Allows more efficient use

of crawl budget.

Reduces duplicate

content issues.

Consolidates ranking

signals to fewer pages.

Suitable for all parameter

types.

Cons:

Moderate technical

implementation time

The rel=”canonical” link attribute calls out that a page has identical

or similar content to another. This encourages search engines to

consolidate the ranking signals to the URL specified as canonical.

You can rel=canonical your parameter based URLs to your SEOfriendly

URL for tracking, identifying or reordering parameters. But

this tactic is not suitable when the parameter page content is not

close enough to the canonical, such as pagination, searching,

translating or some filtering parameters.

Rel=”Canonical” Link Attribute

AN SEO GUIDE TO URL PARAMETER HANDLING

19

Pros:

Relatively easy technical

implementation.

Very likely to safeguard

against duplicate content

issues.

Consolidates ranking signals

to the canonical URL.

Cons:

Wastes crawl budget on

parameter pages.

Not suitable for all parameter

types.

Interpreted by search

engines as a strong hint, not

a directive.

Set a noindex directive for any parameter based page that doesn’t

add SEO value. This tag will prevent search engines from indexing

the page.

URLs with a “noindex” tag are also likely to be crawled less

frequently and if it’s present for a long time will eventually lead

Google to nofollow the page’s links.

Meta Robots Noindex Tag

Pros:

Relatively easy technical

implementation.

Very likely to safeguard against

duplicate content issues.

Suitable for all parameter types

you do not wish to be indexed.

Removes existing parameterbased

URLs from the index.

Cons:

Won’t prevent search engines

from crawling URLs, but will

encourage them to do so less

frequently.

Doesn’t consolidate ranking

signals.

Interpreted by search engines

as a strong hint, not a directive.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

The robots.txt file is what search engines look at first before crawling

your site. If they see something is disallowed, they won’t even go

there.

You can use this file to block crawler access to every parameter

based URL (with Disallow: /*?*) or only to specific query strings you

don’t want to be indexed.

Pros:

Simple technical

implementation.

Allows more efficient use

of crawl budget.

Avoids duplicate content

issues.

Suitable for all parameter

types you do not wish to

be crawled.

Cons:

Doesn’t consolidate ranking

signals.

Doesn’t remove existing

URLs from the index.

Robots.txt Disallow

AN SEO GUIDE TO URL PARAMETER HANDLING

19

Configure the Google’s URL parameter tool to tell crawlers the

purpose of your parameters and how you would like them to be

handled.

Google Search Console has a warning message that using the tool

“could result in many pages disappearing from a search.”

This may sound ominous. But what’s more menacing is thousands

of duplicate pages hurting your website’s ability to rank.

So it’s best to learn how to configure URL parameters in Google

Search Console, rather than letting Googlebot decide.

The key is to ask yourself how the parameter impacts the

page content:

Tracking parameters don’t change page content. Configure them

as “representative URLs”.

URL Parameter Tool in Google Search Console

AN SEO GUIDE TO URL PARAMETER HANDLING

19

Configure parameters that reorder page content as “sorts”. If this

is optionally added by the user, set crawl to “No URLs”. If a sort

parameter it is applied by default, use “Only URLs with value”,

entering the default value.

Configure parameters that filter page down to a subset of content

as “narrows”. If these filters are not SEO relevant, set crawl to “No

URLs”. If they are SEO relevant set to “Every URL”.

Configure parameters that shows a certain piece or group of

content as “specifies”. Ideally, this should be static URL. If not

possible, you will likely want to set this to “Every URL”.

Configure parameters that display a translated version of the

content as “translates”. Ideally, translation should be achieved via

subfolders. If not possible, you will likely want to set this to “Every

URL”.

Configuration parameters that display a component page of a

longer sequence as “paginates”. If you have achieved efficient

indexation with XML sitemaps, you can save crawl budget and

set crawl to “No URL”. If not, set to “Every URL” to help crawlers to

reach all of the items.

Google will automatically add parameters to the list under the

default “Let Googlebot decide”. The challenge is, these can never

be removed, even if the parameter no longer exists. So whenever

possible, it’s best to proactively add parameters yourself. So that if

at any point that parameter no longer exists, you may delete it from

GSC.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

For any parameter you set in Google Search Console to “No URL”,

you should also consider adding it in Bing’s ignore URL parameters

tool.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

Pros:

No developer time needed.

Allows more efficient use of

crawl budget.

Likely to safeguard against

duplicate content issues.

Suitable for all parameter

types.

Cons:

Doesn’t consolidate ranking

signals.

Interpreted by Google as a

helpful hint, not a directive.

Only works for Google and

with lesser control for Bing.

Many people think the optimal way to handle URL parameters is

simply avoid them in the first place. After all, subfolders surpass

parameters to help Google understand site structure and static,

keyword based URLs have always been a cornerstone of on-page

SEO.

To achieve this, you can use server-side URL rewrites to convert

parameters into subfolder URLs.

For example, the URL:

www.example.com/view-product?id=482794

Would become:

www.example.com/widgets/blue

This approach works well for descriptive keyword based parameters,

such as those which identify categories, products, or filter for search

engine relevant attributes. It is also effective for translated content.

But it becomes problematic for non-keyword relevant elements of

faceted navigation, such as price. Having such a filter as a static,

indexable URL offers no SEO value.

It’s also an issue for searching parameters, as every user generated

query would create a static page that vies for ranking against the

canonical – or worse presents to crawlers low quality content pages

whenever a user has searched for a item you don’t offer.

Move From Dynamic to Static URLs

AN SEO GUIDE TO URL PARAMETER HANDLING

19

It’s somewhat odd when applied to pagination (although not

uncommon due to WordPress), which would give a URL such as

www.example.com/widgets/blue/page2

Very odd for reordering, which would give a URL such as

www.example.com/widgets/blue/lowest-price

And is often not a viable option for tracking. Google Analytics will

not acknowledge a static version of UTM parameter.

More to the point, by replacing dynamic parameters with static

URLs for things like pagination, onsite search box results or sorting

does not address duplicate content, crawl budget or internal link

equity dilution.

And having all the combinations of filters from your faceted

navigation as indexable URLs often results in thin content issues.

Especially if you offer multi-select filters.

Many SEO pros argue it’s possible to provide the same user

experience without impacting the URL. For example, by using

POST rather than GET requests to modify the page content. Thus,

preserving the user experience and avoiding the SEO problems.

But stripping out parameters in this manner would remove the

possibility for your audience to bookmark or share a link to that

specific page. And if obviously not feasible for tracking parameters

and not optimal for pagination.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

The crux of the matter is that for many websites, completing

avoiding parameters is simply not possible if you want to provide the

ideal user experience. Nor would it be best practice SEO.

So we are left with this. For parameters that you don’t want to be

indexed in search results (paginating, reordering, tracking, etc)

implement as query strings. For parameters that you do want to be

indexed, use static URL paths.

Pros:

Shifts crawler focus from

parameter based to static

URLs which have a higher

likelihood to rank.

Cons:

Significant investment of

development time for URL

rewrites and 301 redirects.

Doesn’t prevent duplicate

content issues.

Doesn’t consolidate ranking

signals.

Not suitable for all

parameter types.

May lead to thin content

issues.

Doesn’t always provide a

linkable or bookmarkable

URL.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

So which of these six SEO tactics should you implement?

The answer can’t be all of them.

Not only would that create unnecessary complexity. But often the

SEO solutions actively conflict with one another.

For example, if you implement robots.txt disallow, Google would not

be able to see any meta noindex tag. You also shouldn’t combine

a meta noindex tag with a rel=canonical link attribute.

What becomes clear is there is no one perfect solution.

Even Google’s John Mueller can’t decide on an approach. In a

Google Webmaster hangout, he initially recommended against

disallowing parameters, but when questioned on this from a faceted

navigation perspective, answered “it depends.”

There are occasions when crawling efficiency is more important

than consolidating authority signals.

Ultimately, what’s right for your website will depend on your

priorities.

Best Practice URL Parameter

Handling for SEO

AN SEO GUIDE TO URL PARAMETER HANDLING

19

Personally, I don’t use noindex or block access to parameter pages.

If Google can’t crawl and understand all the URL variables, it can’t

consolidate the ranking signals to the canonical page.

I take the following plan of attack for SEO-friendly

parameter handling:

Do keyword research to understand what parameters should be

search engine friendly, static URLs.

Implement correct pagination handling with rel=”next &

rel=”prev”.

For all remaining parameter based URLs, implement consistent

ordering rules, which use keys only once and prevent empty values

to limit the number of URLs.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

Add a rel=canonical link attribute to suitable parameter pages to

combine ranking ability.

Configure URL parameter handling in both Google and Bing as

a failsafe to help search engines understand each parameter’s

function.

Double check no parameter based URLs are being submitted in

the XML sitemap.

No matter what parameter handling strategy you choose to

implement, be sure to document the impact of your efforts on

KPIs.

AN SEO GUIDE TO URL PARAMETER HANDLING

19

How to Perform an

In-Depth Technical SEO

Audit

Chapter 20

Written By

Assistant Editor, Search Engine Journal

Anna Crowe

20

I’m not going to lie:

Conducting an in-depth SEO

audit is a major deal.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

And, as an SEO consultant, there are a few sweeter words than, “Your

audit looks great! When can we bring you onboard?”

Even if you haven’t been actively looking for a new gig, knowing your SEO

audit nailed it is a huge ego boost.

But, are you terrified to start? Is this your first SEO audit? Or, you just

don’t know where to begin? Sending a fantastic SEO audit to a potential

client puts you in the best possible place.

20

It’s a rare opportunity for you to organize your processes and rid

your potential client of bad habits (cough*unpublishing pages

without a 301 redirect*cough) and crust that accumulates like the

lint in your dryer.

So take your time. Remember: Your primary goal is to add value to

your customer with your site recommendations for both the shortterm

and the long-term.

Ahead, I’ve put together the need-to-know steps for conducting

an SEO audit and a little insight to the first phase of my processes

when I first get a new client. It’s broken down into sections below. If

you feel like you have a good grasp on a particular section, feel free

to jump to the next.

This is a series, so stay tuned for more SEO audit love.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Jump to:

When Should I Perform an SEO Audit?

What You Need from a Client Before an SEO Audit

Tools for SEO Audit

Technical > DeepCrawl

Technical > Screaming Frog

Technical > Google Search Console & Bing Webmaster

Tools

Technical > Google Analytics

After a potential client sends me an email expressing interest in

working together and they answer my survey, we set-up an intro call

(Skype or Google Hangouts is preferred).

Before the call, I do my own mini quick SEO audit (I invest at least

one hour to manually researching) based on their survey answers

to become familiar with their market landscape. It’s like dating

someone you’ve never met.

You’re obviously going to stalk them on Facebook, Twitter, Instagram,

and all other channels that are public #soIcreep.

Here’s an example of what my survey looks like:

Here are some key questions you’ll want to ask the client

during the first meeting:

1. What are your overall business goals? What are your channel

goals (PR, social, etc.)?

2. Who is your target audience?

3. Do you have any business partnerships?

4. How often is the website updated? Do you have a web developer

or an IT department?

5. Have you ever worked with an SEO consultant before? Or, had

any SEO work done previously?

When Should I Perform an SEO

Audit?

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Sujan Patel also has some great recommendations on questions

to ask a new SEO client.

After the call, if I feel we’re a good match, I’ll send over my formal

proposal and contract (thank you HelloSign for making this an easy

process for me!).

To begin, I always like to offer my clients the first month as a trial

period to make sure we vibe.

This gives both the client and I a chance to become friends first

before dating. During this month, I’ll take my time to conduct an indepth

SEO audit.

These SEO audits can take me anywhere from 40 hours to 60 hours

depending on the size of the website.

These audits are bucketed into three separate parts and

presented with Google Slides.

Technical: Crawl errors, indexing, hosting, etc.

Content: Keyword research, competitor analysis, content maps,

meta data, etc.

Links: Backlink profile analysis, growth tactics, etc.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

After that first month, if the client likes my work, we’ll begin

implementing the recommendations from the SEO audit. And going

forward, I’ll perform a mini-audit monthly and an in-depth audit

quarterly.

To recap, I perform an SEO audit for my clients:

First month

Monthly (mini-audit)

Quarterly (in-depth audit)

When a client and I start working together, I’ll share a Google doc

with them requesting a list of passwords and vendors.

This includes:

Google Analytics access and any third-party analytics tools

Google and Bing ads

Webmaster tools

Website backend access

Social media accounts

List of vendors

List of internal team members (including any work they outsource)

What You Need from a Client Before

an SEO Audit

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Before you begin your SEO audit, here’s a recap of the

tools I use:

Screaming Frog

Integrity (for Mac users) and Xenu Sleuth (for PC users)

SEO Browser

Wayback Machine

Moz

Buzzsumo

DeepCrawl

Copyscape

Google Tag Manager

Google Tag Manager Chrome Extension

Annie Cushing’s Campaign Tagging Guide

Google Analytics (if given access)

Google Search Console (if given access)

Bing Webmaster Tools (if given access)

You Get Signal

Pingdom

PageSpeed Tool

Sublime Text

Tools for SEO Audit

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Technical

Tools needed for technical SEO audit:

Screaming Frog

DeepCrawl

Copyscape

Integrity for Mac (or Xenu Sleuth for PC users)

Google Analytics (if given access)

Google Search Console (if given access)

Bing Webmaster Tools (if given access)

My 30-Point Technical SEO

Checklist

Tools:

DeepCrawl

Copyscape

Screaming Frog

Google Analytics

Integrity

Google Tag Manager

Google Analytics code

Step 1: Add Site to DeepCrawl and

Screaming Frog

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

What to Look When Using DeepCrawl

The first thing I do is add my client’s site to DeepCrawl. Depending

on the size of your client’s site, the crawl may take a day or two to get

the results back.

Once you get your DeepCrawl results back, here are the things I look

for:

Duplicate Content

Check out the “Duplicate Pages” report to locate duplicate content.

If duplicate content is identified, I’ll make this a top priority in my

recommendations to the client to rewrite these pages and in the

meantime, I’ll add the <meta name=”robots” content=”noindex,

nofollow”> tag to the duplicate pages.

Common duplicate content errors you’ll discover:

Duplicate meta titles and meta descriptions

Duplicate body content from tag pages (I’ll use Copyscape to help

determine if something is being plagiarized).

Two domains (ex: yourwebsite.co, yourwebsite.com)

Subdomains (ex: jobs.yourwebsite.com)

Similar content on a different domain

Improperly implemented pagination pages (see below.)

How to fix:

Add the canonical tag on your pages to let Google know what you

want your preferred URL to be.

Disallow incorrect URLs in the robots.txt.

Rewrite content (including body copy and meta data).

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

These are the steps I took to fix the issue:

I fixed any 301 redirect issues.

Added a canonical tag to the page, I want Google to crawl.

Update the Google Search Console parameter settings to exclude

any parameters that don’t generate unique content.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Here’s an example of a duplicate content issue I had with a client of

mine. As you can see below, they had URL parameters without the

canonical tag.

Added the disallow function to the robots.txt to the incorrect URLs to

improve crawl budget

Pagination

There are two reports to check out:

First Pages: To find out what pages are using pagination, review

the “First Pages” report. Then, you can manually review the pages

using this on the site to discover if pagination is implemented

correctly.

Unlinked Pagination Pages: To find out if pagination is working

correctly, the “Unlinked Pagination Pages” report will tell you if the

rel=”next” and rel=”prev” are linking to the previous and next pages.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

In this example below, I was able to find that a client had reciprocal

pagination tags using DeepCrawl:

How to fix:

If you have a “view all” or a “load more” page, add rel=”canonical”

tag. Here’s an example from Crutchfield:

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

If you have all your pages on separate pages, then add the

standard rel=”next” and rel=”prev” markup. Here’s an example

from Macy’s:

Max Redirections

Review the “Max Redirections” report to see all the pages that

redirect more than 4 times. John Mueller mentioned in 2015 that

Google can stop following redirects if there are more than five.

While some people refer to these crawl errors as eating up the

“crawl budget,” Gary Illyes refers to this as “host load”. It’s important

to make sure your pages render properly because you want your

host load to be used efficiently.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Here’s a brief overview of the response codes you might

see:

301 — These are the majority of the codes you’ll see throughout

your research. 301 redirects are okay as long as there are only

one redirect and no redirect loop.

302 — These codes are okay, but if left longer than 3 months

or so, I would manually change them to 301s so that they are

permanent. This is an error code I’ll see often with e-commerce

sites when a product is out of stock.

400 — Users can’t get to the page.

403 — Users are unauthorized to access the page.

404 — The page is not found (usually meaning the client deleted

a page without a 301 redirect).

500 — Internal server error that you’ll need to connect with the

web development team to determine the cause.

How to fix:

Remove any internal links pointing to old 404 pages and update

them with the redirected page internal link.

Undo the redirect chains by removing the middle redirects. For

example, if redirect A goes to redirect B, C, and D, then you’ll want

to undo redirects B and C. The final result will be a redirect A to D.

There is also a way to do this in Screaming Frog and Google

Search Console below if you’re using that version.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

What to Look For When Using Screaming Frog

The second thing I do when I get a new client site is to add their

URL to Screaming Frog.

Depending on the size of your client’s site, I may configure the

settings to crawl specific areas of the site at a time.

Here is what my Screaming Frog spider configurations look like:

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

You can do this in your spider settings or by excluding areas of the

site.

Once you get your Screaming Frog results back, here are the

things I look for:

Google Analytics Code

Screaming Frog can help you identify what pages are missing the

Google Analytics code (UA-1234568-9).

To find the missing Google Analytics code, follow these

steps:

Go to ‘Configuration’ in the navigation bar, then Custom.

Add analytics\.js to Filter 1, then change the drop down to ‘Does

not contain.’

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

How to fix:

Contact your client’s developers and ask them to add the code to

the specific pages that it’s missing.

For more Google Analytics information, skip ahead to that Google

Analytics section below.

Google Tag Manager

Screaming Frog can also help you find out what pages

are missing the Google Tag Manager snippet with similar

steps:

Go to the ‘Configuration’ tab in the navigation bar, then Custom.

Add <iframe src-“//www.googletagmanager.com/ with ‘Does not

contain’ selected in the Filter.

How to fix:

Head over to Google Tag Manager to see if there are any errors

and update where needed.

Share the code with your client’s developer’s to see if they can

add it back to the site.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Schema

You’ll also want to check if your client’s site is using schema

markup on their site. Schema or structured data helps search

engines understand what a page is on the site.

To check for schema markup in Screaming Frog, follow

these steps:

Go to the ‘Configuration’ tab in the navigation bar, then ‘Custom.’

Add itemtype=”http://schema.\.org/ with ‘Contain’ selected in

the Filter.

Indexing

You want to determine how many pages are being

indexed for your client, follow this in Screaming Frog:

After your site is done loading in Screaming Frog, go to

Directives > Filter > Index to review if there are any missing

pieces of code.

How to fix:

If the site is new, Google may have no indexed it yet.

Check the robots.txt file to make sure you’re not disallowing

anything you want Google to crawl.

Check to make sure you’ve submitted your client’s sitemap to

Google Search Console and Bing Webmaster Tools.

Conduct manual research (seen below).

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Flash

Google announced this year that Chrome will start blocking Flash

due to the slow page load times. So, if you’re doing an audit, you

want to identify if your new client is using Flash or not.

To do this in Screaming Frog, try this:

Head to the ‘Spider Configuration’ in the navigation.

Click ‘Check SWF.’

Filter the ‘Internal’ tab by ‘Flash’ after the crawl is done.

How to fix:

Embed videos from YouTube. Google bought YouTube in 2006,

no-brainer here.

Or, opt for HTML5 standards when adding a video.

Here’s an example of HTML5 code for adding a video:

<video controls=”controls” width=”320” height=”240”>&gt;

<source class=”hiddenSpellError” data-mce-bogus=”1” />src=”/

tutorials/media/Anna-Teaches-SEO-To-Small-Businesses.mp4”

type=”video/mp4”&gt;

<source src=”/tutorials/media/Anna-Teaches-SEO-To-Small-

Businesses.ogg” type=”video/ogg” />

Your browser does not support the video tag.</video>

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Javascript

According to Google’s announcement in 2015, JavaScript is

okay to use for your website as long as you’re not blocking anything

in your robots.txt (we’ll dig into this deeper in a bit!). But, you still

want to take a peek at how the Javascript is being delivered to your

site.

How to fix:

Review Javascript to make sure it’s not being blocked by robots.txt

Make sure Javascript is running on the server (this helps produce

plain text data vs dynamic).

If you’re running Angular JavaScript, check out this article by Ben

Oren on why it might be killing your SEO efforts.

In Screaming Frog, go to the Spider Configuration in the navigation

bar and click ‘Check JavaScript.’ After the crawl is done, filter your

results on the ‘Internal’ tab by ‘JavaScript.’

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Robots.txt

When you’re reviewing a robots.txt for the first time, you want to

look to see if anything important is being blocked or disallowed.

For example, if you see this code:

User-agent: *

Disallow: /

Your client’s website is blocked from all web crawlers.

But, if you have something like Zappos robots.txt file, you should be

good to go.

# Global robots.txt as of 2012-06-19

User-agent: *

Disallow: /bin/

Disallow: /multiview/

Disallow: /product/review/add/

Disallow: /cart

Disallow: /login

Disallow: /logout

Disallow: /register

Disallow: /account

They are only blocking what they do not want web crawlers to

locate. This content that is being blocked is not relevant or useful to

the web crawler.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

How to fix:

Your robots.txt is case-sensitive so update this to be all lowercase.

Remove any pages listed as Disallow that you want the search

engines to crawl.

Screaming Frog by default will not be able to load any URLs

disallowed by robots.txt. If you choose to switch up the default

settings in Screaming Frog, it will ignore all the robots.txt.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

You can also view blocked pages in Screaming Frog under the

‘Response Codes’ tab, then filtered by ‘Blocked by Robots.txt’

filter after you’ve completed your crawl.

If you have a site with multiple subdomains, you should have a

separate robots.txt for each.

Make sure the sitemap is listed in the robots.txt.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Crawl Errors

I use DeepCrawl, Screaming Frog, and Google and Bing

webmaster tools to find and cross-check my client’s crawl errors.

To find your crawl errors in Screaming Frog, follow these

steps:

After the crawl is complete, go to ‘Bulk Reports.’

Scroll down to ‘Response Codes,’ then export the server side

error report and the client error report.

How to fix:

The client error reports, you should be able to 301 redirect the

majority of the 404 errors in the backend of the site yourself.

The server error reports, collaborate with the development team

to determine the cause. Before fixing these errors on the root

directory, be sure to backup the site. You may simply need to

create a new .html access file or increase PHP memory limit.

You’ll also want to remove any of these permanent redirects

from the sitemap and any internal or external links.

You can also use ‘404’ in your URL to help track in Google

Analytics.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Redirect Chains

Redirect chains not only cause poor user experience, but it slows

down page speed, conversion rates drop, and any link love you may

have received before is lost.

Fixing redirect chains is a quick win for any company.

How to fix:

In Screaming Frog after you’ve completed your crawl, go to

‘Reports’ > ‘Redirect Chains’ to view the crawl path of your

redirects. In an excel spreadsheet, you can track to make sure

you’re 301 redirects are remaining 301 redirects. If you see a 404

error, you’ll want to clean this up.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Internal & External Links

When a user clicks on a link to your site and gets a 404 error, it’s

not a good user experience.

And, it doesn’t help your search engines like any better either.

To find my broken internal and external links I use Integrity for

Mac. You can also use Xenu Sleuth if you’re a PC user.

I’ll also show you how to find these internal and external links in

Screaming Frog and DeepCrawl if you’re using that software.

How to fix:

If you’re using Integrity or Xenu Sleuth, run your client’s site URL

and you’ll get a full list of broken URLs. You can either manually

update these yourself or if you’re working with a dev team, ask

them for help.

If you’re using Screaming Frog, after the crawl is completed, go

to ‘Bulk Export’ in the navigation bar, then ‘All Outlinks.’ You can

sort by URLs and see which pages are sending a 404 signal.

Repeat the same step with ‘All Inlinks.’

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

If you’re using DeepCrawl, go to the ‘Unique Broken Links’ tab

under the ‘Internal Links’ section.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

URLs

Every time you take on a new client, you want to review

their URL format. What am I looking for in the URLs?

Parameters – if the URL as weird characters like ?, =, or + it’s a

dynamic URL which can cause duplicate content if not optimized.

User-friendly – I like to keep the URLs short and simple while also

removing any extra slashes.

How to fix:

You can search for parameter URLs in Google by doing site:www.

buyaunicorn.com/ inurl: “?” or whatever you think the parameter

might include.

After you’ve run the crawl on Screaming Frog, take a look at

URLs. If you see parameters listed that are creating duplicates of

your content, you need to suggest the following:

Add a canonical tag to the main URL page. For example, www.

buyaunicorn.com/magical-headbands is the main page and I see

www.buyaunicorn.com/magical-headbands/?dir=mode123$,

then the canonical tag would need to be added to www.

buyaunicorn.com/magical-headbands.

Update your parameters in Google Search Console under ‘Crawl’

> ‘URL Parameters.’

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Disallow the duplicate URLs in the robots.txt.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Tools:

Google Search Console

Bing Webmaster Tools

Sublime Text (or any text editor tool)

Set a Preferred Domain

Since the Panda update, it’s beneficial to clarify to the search

engines the preferred domain. It also helps make sure all your links

are giving one site the extra love instead of being spread across

two sites.

How to fix:

In Google Search Console, click the gear icon in the upper right

corner.

Choose which of the URLs is the preferred domain.

You don’t need to set the preferred domain in Bing Webmaster

Tools, just submit your sitemap to help Bing determine your

preferred domain.

Step 2: Review Google Search

Console and Bing Webmaster

Tools

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Backlinks

With the announcement that Penguin is real-time, it’s vital that

your client’s backlinks meet Google’s standards. If you notice a

large chunk of backlinks coming to your client’s site from one

page on a website, you’ll want to take the necessary steps to

clean it up, and FAST!

How to fix:

In Google Search Console, go to ‘Links’ > then sort your ‘Top

linking sites.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Here’s an example of what my disavow file looks like:

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Contact the companies that are linking to you from one page to

have them remove the links.

Or, add them to your disavow list. When adding companies to your

disavow list, be very careful how and why you do this. You don’t

want to remove valuable links.

Keywords

As an SEO consultant, it’s my job to start to learn the market

landscape of my client. I need to know who their target audience

is, what they are searching for, and how they are searching.

To start, I take a look at the keyword search terms they

are already getting traffic from.

In Google Search Console, under ‘Search Traffic’ > ‘Search

Analytics’ will show you what keywords are already sending your

client clicks.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Sitemap

Sitemaps are essential to get search engines to crawl your

client’s website. It speaks their language.

When creating sitemaps, there are a few things to know:

Do not include parameter URLs in your sitemap.

Do not include any non-indexable pages.

If the site has different subdomains for mobile and desktop, add

the rel=”alternate” tag to the sitemap.

How to fix:

Go to ‘Google Search Console’ > ‘Index’ > ‘Sitemaps’ to compare

the URLs indexed in the sitemap to the URLs in the web index.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Then, do a manual search to determine pages are not getting

indexed and why.

If you find old redirected URLs in your client’s sitemap, remove

them. These old redirects will have an adverse impact on your

SEO if you don’t remove them.

If the client is new, submit a new sitemap for them in both Bing

and Google webmaster tools. HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Crawl

Crawl errors are important to check because it’s not only bad for

the user but it’s bad for your website rankings. And, John Mueller

stated that low crawl rate may be a sign of a low-quality site.

To check this in Google Search Console, go to ‘Coverage’ >

‘Details.’ HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

To check this in Bing

Webmaster Tools, go to

‘Reports & Data’ > ‘Crawl

Information.

How to fix:

Manually check your

crawl errors to determine

if there are crawl

errors coming from old

products that don’t exist

anymore or if you see

crawl errors that should

be disallowed in the

robots.txt file.

Once you’ve determined

where they are coming

from, you can implement

301 redirects to similar

pages that link to the

dead pages.

You’ll also want to crosscheck

the crawl stats in

Google Search Console

with average load time in

Google Analytics to see if there is a correlation between time

spent downloading and the pages crawled per day.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Structured Data

As mentioned above in the schema section of Screaming Frog,

you can review your client’s schema markup in Google Search

Console.

Use the individual rich results status report in Google Search

Console. (Note: The structured data report is no longer available).

This will help you determine what pages have structured data

errors that you’ll need to fix down the road.

How to fix:

Google Search Console will tell you what is missing in the

schema when you test the live version.

Based on your error codes, rewrite the schema in a text editor

and send to the web development team to update. I use Sublime

Text for my text editing. Mac users have one built-in and PC users

can use TextPad.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Tools:

Google Analytics

Google Tag Manager Assistant Chrome Extension

Annie Cushing Campaign Tagging Guide

Step 3: Review Google Analytics

Views

When I first get a new client, I set up 3 different views in Google

Analytics.

Reporting view

Master view

Test view

These different views give me the flexibility to make changes

without affecting the data.

How to fix:

In Google Analytics, go to ‘Admin’ > ‘View’ > ‘View Settings’ to

create the three different views above.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Make sure to check the ‘Bot

Filtering’ section to exclude all hits

from bots and spiders.

Link AdWords and Google Search

Console.

Lastly, make sure the ‘Site search

Tracking’ is turned on.

You want to make sure you add

your IP address and your client’s

IP address to the filters in Google

Analytics so you don’t get any false

traffic.

How to fix:

Go to ‘Admin’> ’View’ > ‘Filters’

Then, the settings should be set

to ‘Exclude’ > ‘traffic from the IP

addresses > ‘that are equal to.’

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Tracking Code

You can manually check the source code, or you can use my

Screaming Frog technique from above.

If the code is there, you’ll want to track that it’s firing realtime.

To check this, go to your client’s website and click around a bit on

the site.

Then go to Google Analytics > ‘Real-Time’ > ‘Locations,’ your

location should populate.

If you’re using Google Tag Manager, you can also check this with

the Google Tag Assistant Chrome extension.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

How to fix:

If the code isn’t firing, you’ll want to check the code snippet to

make sure it’s the correct one. If you’re managing multiple sites,

you may have added a different site’s code.

Before copying the code, use a text editor, not a word processor

to copy the snippet onto the website. This can cause extra

characters or whitespace.

The functions are case-sensitive so check to make sure

everything is lowercase in code.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Indexing

If you had a chance to play around in Google Search Console, you

probably noticed the ‘Coverage’ section. When I’m auditing a client,

I’ll review their indexing in Google Search Console compared to

Google Analytics.

Here’s how:

In Google Search Console, go to ‘Coverage’

In Google Analytics, go to ‘Acquisition’ > ‘Channels’ > ‘Organic

Search’ > ‘Landing Page.’

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Once you’re here, go to ‘Advanced’ > ‘Site Usage’ > ‘Sessions’ > ‘9.’

How to fix:

Compare the numbers from Google Search Console with the

numbers from Google Analytics, if the numbers are widely different,

then you know that even though the pages are getting indexed only

a fraction are getting organic traffic.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Campaign Tagging

The last thing you’ll want to check in Google Analytics is if your

client is using campaign tagging correctly. You don’t want to not

get credit for the work you’re doing because you forgot about

campaign tagging.

How to fix:

Set up a campaign tagging strategy for Google Analytics and

share it with your client. Annie Cushing put together an awesome

campaign tagging guide.

Set up Event Tracking if your client is using mobile ads or video.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Keywords

You can use Google Analytics to gain

insight into potential keyword gems for

your client.

To find keywords in Google Analytics,

follow these steps:

Go to Google Analytics > ‘Behavior’ > ‘Site

Search’ > ‘Search Terms.’ This will give you

a view of what customers are searching for

on the website.

Next, I’ll use those search terms to create a

‘New Segment’ in Google Analytics to see

what pages on the site are already ranking

for that particular keyword term.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Tools:

Google Analytics

Access to client’s server and host

You Get Signal

Pingdom

PageSpeed Tools

Wayback Machine

1 Version of Your Client’s Site is Searchable

Check all the different ways you could search for a website.

For example:

http://annaisaunicorn.com

https://annaisaunicorn.com

http://www.annaisaunicorn.com

As Highlander would say, “there can be only one” website that is

searchable.

How to fix:

Use a 301 redirect for all URLs that are not the primary site to the

canonical site.

Step 4: Manual Check

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Indexing

Conduct a manual search in Google and Bing to determine how

many pages are being indexed by Google. This number isn’t always

accurate with your Google Analytics and Google Search Console

data, but it should give you a rough estimate.

To check, do the following:

Perform a site search in the search engines.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

When you search, manually scan to make sure only your client’s

brand is appearing.

Check to make sure the homepage is on the first page. John

Mueller said it isn’t necessary for the homepage to appear as

the first result.

How to fix:

If another brand is appearing in the search results, you have a

bigger issue on your hands. You’ll want to dive into the analytics

to diagnose the problem.

If the homepage isn’t appearing as the first result, perform a

manual check of the website to see what it’s missing. This could

also mean the site has a penalty or poor site architecture which

is a bigger site redesign issue.

Cross-check the number of organic landing pages in Google

Analytics to see if it matches the number of search results you

saw in the search engine. This can help you determine what

pages the search engines see as valuable.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Caching

I’ll run a quick check to see if the top pages are being cached

by Google. Google uses these cached pages to connect your

content with search queries.

To check if Google is caching your client’s pages, do this:

http://webcache.googleusercontent.com/

search?q=cache:https://www.searchenginejournal.com/

pubcon-day-3-women-in-digital-amazon-analytics/176005/

Make sure to toggle over to the ‘Text-only version.’

You can also check this in Wayback Machine.

How to fix:

Check the client’s server to see if it’s down or operating

slower than usual. There might be an internal server error or a

database connection failure. This can happen if multiple users

are attempting to access the server at once.

Check to see who else is on your server with a reverse IP

address check. You can use You Get Signal website for this

phase. You may need to upgrade your client’s server or start

using a CDN if you have sketchy domains sharing the server.

Check to see if the client is removing specific pages from the

site.

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

Hosting

While this may get a little technical for some, it’s vital to your SEO

success to check the hosting software associated to your client’s

website. Hosting can harm SEO and all your hard work will be for

nothing.

You’ll need access to your client’s server to manually check any

issues. The most common hosting issues I see are having the

wrong TLD and slow site speed.

How to fix:

If your client has the wrong TLD, you need to make sure the

country IP address is associated with the country your client is

operating in the most. If your client has a .co domain and also a

.com domain, then you’ll want to redirect the .co to your client’s

primary domain on the .com.

If your client has slow site speed, you’ll want to address this

quickly because site speed is a ranking factor. Find out what

is making the site slow with tools like PageSpeed Tools and

Pingdom. Here’s a look at some of the common page speed

issues:

Host

Large images

Embedded videos

Plugins

Ads

Theme

Widgets

Repetitive script or dense code

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20

I’m excited to see you test out DeepCrawl, Screaming Frog, and

some of the other tools. And, I’d love to hear about all the creative

ways you perform a site audit. What have you experimented with?

What tools do you use? Let me know if the comments below.

This is a series of posts which I’ll be diving deeper into mobile, site

architecture, site speed, content, and off-site. If there’s anything

particular you want to see, let me know if the comments.

Over to You!

HOW TO PERFORM AN IN-DEPTH TECHNICAL SEO AUDIT

20


No comments:

Post a Comment

News

Followers

Home | About us | Disclaimer | Privacy Policy | Contact us