eCommerce Search Engine Optimization – The Ultimate On-Site SEO Guide…

As bloggers and Store Owners we want to make our websites appeal to both search engines and users.
A search engine ‘looks’ at a website quite differently to
us humans. Your website design, fancy graphics and well written blog
posts count for nothing if you have missed out the critical on-site SEO
covered in this post.
SEO is not dead.
Modern SEO is not about fooling the search engines – it is
about working with the search engines to present your products and
content in the best possible way.
In this post, SEO Expert -Itamar Gero shows us exactly how…
The Ultimate On-Site SEO Guide for store owners and bloggers
Ecommerce is huge business. According to the United States Census Bureau,
ecommerce sales in the US alone hit $394.9 billion in 2016. The rising
trend in eCommerce sales and its influence on offline retail is expected
to continue well into the 2020s worldwide.
Understandably, online store owners are
jockeying for position to take ever-bigger pieces of the eCommerce pie.
To that end, SEO has always been a top priority for most eCommerce
marketing strategies. Search engines drive the most qualified and
motivated traffic that’s easier to convert into paying customers than
visitors from other channels. It’s why websites are waging a constant
“war” behind the scenes to outrank each other for keywords related to
their products and services.
And while having lots of competitors is
daunting, you can take comfort in the fact that a lot of eCommerce sites
aren’t doing their SEO right. With a solid grasp of SEO best practices
and lots of hard work, you can gain high search visibility without
breaking the bank.
SEO Guide To Your Online Store – Remove The Headache!

Special Note about this SEO Guide:
This SEO guide and tutorial is one of the most in depth we have ever produced. [Over 7000 words and 20+ images]
It is easy to be overwhelmed and to feel muddled with SEO
but in reality it is much more straightforward that it may initially
appear.
The rewards for getting Search Engine Optimization right are HUGE!
In addition while this guide is particularly focused on
eCommerce SEO, much of what we detail here is equally applicable to your
blog or company website.
Two SEO resources you will need:
What is Ecommerce SEO?
Simply put, ecommerce SEO is the process
of optimizing an online store for greater visibility on search engines.
It has four main facets, including:
- Keyword research
- On-site SEO
- Link Building
- Usage Signal Optimization
In this post, I’ll tackle the most
foundational and arguably the most crucial among the four areas: on-site
SEO. In our experience, working with 1000s of agencies, we can
attribute the greatest impact on overall organic traffic growth to optimizations that we did within the ecommerce sites we handle.
While link building and other off-page SEO activities are important, on-site SEO sets the tone for success each and every time.
The Resurgence of On-Site SEO
On-site SEO is the collective term used
to describe all SEO activities done within the website. It can be
divided into two segments: technical SEO and on-page SEO. Technical SEO
mostly deals with making sure that the site stays up, running and
available for search bot crawls. On-page SEO, on the other hand, is
geared more towards helping search engines understand the contextual
relevance of your site to the keywords you’re targeting.
I’m focusing on on-site SEO today because
of the undeniable resurgence it’s been having during the past 3 years
or so. You see, the SEO community went through a phase in the late 2000s
to around 2011 when everyone was obsessed with the acquisition of
inbound links. Back then, links were far and away the most powerful
determinant of Google rankings. Ecommerce sites that were locked in
brutal ranking battles were at the core of this movement and competition
eventually revolved around the matter of who’s able to get the most links – ethically or otherwise.
Eventually, Google introduced the Panda
and Penguin updates which punished a lot of sites that proliferated link
and content spam. Quality links that boosted rankings became harder to
come by, making on-site SEO signal influence more pronounced over the
rankings. The SEO community soon started seeing successful ecommerce SEO
campaigns that focused more on technical and on-page optimization than
heavy link acquisition.
This post will show you what you can do on your site to take advantage of the on-site SEO renaissance:
Technical SEO
As mentioned earlier, technical SEO is
mostly about making sure your site has good uptime, loads fast, offers
secure browsing to users and facilitates good bot and user navigation
through its pages. Here’s a list of things you need to monitor
constantly to ensure a high level of technical health:
The Robots.txt File
The robots.txt file is a very small
document that search bots access when they visit your site. This
document tells them which pages can be accessed, which bots are welcome
to do so and which pages are off-limits. When robots.txt disallows
access to a certain page or directory path within a website, bots from
responsible sites will adhere to the instruction and not visit that page
at all. That means disallowed pages will not be listed in search
results. Whatever link equity flowing to them is nullified and these
pages will not be able to pass any link equity as well.
When checking your robots.txt file, make
sure that all pages meant for public viewing don’t fall under any
disallow parameters. Similarly, you’ll want to make sure that pages
which don’t serve the intent if your target audience are barred from
indexing.
Having more pages indexed by search
engines may sound like a good thing, but it really isn’t. Google and
other web portals constantly try to improve the quality of listings
displayed in their SERPs. (search engine results pages)
Therefore, they expect webmasters to be judicious in the pages they submit for indexing and they reward those who comply.
In general, search engine queries fall under one of three classifications:
- Navigational
- Transactional
- Informational
If your pages don’t satisfy the intents
behind any of these, consider using robots.txt to prevent bot access to
them. Doing so will make better use of your crawl budget and make your
site’s internal link equity flow to more important pages. In an
ecommerce site, the types of URLs that you usually want to bar access to
are:
- Checkout pages – These are series of pages that shoppers use to choose and confirm their purchases. These pages are unique to their sessions and therefore, are of no interest to anyone else on the Internet.
- Dynamic pages – These pages are created through unique user requests such as internal searches and page filtering combinations. Like checkout pages, these pages are generated for one specific user who made the request. Therefore, they’re of no interest to most people on the Web, making the impetus for search engines to index them very weak. Further, these pages eventually expire and send out 404 Not Found responses when re-crawled by search engines. That can be taken as a signal of poor site health that can negatively impact an online store’s search visibility.
Dynamic pages can easily be identified by
the presence of the characters “?” and “=” in their URLs. You can
prevent them from being indexed by adding a line in the robots.txt file
that says something like this: disallow: *?
- Staging Pages – These are pages that are currently under development and unfit for public viewing. Make sure to set up a path in your site’s directory specifically for staging webpages and make sure the robots.txt file is blocking that directory.
- Backend pages – These pages are for site administrators only. Naturally, you’ll want the public not to visit the pages – much less find them in search results. Everything from your admin login page down to the internal site control pages must be placed under a robots.txt restriction to prevent unauthorized entry.
Note that the robots.txt file isn’t the
only way to restrict the indexing of pages. The noindex meta directive
tag, among others, can also be used for this purpose. Depending on the
nature of the deindexing situation, one may be more appropriate than the
other.
The XML Sitemap
The XML sitemap is another document that
search bots read to get useful information. This file lists all the
pages that you want Google and other spiders to crawl. A good sitemap
contains information that gives bots an idea of your information
architecture, the frequency at which each page is modified and the
whereabouts of assets, such as images, within your domain’s file paths.
While XML sitemaps are not a necessity in
any website, they’re very important to online stores due to the number
of pages that a typical ecommerce site has. With a sitemap in place and
submitted to tools like Google Search Console, search engines tend to find and index pages that are deep within your site’s hierarchy of URLs.
Your web developer should be able to set
up an XML sitemap for your ecommerce site. More often than not,
ecommerce sites already have it by the time their development cycles are
finished. You can check this by going to
[www.yoursite.com/sitemap.xml]. If you see something like this, you
already have your XML sitemap up and running:

Having an XML Sitemap isn’t a guarantee that all the URLs listed on it will be considered for indexing.
Submitting the sitemap to Google Search Console
ensures that the search giant’s bot finds and reads the sitemap. To do
this, simply log in to your Search Console account and find the property
you want to manage. Go to Crawl>Sitemaps and click the “Add/Test New Sitemap” button on the upper right. Just enter the URL slug of your XML sitemap and click “Submit.”
You should be able to see data on your sitemap in 2-4 days. This is what it will look like:

Notice that the report tells you how many
pages are submitted (listed) in the sitemap and how many Google
indexed. In a lot of cases, ecommerce sites will not get every page they
submit in the sitemap indexed by Google. A page may not be indexed due
to one of several reasons including:
- The URL is Dead – If a page has been deliberately deleted or is afflicted with technical problems, it will likely yield 4xx or 5xx errors. If the URL is listed in the sitemap, Google will not index a page from a URL that is not working properly. Similarly, if a once-live page that’s listed in the sitemap goes down for long periods of time, it may be taken off the Google index.
- The URL is Redirected – When a URL is redirected and is yielding either a 301 or a 302 response code, there’s no sense to have it in the sitemap. The redirect’s target page should instead be listed if it’s not there already. If a redirecting URL is listed in a sitemap, there’s a good chance Google will simply ignore it and report it as not indexed.
- The URL is being Blocked – As discussed under the robots.txt section, not all pages in an ecommerce site need to be indexed. If a webpage is being blocked by robots.txt or the noindex meta tag, there’s no sense listing it on the XML sitemap. Search Console will count it as not being indexed precisely because you asked it not to be.
Checkout pages, blog tag pages and other
product pages with duplicate content are examples of pages that need not
be listed in the XML sitemap.
- The URL has a Canonical Link – The rel=canonical HTML tag is often used in online stores to tell search engines which page they want indexed out of several very similar pages. This often happens when a product has multiple SKUs with very small distinguishing attributes. Instead of having Google choose which one to display on the SERPs, webmasters gained the ability to tell search engines which page is the “real” one that they want featured.
If your ecommerce site has product pages
that have the rel=canonical element, there’s no need to list them on
your sitemap. Google will likely ignore them anyway and honor the one
they’re pointing to.
- The Page has Thin Content – Google defines thin content as pages with little to no added value. Examples include pages with little to no textual content or pages that do have text but are duplicating other pages from within the site or from elsewhere in the web. When Google deems a page as being thin, it either disfavors it in the search results or ignores it outright.
If you have product pages that carry
boilerplate content lifted from manufacturer sites or other pages from
your site, it’s usually smart to block indexing on them until you have
the time and manpower to write richer and more unique descriptions. It
also follows that you should avoid listing these pages on your XML
sitemap just because they’re less likely to be indexed.
- There is a Page-Level Penalty – In rare instances, search engines might take manual or algorithmic actions against sites that violate their quality guidelines. If a page is spammy, or has been hacked and infused with malware, it may be taken off the index. Naturally, you’ll want pages like these off your sitemap.
- The URL is Redundant – Duplicate URLs in the XML sitemap, as you may expect, will not be listed twice. The second one will likely be ignored and left off the index. You can solve this issue by opening your sitemap on your browser and saving it as an XML document that you can open in Excel. From there, go to the Data tab. Highlight the column where the URLs are in your sitemap and click on remove Duplicates.
- Restricted Pages – Pages that are password-protected or are only granting access to specific IPs will not be crawled by search engines and therefore not indexed.
The fewer inappropriate pages you list in
your sitemap, the better your submission to indexing ratio will be.
This helps search engines understand which pages within your domain hold
the highest degrees of importance, allowing them to perform better for
keywords they represent.
Crawl Errors
In online stores, products are routinely
added and deleted depending on a lot of factors. When indexed product or
category pages are deleted, it doesn’t necessarily mean that search
engines automatically forget about them. Bots will continue to attempt
crawls of these URLs for a few months until they’re fixed or taken off
the index due to chronic unavailability.
In technical terms, crawl errors are
pages that bots can’t successfully access because they return HTTP error
codes. Among these codes, 404 is the most common but others in the 4xx
range apply. While search engines recognize that crawl errors are a
normal occurrence in any website, having too many of them can stunt
search visibility. Crawl errors tend to look like loose ends on a site
which can disrupt the proper flow of internal link equity between pages.
Crawl errors are usually caused by the following occurrences:
- Deliberately deleted pages
- Accidentally deleted pages
- Expired dynamic pages
- Server issues
To see how many crawl errors you have in
your ecommerce site and which URLs are affected, you can access Google
Search Console and go to the property concerned. At the left sidebar
menu, go to Crawl>Crawl Errors. You should see a report similar to this:

Depending on the types of pages you find in your crawl error report, there are several distinct ways to tackle them, including:
- Fix Accidentally Deleted Pages – If the URL belongs to a page that was deleted unintentionally, simply re-publishing the page under the same web address will fix the issue.
- Block Dynamic Page Indexing – As recommended earlier, dynamic pages that expire and become crawl errors can be prevented by blocking bot access using robots.txt. If no dynamic page is indexed in the first place, no crawl error will be detected by search engines.
- 301 Redirect the Old Page to the New Page– If a page was deleted deliberately and a new one was published to replace it, use a 301 redirect to lead search bots and human users to the page that took its place. This not only prevents the occurrence of a crawl error, it also passes along any link equity that the deleted page once held. However, don’t assume that the fix for every crawl error is a 301 redirect. Having too many redirects can affect site speed negatively.
- Address Server Issues – If server issues are the root cause of downtime, working with your web developer and your hosting service provider is your best recourse.
- Ignore Them – When pages are deleted deliberately but they’re of no great importance and no replacement page is planned, you can simply allow search engines to flush them out of the index in a few months.
- Having as few crawl errors as possible is a hallmark of responsible online store administration. A monthly check on your crawl error report should allow you to stay on top of things.
SEO Guide To Broken Links
Broken links prevent the movement of search spiders from page to page. They’re also bad for user experience because they lead visitors to dead ends on a site. Due to the volume of pages and the complex information architectures of most ecommerce sites, it’s common for broken links to occur here and there.Broken links are usually caused either by errors in the target URLs of links or by linking to pages that are displaying 404 Not found server response codes. To check your site for crawl errors, you can use a tool called the Screaming Frog SEO Spider. It has a free, limited version and a full version that costs 99GBP per year. For smaller online stores, the free version should suffice. For sites with thousands of pages, you’ll need the paid version for a thorough scan.To check for broken links using Screaming Frog, simply set the app to function on the default Spider mode. Enter the URL of your home page and click the Start button.Wait for the crawl to finish. Depending on the number of pages and your connection speed, the crawl could take several minutes to an hour.When the crawl finishes, go to Bulk Export and click on “All Anchor Text.” You will then have to save the data as a CSV file and open it in Excel. It should look like this:Go to Column F (Status Code) and sort the values from largest to smallest. You should be able to find the broken links on top. In this case, this site only has 4 broken links.Column B (Source) refers to the page where the link can be found and edited. Column C (Destination) refers to the URL where the link is pointing to. Column E (Anchor) pertains to the text where the link on the source page is attached.You can fix broken links through one of the following methods: - Fix the Destination URL – If the destination URL was misspelled, correct the typo and set the link live.
- Remove the Link – If there was no clerical error on the destination URL but the page it used to link to no longer exists, simply remove the hyperlink.
- Replace the Link – If the link points to a page that has been deleted but there’s a replacement page or another page that can sub for it, replace the destination URL in your CMS.
Fixing broken links improves the
circulation of link equity and gives search engines a better impression
of your site’s technical health.
Duplicate Content and SEO for Your Online Store
As mentioned earlier, online stores have a
higher tendency to suffer from content duplication issues due to the
number of products they carry and similarities in the names of their
SKUs. When Google detects enough similarities between pages, it makes
decisions on which pages to show in its search results. Unfortunately,
Google’s choice of pages usually aren’t consistent with yours.
To find pages in your ecommerce site that are possibly affected by duplication problems, go to Google Search Console
and click on Search Appearance in the left sidebar. Click on the HTML
Improvements report and you should see something like this:

The blue linked text indicates that your
site has problems with a specific type of HTML issue. In this case, it’s
duplicate title tags. Clicking on this allows you to see the pages
concerned. You can then export the report to a CSV file and open it in
Excel.
Investigate the report and the URLs
involved. Analyze why title tags and meta descriptions might be
duplicating. In online stores, this could be due to:
- Lazy Title Tag Writing – In some poorly optimized sites, web developers might leave all the pages with the same title tags. Usually, it’s the site’s brand name. This can be addressed by editing the title tags and appropriately naming each page according to the essence of its content.
- Very Similar Products – Some online stores have products that have very similar properties. For example, an ecommerce store that sells garments can sell a shirt that comes in 10 different colors. If each color is treated as a unique SKU and comes with its own page, the title tags and meta descriptions can be very similar.
As mentioned in a previous section, using
the rel=canonical HTML element can point bots to the version of these
pages that you want indexed. It will also help search engines understand
that the duplication in your site is by design.
- Accidental Duplication – In some cases, ecommerce CMS platforms could be misconfigured and run amok with page duplications. If this happens, a web developer’s help in addressing the root cause is necessary. You will likely need to delete the duplicate pages and 301 redirect them to the original.
Bonus: the pages identified as
having short or long title tags and meta descriptions can be dealt with
simply by editing these fields and making sure length parameters are
followed. More on that in the Title Tags and Meta Descriptions section
of this guide.
Site Speed
Over the years, site speed has become one
of the most important ranking factors in Google. For your online store
to reach its full ranking potential, it has to load quickly for both
mobile and desktop devices. In competitive keyword battles between rival
online stores, the site that has the edge in site speed usually
outperforms its slower competitors.
To test how well your site performs in the speed department, go to Google PageSpeed insights. Copy and paste the URL of the page you want to test and hit Enter.

Google will rate the page on a scale of
1-100. 85 and above are the preferred scores. In this example, the site
tested was way below the ideal speed on desktop and mobile. Fortunately,
Google provides technical advice on how to address the load time
issues. Page compression, better caching, minifying CSS files and other
techniques can greatly improve performance. Your web developer and
designer will be able to help you with these improvements.
Secure URLs
A couple of years ago, Google announced
that they’re making secure URLs a ranking factor. This compelled a lot
of ecommerce sites to adopt the protocol in pursuit of organic traffic
gains. While we observed the SEO benefits to be marginal, the true
winners are end users who enjoy more secure shopping experiences where
data security is harder to compromise.

You can verify if your online store uses
secure URLs by checking for the padlock icon in the address bar of your
web browser. If it’s not there, you might want to consider having it
implemented by your dev. To date, most ecommerce sites only have secure
URLs in their checkout pages. However, more and more online store owners
are making the switch to the SSL protocol.
Implementing secure URLs in an ecommerce
site can be a daunting task. Search engines see HTTP and HTTPS
counterpart pages as two different web addresses. Therefore, you’d have
to re-create every page in your site in HTTPS and 301 redirect all the
old HTTP equivalents to make things work. Needless to say, this is a
major decision where SEO is just one of several factors to consider.
On-Page SEO
On-page SEO refers to the process of
optimizing individual pages for greater search visibility. This mainly
involves increasing content quality and making sure keywords are present
in elements of each page where they count.
On-page SEO can be a huge task for big
ecommerce sites but it has to be done at some point. Here are the
elements of each page that you should be looking to tweak:
Canonical URL Slugs
One of the things search engines look at
when gauging a page’s relevance to keywords is its URL slug. Search
engines prefer URLs that contain real words over undecipherable
characters. For instance, a URL like [www.clothes.com/112-656-11455]
might technically work, but search engines would get a better idea on
what it’s about if it looked more like [www.yourshop.com/pants]. URLs
that use real words instead of alphanumeric combinations are referred to
as “canonical URLs.” These are not to be confused with the
earlier-mentioned rel=canonical links, which have nothing to do with URL
slugs.
Most online shopping platforms make ready
use of canonical URLs. However, some might not and this is something
you need to look out for. Your web developer will be able to help you
set up canonical URL slugs if it’s not what your CMS uses out of the
box.
SEO Guide – Meta Directives
Aside from the robots.txt file, there’s
another way to restrict indexing on your pages. This is by using what’s
known as meta directives tags. Simply put, these are HTML instructions
within the <head> part of each page’s source code which tell bots
what they can and can’t do with a page. The most common ones are:
- Noindex – This tag tells search engines not to index a page. Similar to the effects of a robots.txt disallow, a page will not be listed in the SERPs if it has this tag. The difference, however, is the fact that a responsible bot will not crawl a page restricted by robots.txt while a page with noindex will still be crawled – it just won’t be indexed. In that regard, search spiders can still pass through a noindexed page’s links and link equity can still flow through those links.
The noindex tag is best used for excluding blog tag pages, unoptimized category pages and checkout pages.
- Noarchive – This meta tag allows bots to index your page but not to keep a cached version of it in their storage.
- Noodp – This tag tells bots not to list a site in the Open Directory Project (DMOZ)
- Nofollow – This tag tells search engines that the page may be indexed but the links in it should not be followed.
As discussed earlier, it’s best to be
judicious when deciding which pages you should allow Google to index.
Here are a few quick tips on how to handle indexing for common page
types:
- Home Page – Allow indexing.
- Product Category Pages – Allow indexing. As much as possible, add breadth to these pages by enhancing the page’s copy. More on that under the “Unique Category and Product Copy” section.
- Product Pages – Allow indexing only if your pages have unique copy written for them. If you picked up the product descriptions from a catalog or a manufacturer’s site, disallow indexing with the “noindex,follow” meta tag.
- Blog Article Pages – If your online store has a blog, by all means allow search engines to index your articles.
- Blog Category Pages – Allow indexing only if you’ve added unique content. Otherwise, use the “noindex,follow” tag.
- Blog Tag Pages – Use the “noindex,follow” tag.
- Blog Author Archives – Use the “noindex,follow” tag.
- Blog Date Archives – Use the “noindex,follow” tag.
Popular ecommerce platforms such as
Shopify, Magento and WordPress all have built-in functionalities or
plugins that allow web admins to easily manage meta directives in a
page.
In cases where you’re wondering why a
particular page isn’t being indexed or why a supposedly restricted page
is appearing in the SERPs, you can manually check their source codes.
Alternatively, you can audit your entire site’s meta directives list by
running a Screaming Frog crawl.

After running a crawl, simply check the
Meta Robots column. You should be able to see what meta directive tags
each page has in your site. As with any Screaming Frog report, you can
export this to an Excel sheet for easier data management.
Title Tags
The title tag remains the most important
on-page ranking factor to most search engines. This is the text that
headlines search result and its main function is to tell human searchers
and bots what a page is about in 60 characters or less.

Due to the brevity and importance that
title tags innately have, it’s crucial for any ecommerce SEO campaign to
get these right. Well-written title tags must have the following
qualities to maximize a page’s ability to rank:
- 60 characters or less including spaces
- Just give the gist of what the page is about
- Must mention the page’s primary keyword early on
- Must cater to buying intent by mentioning purchase action words
- Optionally,mention the site’s brand
- Uses separators such as “-“ and “|” to demarcate the title tag’s core and the brand name.
Here’s an example of a good title tag:
Buy Denim Jeans Online | Men’sWear.com
We see that the purchase word “Buy” is
present but the product type “Denim Jeans” is still mentioned early on.
The word “Online” is also mentioned to indicate that the purchase can be
made over the Internet and the page won’t just tell the searcher where
he or she can buy physically. The brand of the online store is also
mentioned but the whole text block is still within the 60-character
recommended length.
SEO Guide – Meta Descriptions
Meta descriptions are the text blocks
that you can see under the title tags in search results. Unlike title
tags, these aren’t direct ranking factors. As a matter of fact, you can
get away with not writing them and Google would just pick up text from
the page’s content that it deems most relevant to a query.

That doesn’t mean you shouldn’t care
about meta descriptions, though. When written in a well-phrased and
compelling manner, these can determine whether a user clicks on your
listing or goes to a competitor’s page. The meta description may not be a
ranking factor but a page’s click-through rate (CTR) certainly is. CTR
is an important engagement signal that Google uses to see which search
results satisfy their users’ intent best.
Good meta descriptions have the following qualities:
- Roughly 160 characters in length (including spaces)
- Provides users an idea of what to expect from the page
- Mentions the page’s main keyword at least once
- Makes a short but compelling case on why the user should click on the listing
Here’s a nice sample from an online store which ranks number 1 on Google for the keyword “cheap printer inks.”

Notice how the meta description makes it
clear from the start that they sell the products mentioned in the query
with the phrase “Shop for ink cartridges, printer cartridges and toner
cartridges.” This part also mentions the main keywords that the site
targets. Meanwhile, the phrase “Find cheap printer ink for all brands at
Carrotink.com. Free Shipping!” conveys additional details such as the
brands covered and the value-add offer that promises free shipping. In
much less than 160 characters, the listing tells shoppers looking for
affordable inks that CarrotInk.com is a good destination for them.
SEO Guide – Header Text
Header text is the general term for text
on a webpage that’s formatted with the H1, H2, and H3 elements. This
text is used to make headline and sub-headline text stand out, giving it
more visual weight to users and contextual weight to search engines.
In online stores, it’s important to
format main page headlines with the H1 tag. Typically, the product or
category name is automatically made the main header text like we see in
the example blow. Sub-headings that indicate the page’s various sections
can then be marked up as H2, H3 and so on.

If your online store publishes articles
and blog posts, remember to format main headlines in H1 and sub-headings
in H2 and H3 as well. Effective header text is brief, direct and
indicative of the context that the text under it conveys. It preferably
mentions the main keyword in the text it headlines at least once.
If you want to check whether all your
pages have proper header text, you can do an audit with Screaming Frog.
Simply run a crawl and filter the results to just HTML using the Filter
drop-down on the top left section of the app window. You can find column
headings that say “H1-1”and “H2-1” when you scroll right. The text
there are the headings in the page.

You can export these to Excel for easy data management and reference when applying changes.
SEO Guide – Image Alt Text
Google and other search engines have
gotten dramatically smarter these past two decades. They’re now so much
better at understanding language, reading code and interpreting user
behavior as engagement signals. What they haven’t refined to a science
just yet is their ability to understand what a picture shows and what it
means in relation to a page’s content.
Since bots read code and don’t
necessarily “see” pages the way human eyes do, it’s our responsibility
to help them understand visual content. With images, the most powerful
relevance signal is the alt text (shorthand for alternate text). This is
a string of text included in the image’s code that provides a short
description of what’s being shown.
Adding accurate, descriptive and
keyword-laced anchor text is very important to online stores since
online shoppers tend to buy from sites with richer visual content.
Adding alt text helps enhance your pages’ relevance to their target
keywords and it gets you incremental traffic from Google Image Search.
To check if your site uses proper image
alt text, we can use trusty Screaming Frog once more. Do a crawl of your
site from the home page and wait for the session to finish. Go to the
main menu and click Bulk Export>Images>All Alt Text. Name the file and save it. Open the file in Excel and you should see something like this:

Column B (Source) tells us which page the
image was found on. Column C (Destination) tells us the web address of
the image file. Column D (Anchor) is where we’ll find the anchor text if
something has already been written. If you see a blank cell, it means
no anchor text is in place and you’ll need to write it at some point.
Good image alt texts have the following qualities:
- About 50 characters long at most
- Describes what’s depicted in the image directly in one phrase
- Mentions the image’s main keyword at least once
Different CMS platforms have different
means of editing image alt texts. In a lot of cases, a spreadsheet being
fed into the CMS for bulk uploads can contain image alt text
information.
SEO Guide – Unique Category and Product Copy
SEO has everything to do with the age-old
Internet adage “content is king” and that applies even to ecommerce
websites. Analyses have continually shown that Google tends to favor
pages and sites with more breadth and depth in their content. While most
online store owners would cringe at the idea of turning their site into
“some kind of library,” smarter online retailers know that content can
co-exist with their desired web design schemes to produce positive
results.
One of the fundamental flaws that a lot
of ecommerce sites face from an on-page SEO perspective is the
proliferation of thin content on unoptimized category pages and product
pages with boilerplate content. Regular category pages have little text
and no fundamental reason for existence other than to link to
subcategory pages or product pages. Product pages with boilerplate
content, on the other hand, don’t deliver unique value to searchers. If
your online store is guilty of both shortcomings, search engines won’t
look at it with much favor, allowing other pages with better content to
outrank yours.
Category pages are of particularly high
importance due to the fact that they usually represent non-branded,
short and medium-length keywords. These keywords represent query intents
that correspond with the top and middle parts of sales funnels. In
short, these pages are the ones that users look for when they’re in the
early stages of forming their buying decisions.
Category pages can be optimized for SEO by adding one or both of the following:
- A short text blurb that states what the category is about and issues a call to action. This is usually a 2-3 sentence paragraph situated between the H1 text and the selection of subcategory product links.

- Longer copy that further discusses the category and the items listed under it. This can be one or more paragraphs and is positioned near the bottom of the page’s body, under the list of subcategories or products in the page. This text isn’t meant for human readers as much as it is for bots to consume and understand text.

For both text block types, the goal is to
increase the amount of text and provide greater relevance between the
page and its target keyword. The text blocks also provide opportunities
for internal linking as the text can be used as anchors for links to
other pages.
Product pages, on the other hand,
represent query intents that are near the bottom of the sales funnel.
Product pages match the intents of users who already have a good idea of
what they want to buy and are just looking for the best vendor to sell
it to them. The keywords that product pages represent are typically
lower in search volume but higher in conversion rate.
Optimizing the copy on product pages
usually isn’t too difficult as it is work intensive. Researching
hundreds or even thousands of product details, then rewriting them for
uniqueness can consume a ton of man hours. Typically, SEO savvy online
store owners selectively do this for just their priority product lines.
Product pages that are left unpopulated with unique content are
typically restricted using the noindex tag to focus link equity and
crawl budget on pages that have been optimized with unique copy.
When writing unique copy for product pages, focus on writing the following in your own way:
- The Benefit Statement – The part of the copy that lists 3 or 4 bullet points stating how the product solves a problem or makes life better for the target customer.
- Use Cases – A section in the copy where you mention possible uses and applications of the product.
- How-to Section – A part of the copy where you provide instructions on how the product can be used or installed.
- Customer Reviews – The part of the copy where you allow buyers to write feedback about a product. Just make sure to allow only verified customers to post. Moderating the content for quality and readability can also be helpful.
If you want to see a good example of rich, unique and useful product page copy, here’s one.
Category and product page content can
also help your pages achieve position zero on Google. This is the
instant answer box that’s displayed atop everything else in some
queries. If the content you wrote answers a question in a clear and
direct manner, uses proper header tags, and uses proper HTML tags for
bulleted and numbered lists, you may just find that your content is
being featured prominently in the SERPs.
Pagination
Product category pages sometimes contain
so many items that it would look awful to display them all in one page.
To solve that design problem, web designers make use of paginated
content. This is a method that allows a category page to display a
portion of the listings under it while keeping the rest hidden. If a
user wants to see the rest, he or she can click on page numbers that
take them to an identical page where other items in the list are
displayed.

While this works great for design and
usability, it can become an SEO issue if you’re not careful. Search
engines aren’t great at figuring out that a series of pages can
sometimes belong to a bigger whole. As such, they have the tendency to
crawl and index pages in a paginated series and view them as duplicates
because of the identical title tags, meta description and category page
copy displayed.
To help search engines identify a
paginated series in your category pages, simply have your web developer
implement the rel=”prev” and rel=”next” element in paginated content.
These elements tell search engines that pages belong to a series and
only the first one should be indexed. You can read about these HTML
elements in greater detail here.
SEO GUIDE – Aggregate Review Schema
You’ve probably seen search results for
products and services with review stars affixed to listings. These stars
do more than just occupy space on SERPs and catch visual attention –
they also encourage better clicking behavior on the part of the
searchers. For online stores, having review stars appear in product page
listings can be very handy. These rich snippets can be the deciding
factor on whether a user clicks on your listing or your competitor’s.

Fortunately, implementing review stars is
relatively easy. All you need to do is add this string of code to your
product pages and you should be all set:
<script type=”application/ld+json”>
{ “@context”: “http://schema.org”,
“@type”: “Product”,
“name”: “##PRODUCT###”,
“aggregateRating”:
{“@type”: “AggregateRating”,
“ratingValue”: “##RATING##”,
“reviewCount”: “##REVIEWS##”
}
}
</script>
You’ll have to fill in the ##PRODUCT###
part with the product’s name, the ##RATING## part with the aggregated
review scores of the product and the ##REVIEWS## part with the number of
reviews. If you don’t feel comfortable handling code, assigning this
task to a web developer will get things moving in the right direction.
Of course, adding this markup is not a
guarantee that Google will actually show the review stars. The search
giant has a set of algorithms to follow. To help your chances of getting
aggregate review stars to show up, keep the following in mind:
- Add the markup code only to specific product pages. Adding it to the ecommerce site’s home and category pages will have no effect.
- Encourage real customer reviews and avoid fabricating your own review scores.
- Try to promote a healthy balance between reviews from within the site and from external sources.
- If you’re planning to have people assign ratings to your pages at your behest, don’t do it for every product all at once. Having every page suddenly get ratings looks unnatural and may be viewed disfavorably by Google.
Note that some CMS platforms are better
than others with Schema markups. WordPress, for instance, has Schema
markups built in if you’re running a Genesis-based theme.
Bonus: Link Building Strategy
While this post is about on-site SEO,
it’s inevitable for ecommerce site owners to eventually think about link
building if they want to dominate their search market niche. Income
Diary has posted about how to build links the right way in the past, so we won’t go into detail on that. We will, however, discuss where to point those backlinks.
Generally, websites follow information
architectures that are pyramid shaped. Imagine the home page as the tip
of the pyramid while its main categories and subcategories make up the
next few layers below it. At the base of the pyramid are the product
pages. They’re the most numerous page type but they’re at the bottom of
the hierarchy.

Now, picture the pyramid as one that has
an internal water distribution system. In this case, the water is the
link equity. Most of it comes from the tip (the home page) since this is
where most inbound links point to. The equity flows down to the
category pages that are linked from the home page, then descends and
distributes throughout the product pages.
Having said that, the home page is the
best page to point inbound links to if your goal is to boost the ranking
power of all your pages. This helps the home page rank better for the
short keywords that it represents while providing incremental boosts to
category and product pages.
If you happen to have a product line that
you are prioritizing over the rest, it would serve you well to build
links to their respective category pages. This not only helps the
category page perform better in the SERPs, it also cascades more potent
link equity to your product pages.
Ultimately, good SEO for online stores is
all about promoting a higher degree of crawlability and relevance to
your pages. Adding the these optimization tasks to your site maintenance
bucket list may look like a major task, but the rewards are well worth
the effort.