Discover practical ways to fix any flaws & weaknesses in your website!
Get a free audit of your ecommerce store. Fix issues before they start costing you a fortune. 🛠️
Magento Technical SEO Issues You Shouldn't Be Ignoring
When’s the last time you checked your store for technical SEO issues? Our technical SEO checklist will help you to find errors that hurt your Magento store.
What’s the problem with my store? What are the common technical SEO issues?
Many eCommerce websites don’t take SEO seriously and focus mostly on PPC and social media ads to promote their sites. eCommerce SEO is not easy, and it may take a lot of time and effort to set up. But if set up properly, it will bring you quality traffic that converts.
To audit your site, you should look at on-page SEO, off-page SEO, and technical SEO.
But don’t start your efforts with the first two.
Technical SEO is one of the most crucial parts of your strategy, and you have to focus on it first. Technical SEO is the first step in creating a better search experience. Other SEO projects should come after you’ve guaranteed your site has proper usability.
NOTE:
Technical SEO includes components like page titles, title tags, HTTP header responses, XML sitemaps, 301 redirects, and metadata. Technical SEO does NOT include analytics, keyword research, backlink profile development, or social media strategies.
Technical SEO Checklist
At MageCloud, during hundreds of Magento site audits over the years, we have consistently encountered the same technical SEO issues.
Here we share a few tips on what to start out. Hopefully, ourtechnical SEO checklist will help you to find the SEO errors that hurt your Magento online store.
Be honest, when’s the last time you checked that everything was working right and you were getting the right results in terms of traffic and online authority?
It’s definitely high time to take another look.
#1 HTTPS Security
Security has always been “a top priority” for Google. Since October 2017, Google has been showing “Not secure” caution signs in Chrome every time a web user lands on an HTTP site.
The first step for this quick fix is to check if your site is HTTPS.
How to Check?
✔️ Type your domain name into Google Chrome. If you see the “secure” message (pictured below), your site is secure.
✔️ If your site is not secure, when you type your domain name into Google Chrome, it will display a gray background — or even worse, a red background with a “not secure” warning. This warning in your address bar is going to kill your conversions.
How to Fix?️
✔️ To convert your site to HTTPS, you need an SSL certificate from a Certificate Authority.
✔️ Once you purchase and install your certificate, your site will be secure.
#2 Site Isn't Indexed Correctly
When you search for your brand name in Google, does your website show up in the search results? If the answer is no, there might be an issue with your indexation. As far as Google is concerned, if your pages aren’t indexed, they don’t exist — and they certainly won’t be found on the search engines.
How to Check?
✔️ Type the following into Google’s search bar: “site:yoursitename.com” and instantly view the count of indexed pages for your site.
✔️ Your site is indexed, but there are many MORE results than expected. Look deeper for either site-hacking spam or old versions of the site that are indexed instead of appropriate redirects in place to point to your updated site.
✔️ In case your site is indexed, but you see quite a bit LESS than expected, perform an audit of the indexed content and compare it against the pages you want to rank. If you’re not sure why the content isn’t ranking, check Google’s Webmaster Guidelines to ensure your site content is compliant.
✔️ If the results are different from what you expected in any way, verify that your important website pages are not blocked by your robots.txt file (see #4 on this list). You should also verify you haven’t mistakenly implemented a NOINDEX meta tag (see #5 on this list).
#3 No XML Sitemaps
XML sitemaps help Google search bots understand your site pages better, so they can crawl your site more effectively and intelligently.
How to Check?
✔️ Type your domain name into Google and add “/sitemap.xml” to the end, as pictured below.
If your website has a sitemap, you will see something like this:
How to Fix?️
✔️ If your website doesn’t have a sitemap (and you end up on a 404 page), you can create one yourself or hire a web developer to create one for you. The easiest option is to use an XML sitemap generating tool.
#4 Missing or Incorrect Robots.txt
A missing robots.txt file is a big red flag — but did you also know that an improperly configured robots.txt file destroys your organic site traffic?
✔️ If you get a result that reads “User-agent: * Disallow: /” then you have an issue, cause this example tells all robots to stay out of your website:
Photo credit for robots.txt bad file [Source: robotstxt.org]
How to Fix?️
✔️ If you see any “NOINDEX” or “NOFOLLOW” in your source code, check with your web developer, as they may have included it for specific reasons.
✔️ If there’s no known reason, have your developer change it to read <meta name=”robots” content=” INDEX, FOLLOW”> or remove the tag altogether.
#5 Meta Robots NOINDEX
NOINDEX can be even more damaging than a misconfigured robots.txt at times.
Most commonly, the NOINDEX is set up when a website is in its development phase, but once the website goes live, it’s imperative to remove the NOINDEX tag.
Since so many web development projects are running behind schedule and pushed to live at the last hour, this is where the mistake can happen. Do not blindly trust that it was removed, as the results will destroy your site’s search engine visibility.
When the NOINDEX tag is appropriately configured, it signifies certain pages are of lesser importance to search bots. However, when configured incorrectly, NOINDEX can immensely damage your search visibility by removing all pages with a specific configuration from Google’s index.
How to Check?
✔️ Manually do a spot-check by viewing the source code of your page (right-click on your site’s main pages and select “View Page Source”), and looking for or lines in the source code that read “NOINDEX” or “NOFOLLOW” such as:
Photo credit for robots.txt bad file [Source: robotstxt.org]
90% of the time, you’ll want it to be either “INDEX, FOLLOW” or nothing at all. If you see one of the above, you need to take action.
✔️ If you see “Disallow: /”, immediately talk to your developer. There could be a good reason it’s set up that way, or it may be an oversight.
✔️ If you have a complex robots.txt file, like many e-commerce sites, you should review it line-by-line with your developer to make sure it’s correct.
#6 Multiple Home URL Versions
Remember when you discovered that “yourwebsite.com” and “www.yourwebsite.com” go to the same place? While this is convenient, it also means Google may be indexing multiple URL versions, diluting your site’s visibility in search.
Every site should have the following two properties setup in Search Console:
http://www.yourwebsite.com
http://yourwebsite.com
If your website has an SSL certificate set up, such as an eCommerce website, you would also add two additional properties:
https://www.yourwebsite.com
https://yourwebsite.com
That brings our total to four site versions, all being treated differently! The average user doesn’t really care if your home page shows up as all of these separately, but the search engines do.
If Search Console is successfully set up for your website (preferred site version is set), you should begin to only see one of the site versions getting all the attention.
Setting the preferred site version in Google Search Console is a good first step in addressing the numerous versions of your site, but it’s more than just selecting how you want Google to index your website. Your preferred site version affects every reference of your website’s URL – both on and off the site.
How to Check?
✔️ Check manually if different versions of your URL successfully flow to one standard URL. This can include HTTPS and HTTP versions, as well as versions like “www.yourwebsite.com/home.html”. Check each possible combination.
✔️ Another way is to use “site:” operator in Google search (“site:yoursitename.com”) to find out which pages are indexed and if they stem from multiple URL versions.
How to Fix?️
✔️ If you discover multiple indexed versions, you’ll need to set up 301 redirects or have your developer set them up for you.
✔️ Set up a schedule to monitor your URL canonicalization on a weekly or monthly basis.
#7 Incorrect Rel=Canonical
Rel=canonical is particularly important for all sites with duplicate or very similar content (especially e-commerce sites).
Dynamically rendered pages (like a category page of blog posts or products) can look like duplicate content to Google search bots. The rel=canonical tag tells search engines which “original” page is of primary importance (hence: canonical) — similar to URL canonicalization.
How to Check?
✔️ Spot-check important pages to see if they’re using the rel=canonical tag.
✔️ Use a site scanning software to list out all the URLs on your site and determine if there are duplicate page problems that can be solved with a rel=canonical tag.
How to Fix?️
✔️ If you discover multiple indexed versions, you’ll need to set up 301 redirects or have your developer set them up for you.
✔️ Set up a schedule to monitor your URL canonicalisation on a weekly or monthly basis.
#8 Duplicate Content
With more and more brands using dynamically created websites, content management systems, and practicing global SEO, the problem of duplicate content plagues many websites.
The problem with duplicate content is that it may “confuse” search engine crawlers and prevent the correct content from being served to your target audience.
Usually, duplicate content can occur for many reasons:
– E-commerce site store items appear on multiple versions of the same URL.
– Printer-only web pages repeat content on the main page.
– The same content appears in multiple languages on an international site.
How to Check?
✔️ There are a lot of tools to find duplicate content. The best-known duplicate content checkers are probably Copyscape and Siteliner.
How to Fix?️
Each of the issues mentioned above can be resolved with:
Structured data is a simple way to help Google search crawlers understand the content and data on a page. For example, if your page contains a recipe, an ingredient list would be an ideal type of content to feature in a structured data format. Address information like this example from Google, is another type of data perfect for a structured data format.
How to Check?
✔️ Use the Rich Results Tool, created by Google, to help you validate and test the syntactic correctness and alignment to Google recommendations of your schema markup.
How to Fix?️
✔️ As you roll out new content, identify opportunities to include structured data on the page, and coordinate the process between content creators and your SEO team. Better use of structured data may help improve CTR and possibly improve rank position in the SERP.
Once you implement structured data, review your Google Search Console report regularly to make sure that Google is not reporting any issues with your structured data markup.
#10 Broken Links
Good internal and external links show both users and search crawlers that you have high-quality content. Over time, content changes, and once-great links break. Broken links create a poor user experience and reflect lower-quality content, a factor that can affect page ranking.
A website migration or relaunch project can spew out countless broken backlinks from other websites. Some of the top pages on your site may have become 404 pages after a migration.
How to Check?
✔️ Two types of tools are great to find all external links that are broken — Google Search Console, and a backlink checker such as Moz, Majestic, or Ahrefs. From there, you can reach out to the sites with broken links and provide them with the correct link or new page.
How to Fix?️
✔️ After identifying your top pages with backlinks that are dead, 301 redirect these to the best pages
✔️ While internal links should be confirmed every time a page is removed, changed, or a redirect is implemented, the value of external links requires regular monitoring. The best and most scalable way to address broken links is to run regular site audits.
#11 301 & 302 Redirects
Redirects are an amazing tool in an SEO’s arsenal for managing and controlling dead pages, for consolidating multiple pages, and for making website migrations work without a hitch.
301 redirects are permanent and 302 redirects are temporary. The best practice is to always use 301 redirects when permanently redirecting a page.
301 redirects can be confusing for those new to SEO, trying to properly use them. They’re a lifesaver when used properly, but a pain when you have no idea what to do with them.
How to Check?
✔️ Do a full review of all the URLs on your site and look at a high level. You can use Redirect Checker or the RedirectCheck tools.
How to Fix?️
✔️ If using 302 redirects incorrectly for permanent redirects, change these to 301 redirects (But at first, discuss using 302 redirects with your development team).
✔️ Include redirect checking in your monthly or weekly site scan process.
NOTE:
1. Never redirect all the pages from an old site to the home page unless there’s a really good reason
2. Don’t go redirect-crazy on all 404 errors — use them for pages receiving links or traffic only to minimize your redirects list
#12 Slow Page Speed
If your site doesn’t load quickly, your users will go elsewhere. Site speed – and other usability factors – matter to the user experience and to Google.
NOTE:
Your site should pass Core Web Vitals assessment - quality signals key to delivering great UX on the web.
How to Check?
✔️ Use Google PageSpeed Insights to detect specific speed problems with your site. (Be sure to check desktop as well as mobile performance.)
How to Fix?️
✔️ The solution to site speed issues can vary from simple to complex. Common site speed solutions can include image optimization/compression, browser caching improvement, server response time improvement, and JavaScript minifying.
✔️ Speak with your web developer or contact us to ensure the best possible solution for your site’s particular page speed issues.
NOTE:
1. Never redirect all the pages from an old site to the home page unless there’s a really good reason
2. Don’t go redirect-crazy on all 404 errors — use them for pages receiving links or traffic only to minimize your redirects list
As your mobile site will be the one indexed, you’ll need to do the following for all “.m” web pages:
Whether Google has transitioned you to mobile-first indexing yet or not, you need to guarantee your site is mobile friendly to ensure exceptional mobile user experience. Anyone using responsive website design is probably in good shape. If you run a “.m” mobile site, you need to make sure you have the right implementation on your m-dot site so you don’t lose your search visibility in a mobile-first world.
How to Fix?️
As your mobile site will be the one indexed, you’ll need to do the following for all “.m” web pages:
✔️ Guarantee the appropriate and correct hreflang code and links.
✔️ Update all metadata on your mobile site. Meta descriptions should be equivalent on both mobile and desktop sites.
✔️ Add structured data to your mobile pages and make sure the URLs are updated to mobile URLs.
#14 Missing Alt tags
Broken images and those missing alt tags are a missed SEO opportunity. The image alt tag attribute helps search engines index a page by telling the bot what the image is all about. It’s a simple way to boost the SEO value of your page via the image content that enhances your user experience.
How to Fix?️
✔️ Most SEO site audits will identify images that are broken and missing alt tags. Running regular site audits to monitor your image content as part of your SEO standard operating procedures makes it easier to manage and stay current with image alt tags across your website.
#15 Missing or Non-Optimised Meta Descriptions and H1 tag
Meta descriptions are those short, up to 160-character content blurbs that describe what the web page is about. These little snippets help the search engines index your page, and a well-written meta description can stimulate audience interest in the page.
It’s a simple SEO feature, but a lot of pages are missing this important content. You might not see this content on your page, but it’s an important feature that helps the user know if they want to click on your result or not. Like your page content, meta descriptions should be optimized to match what the user will read on the page, so try to include relevant keywords in the copy.
And although H1 tags are not as important as they once were, it’s still an on-site SEO best practice to prominently display.
This is actually most important for large sites with many, many pages such as massive eCommerce sites. It’s most important for these sites because they can realistically rank their product or category pages with just a simple keyword-targeted main headline and a string of text.
How to Fix?️
✔️ For pages missing meta descriptions: run an SEO site audit to find all pages that are missing meta descriptions. Determine the value of the page and prioritize accordingly.
✔️ For pages with meta descriptions: evaluate pages based on performance and value to the organization. An audit can identify any pages with meta-description errors. It’s important to make sure that meta descriptions are unique to a page.
✔️ Look for whether H1 and H2 tags are being found on pages across your site; if not, fix it.
In Summary
An SEO audit reveals a complete overview of site health and optimization efforts.
A website SEO audit is like a health check for your site. It allows you to check how your web pages appear in search results, so you can find and fix any weaknesses.