Follow us on Facebook
Follow us on Twitter
Signalogic on LinkedIn

Google Canonical Page Algorithm Problems

The Problem

The problem is that Google's indexing algorithms may inexplicably select incorrect "canonical pages" for a website, mistakenly selecting as "canonical" unrelated pages elsewhere on the site. This allows Google's algorithms to label completely original content as "duplicated" -- effectively making a mistake that no human would ever make -- and thus blocking original pages from appearing in Google search results. This can happen for 100s of pages on a website, blocking the entire site from appearing in Google search results, or if not the entire site, then its crucial pages (also known as "landing pages"). And, this can happen even though the blocked pages contain 100% original content, not duplicated anywhere else on the web, even including registered trademarks or copyrighted material, and otherwise following all of Google's published rules and guidelines.

Probably, the new canonical page algorithm was implemented due to e-commerce abuses, for example 100s of e-commerce sites with pages having nearly identical content. Google has attempted to address this in the past, but -- as might be expected in any measure/counter-measure situation -- the e-commerce sites got better at subtly changing their pages so that Google's algorithms would see them as different. Basically, it's bot vs. bot warfare, and now Google has the upper hand again. Of course Google's focus on e-commerce completely ignores independent vendors providing infrastructure software solutions, which is the case for Signalogic. Below is a 16-mo performance history for signalogic.com, showing the time period when the canonical algorithm change "hit":

Signalogic website 16 month performance history, showing effects of Google canonical algorithm changes in April 2019
Signalogic website 16 month performance history, showing effects of Google's major canonical algorithm change

As you can see, complete disaster -- virtually no web presence for more than a month, then a long, slow slog back to a medium presence. Finally after 5 months clicks and impressions are approaching normal levels. 2 The steps below document the work and effort that went into this 5 month "recovery".

Update

Our original DoJ and other complaints are given below. Since then, I've been able to work around Google's canonical page algorithm problems. Below I've given a step-by-step procedure. By no means is this guaranteed to work -- as everyone knows, Google's algorithms are complex and closed. All anyone can do is spend 1000s of hours of time to deduce what Google might be doing, run 100s of tests, and try to come up with something that works. Of course that's a waste of time, but it's the "search tax" imposed by Google as a monopoly. As things stand people have no choice, other than to pay Google monthly for advertising (they quoted me $2,000 per month to fix canonical issues). If there was valid search competition Google would need to be clear and definitive about what they're doing to select canonical pages and otherwise detect duplicate content. For further discussion, see "Analysis", below.

Also, I believe these steps are likely to apply more so if your site runs a page generation script (e.g. Perl or PHP). For example our URLs contain a Perl script that generates the visible page, for instance https://www.signalogic.com/index.pl?page=directcore 1

Steps

1) First and foremost, the script parameter that controls page selection (for our site, this is the text following "page=") should be long and descriptive, and contain actual words in the dictionary, nouns, or names. Abbreviations and short-hands, and even short but actual words, seem to be interpreted by Google's anti e-commerce algorithms as bot-generated, and thus likely to be associated with duplicate content.

2) Second, related to 1), in your Google Search Console (GSC), set "URL Parameters" for your script parameter as follows:

   Q: "Does this parameter change page content seen by the user?"
   A: "Yes, changes, reorders, or narrows page content"

   Q: "How does this parameter affect page content?"
   A: "Specifies"

   Q: "Which URLs with this parameter should Googlebot crawl?"
   A: "Every URL"

Below is a screenshot showing an example.

GSC URL Parameters screen capture, showing correct settings for page generation scripts

3) Also related to 1), do not leave any short page names as active, or hanging around that Google can find, for example if you have old pages or are moving from http to https. Clean up any shorthand script parameters.

4) When you remove a page, make sure Google sees it as a "hard 404". It's not enough to return a soft 404. For page generation scripts, this might be an issue, depending on the programming of the script.

5) Any time Google mis-identifies your correct (target) canonical page, immediately attack the problem as follows: What you may find is that fairly soon GSC will still show the target page as not indexed, but now the canonical URL will show as "N/A". That's good. Any time you force Google's anti e-commerce algorithms to recalculate, that's progress. As soon as you see N/A, re-index both pages again. Google might then latch on to another wrong page; if so keep repeating this cycle. It may take some time.

6) Constantly monitor around 20 or so of your key pages in GSC; try to choose these pages as a general, broad representation of your site (i.e some of A, some of B, etc). At any time, Google might again mis-identify a canonical due to periodic crawls, and you're back in the meat grinder again. But as I noted above, that's the "search tax" we are forced to pay, there isn't a choice. The objective is to catch and fix canonical mistakes early -- what I encountered is that if you let Google incorrectly select a large number of pages as duplicates, it can take a long time to dig your way out. Don't let their bad-ass algorithms get the jump on you.

7) Things I found that did not help: submitting/removing sitemaps, temporarily removing URLs ("Remove URLs" in GSC), no-index tags, and changing page content.

Analysis

I can only guess at what Google is actually doing. It seems that in general they are suspicious of page generation scripts, and employ hidden algorithms and decisions against such scripts, as they can be weaponized by e-commerce. But as we know, page generation scripts are used for legitimate reasons, including:
  1. They avoid extensive "dynamic content" to generate the page, which usually amounts to extensive reliance on Java script. Unfortunately Java script is increasingly seen as a security risk and an increasing number of users de-activate it
  2. They are widely used by small businesses and organizations that cannot afford dozens of Java script programmers and IT personnel to maintain highly complex web sites with 100s of pages of code (i.e. code vs. content)

This may also be true for fixed URLs, if they contain abbreviations or shorthands, for example for a site with many such pages Google's algorithms might decide such URLs are bot-generated for duplicate content and "rankings spam" purposes. But that is just a guess, I did not find evidence of that as our site is not organized in that way.

The latter point in 7) above is worth emphasizing: actual content is not what Google is using when it makes canonical mistakes. You can change text and images all day long and Google will still incorrectly identify canonical pages, if you have otherwise triggered their anti e-commerce thresholds.

1 We use .htaccess rewrites to allow shortened URLs, for example if you enter https://signalogic.com/directcore, our web server rewrites to https://www.signalogic.com/index.pl?page=directcore. We never submit the shortened URLs to Google for indexing, as they would be rejected as "re-directs".
2 For anyone wondering, the weekly click/impression oscillations are due to the nature of Signalogic's software products, which are B2B oriented (business-to-business). Typically business personnel are not searching for such products on weekends. Also by a ratio of more than 90%, they tend to use desktop search, not mobile.

Original Complaints

As a veteran-owned small business CEO, I have written this page to document a non-obvious, but crucial problem with Google Search. Google uses a "Canonical Page" algorithm to identify duplicate pages within a website and select one for indexing in Google Search. At best, this algorithm is buggy and does not function correctly. At worst, Google intentionally wrote this algorithm to deliberately disadvantage small business.

Below is a "Google Search Console" screenshot showing the Canonical Page algorithm selecting a random, unrelated page on my company's website (labeled "hw") as canonical to other pages that are crucially important to our business. This blocks the important pages from being indexed in Google Search. In the screen shot, Google is selecting the hw page as canonical to an important page describing a software product named "DirectCore". As you can see from the hardware_products and DirectCore pages themselves, and the screenshots below, they are clearly and substantially different. A human would not label them as identical.

Over the last few months I have followed Google's published instructions carefully, ensuring all of our pages use the https secure protocol, are mobile friendly, and describe unique content. For the hw page I tried a "noindex tag", submitting a request to Google's "URL Removal Tool" to de-index the page, and setting the page's priority to 0.1 in our sitemap. Nothing works -- Google for whatever reason is locked on to this page. I have sent countless feedbacks to Google on this issue. I have asked them for guidance and advice, as well as any possible workaround or fix that I can implement, and received no answer from them whatsoever. None of my efforts have had any effect, and the Canonical Page algorithm still inexplicably considers the hw page as a "phantom duplicate".

The problem continues to occur with dozens of our company's website pages. What may be worse is that DirectCore is a registered trademark, and Google is unable to distinguish content even to the level of the US Patent and Trademark office. I can hardly find that believable, but evidently it is the case. I feel this is very unfair to small business; Google should work hard and pull their own weight like everyone else. If my company had a buggy algorithm in software we provide to customers we would be out of business. I am assuming this is a bug they can fix, or provide a workaround for, as I certainly hope Google's behavior is not something like Uber's "Greyball" that directly targets a subset of customers.

       

Forum / Thread Technical Discussions

    SE Roundtable Thread, 26Dec18
    "Google is selecting an old, unrelated page as Canonical, blocking our pages from indexing", 15May19, Google WebMaster Forum
    "Google selects one unrelated page as Canonical for many other pages, blocking our pages from indexing", 7Jun19, Webmasters Stack Exchange Thread
    "Google selecting wrong canonical", 13Jun19, Google WebMaster Forum

Texas Attorney General Complaint

Complaint to Texas Attorney General (pdf)

FTC Complaint

Complaint to FTC (pdf)

DoJ Complaint

Complaint to DoJ and initial reply (png)