7 indexability bugs killing small business SEO in 2026 (and how to find them)
Most small businesses I audit have at least one of these bugs. They’re silent, the site looks fine in a browser, and they’re responsible for more lost traffic than any algorithm update.
The sneaky thing about indexability bugs is that they don't break anything visible. The page renders correctly. The site looks healthy. Analytics keep showing traffic from old sources. And meanwhile Google has quietly stopped indexing 40% of your URLs and you don't notice for months.
I see these bugs constantly on small business sites. Here are the seven most common, in rough order of how often I find them.
1. Noindex left over from staging
Someone copied a layout file from staging to production and brought along the <meta name="robots" content="noindex"> tag. The page renders. Google crawls it. Google obediently doesn't index it. You don't notice because the URL still loads in your browser.
How to find it: Search Console > Pages > "Excluded by 'noindex' tag." Sort by impressions to find the most damaging ones first. Fix: remove the tag. Re-submit via URL inspection.
2. Canonical pointing at the wrong URL
A canonical tag says "this page is a duplicate, please rank this other URL instead." When the canonical points at the wrong URL — usually the homepage, or a parent category, or a 404 — Google obediently ranks the wrong URL and ignores the page that has the actual content.
Common cause: a templated canonical that wasn't customized per page. Common symptom: detail pages that have great content disappearing from search.
How to find it: Search Console > URL inspection > "Google-selected canonical." If it disagrees with your declared canonical on multiple pages, dig in.
3. JavaScript-rendered content that Googlebot can't see
Single-page apps and heavily-hydrated React sites often render their main content via JavaScript after the initial HTML loads. Googlebot does render JavaScript — eventually — but with significant delay and some content losses. AI crawlers like GPTBot don't render JavaScript at all.
How to find it: Search Console > URL inspection > "View tested page" > HTML. Compare what Google saw against what your browser sees. If the rendered HTML is missing your main content, you have a server-side rendering problem.
Fix: server-side rendering or static generation for any page you want indexed. For Next.js sites, that means using getStaticProps / getServerSideProps or the App Router equivalents instead of fetching client-side.
4. Pagination collapsed into one infinite scroll
Replacing paginated category pages (page 1, 2, 3) with infinite scroll often nukes indexability of everything past the first scroll. Googlebot doesn't scroll. If your second page of products is only reachable via JavaScript scroll handlers, those products effectively don't exist in search.
Fix: keep paginated URLs as a fallback even if you ship infinite scroll for users. ?page=2 URLs that work directly preserve indexability.
5. Robots.txt blocking JavaScript or CSS
Old SEO advice said to block /wp-content/ or /static/ directories to "save crawl budget." Modern Google needs to render those resources to evaluate the page. Blocking them tells Google your site looks broken.
How to find it: PageSpeed Insights will warn about blocked resources. Search Console will mark pages as having issues if critical CSS or JS is unreachable. Fix: remove the blocks. The crawl budget concern was always overblown for sites under 10K pages.
6. HTTPS migration that left HTTP variants live
Migrations are stressful and corners get cut. The most common shortcut: switch the site to HTTPS, set the canonical to HTTPS, and assume Google will figure it out. Months later, half your URLs are still indexed at the HTTP address with split link equity.
Fix: 301 redirect every HTTP URL to its HTTPS equivalent at the server level (not via meta refresh). Update internal links to use the HTTPS protocol. Resubmit your sitemap.
7. Sitemap full of garbage URLs
WordPress and most CMSs auto-generate sitemaps. By default, those sitemaps include archive pages, category pages, tag pages, attachment pages, author archives, and date-based archives — most of which are thin or duplicative. Submitting them all to Google trains Google to think your site is mostly low-quality.
Fix: noindex the thin archives, exclude them from the sitemap, and only submit pages you actually want ranked. Yoast and Rank Math both have settings for this; on custom platforms you'll need to filter the sitemap output.
The diagnostic order
If you suspect indexability problems but don't know where to start, check in this order:
- Search Console > Pages > Indexed count vs. Submitted count. The gap is your problem.
- Search Console > Pages > Excluded reasons. Each reason has a specific fix.
- URL inspection on three URLs that should rank but don't. Look at canonical, mobile usability, and the rendered HTML.
- PageSpeed Insights on a few key pages to catch render-blocking and mobile usability issues.
- Sitemap audit (use our Sitemap Analyzer) to catch garbage URLs and stale lastmod dates.
Most of these bugs take less than an hour to fix once you find them. The hard part is finding them, because the site looks fine. That's why indexability audits are the highest-leverage SEO work for most small business sites — the upside is recovered traffic that was already earned but quietly lost.
Ready to Audit Your Website?
Put these insights into action with our free SEO audit tool. Get instant analysis and recommendations.
Start Free SEO Audit✨ 100% Free • AI-Powered • Instant Results