If your website is live but not appearing in Google, the problem is usually not that the page exists. The problem is that Google cannot discover it properly, cannot crawl it properly, or does not think it is worth indexing. For South African businesses, that matters because a page that is not indexed cannot support visibility, enquiries, or sales.
A new service page, city page, or ecommerce category can be fully published and still do nothing if Google is receiving weak, conflicting, or low-value signals. Indexing is rarely the real problem on its own. It is often the first clear sign that the site is not making its priorities easy for Google to understand.
What causes a website not to be indexed
Most indexing problems fall into four groups: the page is blocked, the page is weak, the page is sending conflicting signals, or the site is wasting crawl attention on the wrong URLs.
Noindex tags and blocked directives
A common cause is an accidental noindex directive in the HTML or HTTP header. That tells Google not to include the page in its index. It often happens after a staging environment, plugin update, theme change, or redesign.
Blocked crawl paths can also interfere. robots.txt does not work the same way as noindex, but it can still prevent Google from accessing the page properly. If Google cannot crawl the page, it cannot assess the content, canonical, or internal linking signals properly.
Canonical mistakes
A canonical tag tells Google which URL should be treated as the preferred version. If a page points to another URL by mistake, Google may ignore the page you want indexed and consolidate signals elsewhere.
This is common on ecommerce templates, filtered category pages, paginated URLs, and service pages built from reused templates.
Poor internal linking and orphan pages
Pages outside the main internal linking structure are harder to discover and easier to deprioritise. If a page is only listed in the XML sitemap, or buried too deeply with no contextual links, Google may treat it as low priority.
Orphan pages are a common cause of poor indexing on service sites and large content libraries. The page exists, but the site is not helping Google find it or understand why it matters.
Duplicate, thin, or low-value content
Google does not index every live page. If a page overlaps too heavily with another page, adds very little unique value, or looks like a near-copy built for keyword coverage, Google may crawl it and still choose not to index it.
This often happens with thin city pages, lightly rewritten service variations, duplicate product categories, and template-driven pages with weak original content.
Crawl inefficiency and rendering issues
Some pages are live but hard to process. Heavy JavaScript, broken rendering, unstable server responses, redirect chains, or bloated URL patterns can all make indexing less reliable.
On larger sites, crawl inefficiency becomes more obvious. If Google spends too much time on filter URLs, duplicate URLs, or low-value pages, it may delay or deprioritise important URLs.
Sitemap problems
Sitemaps do not guarantee indexing, but they do support discovery and canonical consistency. If the XML sitemap includes redirected URLs, non-indexable URLs, outdated URLs, or excludes important pages, it weakens the site’s indexing signals.
The page does not offer a strong reason to index
Some pages are technically accessible but still not index-worthy. Usually that means the page is too generic, too repetitive, too thin, or clearly weaker than another page on the site.
A live page is not automatically an indexable page.
What to check first
Do not start by changing random settings. First identify what Google is actually seeing.
Check Google Search Console first
Use URL Inspection on the exact URL. That will show whether the page is indexed, whether it can be indexed, whether Google crawled it, and whether Google selected a different canonical.
Then review the Page Indexing report and note the status message attached to the page or group of pages. Common examples include:
- Crawled – currently not indexed: Google fetched the page but chose not to index it. This often points to weak content, duplication, soft-quality problems, or mixed signals.
- Discovered – currently not indexed: Google knows the page exists but has not crawled it yet. This often points to weak internal linking, crawl inefficiency, low priority, or too many competing URLs.
- Excluded by ‘noindex’ tag: The page is being explicitly blocked.
- Duplicate, Google chose different canonical than user: Your canonical says one thing, Google prefers another. This usually means duplication or weak canonical signals.
- Alternate page with proper canonical tag: Google is treating the page as a duplicate by design. That may be correct, or it may reveal a template issue.
- Soft 404: The page loads, but Google thinks it behaves like a low-value or error-like page.
Quick Search Console checklist
| Search Console statusLikely causeFirst action | ||
|---|---|---|
| Crawled – currently not indexed | Thin content, duplication, weak value, mixed signals | Compare the page to similar URLs, strengthen uniqueness, review canonicals and internal links |
| Discovered – currently not indexed | Weak discovery, crawl inefficiency, low priority | Improve internal linking, add sitemap inclusion, reduce crawl waste elsewhere |
| Excluded by noindex | Meta robots or header directive | Remove unintended noindex and confirm live source output |
| Duplicate, Google chose different canonical than user | Duplicate content or weak canonical handling | Review canonical target, compare overlapping pages, consolidate if needed |
| Alternate page with proper canonical tag | Intentional or accidental duplication | Check whether the duplicate should exist at all |
| Soft 404 | Thin or weak page, broken template, misleading URL state | Improve page substance, confirm 200 response, check page purpose |
These labels matter because each one points to a different problem.
Confirm the page is actually indexable
Check the live page source and HTTP response. Confirm that the page has:
- a 200 status code
- no unintended
noindex - a self-referencing canonical, unless there is a valid reason not to
- visible main content in rendered HTML
- no broken redirect logic or server instability
Do not rely only on the CMS interface. Check the live URL.
Check whether the page can be discovered
Ask a practical question: how would Google find this page without you manually submitting it?
Look for contextual internal links from relevant pages. A new service page linked only from the sitemap is weak. A city page buried four clicks deep with no supporting links is weak. A category page accessible only through filter combinations is weak.
Also confirm that the page is included in a clean XML sitemap.
Check whether Google can render the page properly
If the page depends heavily on JavaScript, confirm that the main content, canonical, meta directives, and important internal links are available when rendered. If the content loads late, breaks during rendering, or disappears behind scripts, indexing becomes unreliable.
Similar problems people confuse with indexing
A lot of “not indexed” complaints turn out to be something slightly different. That matters because the fix changes depending on which problem you actually have.
Live but blocked
The page exists, but Google is being told not to index it or cannot access it properly. This usually comes down to noindex, blocked crawl paths, or server access problems.
Crawled but not indexed
Google saw the page and chose not to keep it in the index. That usually means low value, duplication, soft-quality issues, or conflicting signals.
Indexed but not ranking
This is not an indexing problem. The page is already in Google, but it is not competitive or relevant enough to perform. That shifts the discussion away from crawl and indexability and toward targeting, content quality, competition, and authority.
Canonicalised to another URL
The page exists and may be crawled, but Google is treating another URL as the main version. That is usually a duplication or canonical issue, not a discovery issue.
A useful rule here is simple: if Google cannot find or keep the page, you have an indexing problem. If Google can find it but does not surface it well, you have a ranking problem. Mixing those up leads to the wrong fixes.
What to fix or change
The right fix depends on whether the issue affects one page, a set of related pages, or the site’s template logic.
If the issue is one service page
A newly launched service page may fail to index because it is thin, weakly linked, or too similar to another service page. In that case:
- remove any accidental
noindex - confirm the page self-canonicalises correctly
- add strong internal links from parent and related service pages
- improve unique value, not just word count
- make sure the page has a clear role in the site architecture
For example, a Cape Town SEO page that repeats the same copy as a Johannesburg page with only the city name changed may be crawled but not indexed because Google sees little reason to keep both.
Another common example is a new “SEO migration services” page added under technical SEO, but linked only from the sitemap and not from the main technical SEO hub or related pages. In that case, discovery and internal priority are both weak even if the page itself is decent.
If the issue follows a new site launch
After launches and redesigns, indexing problems are often technical rather than editorial. Check for:
- staging
noindexdirectives left behind - canonicals copied from old templates
- broken internal links
- sitemap errors
- redirect mistakes
- rendering failures introduced by the new build
A newly launched South African law firm site, for example, may publish all its core service pages correctly in the CMS but still fail to index them because the service template outputs the homepage as the canonical across the whole set.
If the issue affects ecommerce category pages
Category pages often run into duplication, thinness, crawl waste, or faceted navigation problems. Fixes may include:
- consolidating overlapping categories
- tightening canonical rules
- reducing indexable filter combinations
- strengthening internal links to priority categories
- improving category copy so the page offers more than a product grid
A local retailer might have “men’s running shoes”, “running shoes for men”, and multiple filter-generated brand versions competing for the same intent. Google may choose one and ignore the rest.
Another common case is when colour, size, brand, and sale filters each generate crawlable URL variants. Google then spends time crawling dozens of low-value combinations while important category pages are discovered late or revisited less often.
If the issue is sitewide template misconfiguration
When many pages show the same exclusion pattern, stop treating it as a page-level issue. Check the template rules behind canonicals, robots directives, sitemap generation, pagination, and internal links.
That is not a one-page problem. It usually means the CMS, SEO plugin, theme logic, or development implementation is outputting the wrong signals across a whole template set.
Quick fixes vs broader fixes
Quick fixes include removing noindex, correcting canonicals, fixing status codes, improving internal links, and updating sitemap inclusion.
Broader fixes include consolidating duplicate page sets, cleaning up faceted navigation, correcting template output, reducing crawl waste, and improving how important pages are linked from hubs and parent sections.
If multiple commercially important pages are affected, the problem is usually structural rather than isolated.
How this connects to broader SEO priorities
Indexing problems are rarely the root problem. They are usually the first visible symptom of a site that is sending unclear signals about what matters.
If Google keeps crawling filter URLs instead of priority categories, the issue is crawl waste. If service pages sit too deep in the site and attract few internal links, the issue is architecture. If several pages are being canonicalised to the same URL, the issue is duplication or poor page differentiation. If whole groups of pages share the same exclusion pattern, the issue is often template logic, CMS output, or faulty SEO settings applied at scale.
That distinction matters because fixing the visible symptom does not remove the cause. You can get one page indexed by correcting a tag or submitting it manually, but that will not solve weak internal linking, repeated template copy, uncontrolled filters, or conflicting canonicals across a wider page set.
For a business owner or marketing lead, the more useful question is not only “why is this page not indexed?” but “what in the site setup keeps creating this outcome?” On a service site, that may mean thin page variation, weak parent-child structure, or poor internal linking from stronger pages. On an ecommerce site, it often means faceted URL sprawl, overlapping categories, or too many low-value URLs competing for crawl attention.
When the same problem appears across more than one or two important pages, treat it as a site decision problem, not just a page problem.
When to get expert help
Some indexing problems can be solved internally. A clear noindex, missing internal link, or broken sitemap entry may be straightforward to fix.
Expert review becomes more useful when:
- the cause is unclear
- multiple pages are affected
- Search Console shows repeated exclusion patterns
- a migration, redesign, or CMS change preceded the problem
- JavaScript, template output, or ecommerce scale may be involved
At that point, the job is not to guess. It is to identify which signals are blocking indexing, which pages are being consolidated away, and whether the issue is localised or structural.
That is where structured SEO audit services can help. A proper review can separate one-off page fixes from wider crawl and indexing problems affecting visibility and enquiries. If you need broader support, technical SEO consulting can help clarify whether the underlying issue is architecture, crawl waste, template output, or duplication. If you are comparing options first, it also helps to understand what affects technical SEO audit cost.
FAQs
Why is my page not indexed even though it is live?
Because live and indexable are not the same thing. The page may be blocked, duplicated, weakly linked, canonically overridden, or judged too low-value to include.
How long does Google take to index a page?
There is no fixed timeline. Some pages are indexed quickly. Others take longer, and some are never indexed. Discoverability, crawl access, technical signals, and page value all influence the outcome.
Can a page rank if it is not indexed?
No. A page must be indexed before it can rank in Google search results.
Does robots.txt stop indexing?
It can stop crawling, which can interfere with indexing. It is not the same as a noindex tag, but it can still cause indexing problems by preventing Google from accessing the page properly.
Can poor internal linking cause indexing problems?
Yes. Weak internal linking reduces discovery and weakens importance signals. Orphan pages are especially vulnerable.
Can duplicate pages affect indexing?
Yes. When pages overlap too heavily, Google may choose one version, ignore the others, or treat the whole set as lower value.
Does requesting indexing in Search Console help?
Sometimes, but only when the page is already indexable and worth indexing. A manual request can speed up discovery in some cases. It does not solve noindex directives, weak content, duplication, broken canonicals, or wider crawl issues.
Why would a page drop out of Google after being indexed?
A page can drop out when Google reassesses its value or receives stronger competing signals elsewhere. Common causes include duplication, content changes that weaken the page, new canonical conflicts, internal links being removed, rendering problems, or sitewide template changes that alter indexability.
Next step
If an important page is not indexing, start by confirming what Google is actually seeing before changing templates, rewriting content, or requesting indexing repeatedly. In many cases the answer sits in a small set of checks: index status, canonical handling, internal links, crawl access, and page uniqueness.
If the issue affects several pages, or keeps returning after individual fixes, it usually makes sense to review the wider technical setup. That is where technical SEO consulting, a clearer view of technical SEO audit cost, or more structured SEO audit services become useful.