SEO Research Framework

Official-Source SEO Framework

This document is now evolving into a reusable SEO research framework. Green blocks are official guidance from Google Search Central and official OpenAI documentation. Yellow blocks are high-quality field evidence: observed SERP behaviour, reputable studies, and practitioner reporting that help explain what actually happens in the wild. Blue is reserved for conclusions and contradiction notes that synthesise the green and yellow evidence.
Scope: SEO foundations first
Green = official guidance
Yellow = field evidence
Blue = conclusions / contradictions
Workflow: official baseline → field evidence → ranking
LLM layer: kept separate from classic SEO

Method & source standard

Working rules
We are not carrying over assumptions from the earlier brief. This framework starts fresh. Each SEO area is separated into its own section so we can later assign evidence-based importance for (1) classic Google SEO and (2) LLM visibility. The framework now uses three layers: green for official documentation, yellow for strong field evidence, and blue for conclusions and contradiction notes that synthesise the evidence.
Green blocks
Official guidance from Google Search Central and official OpenAI documentation/help content.
Yellow blocks
High-quality field evidence: first-party observation, live SERP behaviour, and reputable studies or reporting from trusted search publications.
Blue blocks
Conclusions, synthesis, contradiction handling, and decision rules based on the green and yellow evidence.
Two scoring lenses — one shared framework
Green — official framework rule
Google says AI features like AI Overviews and AI Mode use the same fundamental SEO best practices, and that there are no special extra requirements just to appear in them. OpenAI separately documents crawler access, search inclusion, and noindex behaviour for ChatGPT search contexts. That means the foundation is shared, while access, measurement, and visibility controls are not identical across systems.

Sources:
Google — AI features and your website
OpenAI — Publishers & Developers FAQ
OpenAI — Overview of crawlers
Blue — document decision rule
This framework uses one shared set of SEO rules, then scores each topic through two separate lenses: Google Search impact and LLM / AI search impact. We are not building two completely separate playbooks. We are using one core framework, then judging how strongly each topic matters in each environment.
Overarching writing rule — useful, natural, non-spammy language
Green — official writing standard
Google says content should be created primarily to benefit people, not to manipulate search rankings. It also says pages should provide original, helpful, descriptive information, and that using generative AI does not change the standard: the output still needs to add value, remain accurate, and avoid scaled low-value content.

Sources:
Google — Creating helpful, reliable, people-first content
Google — Using generative AI content on Search
Blue — document-level decision rule
Every section in this framework should be written in natural language that sounds useful to a real person, not engineered purely to carry keywords. Avoid repetitive phrasing, robotic wording, interchangeable template copy, and any sentence that feels like it exists only for SEO. If the writing feels spammy, thin, or obviously manufactured for rankings, it fails the standard even if the keywords are technically present.
Current state
Starter structure only. No rankings locked yet.
Next step
Review each section one by one and assign importance based on official evidence.

Official source register

Approved base set
Yellow-note source standard
Approved field evidence types
Search Engine Land, Search Engine Roundtable, and methodology-backed studies from vendors like Ahrefs or Semrush are allowed for yellow notes when they are clearly observational and not dressed up as doctrine.
What yellow does not mean
Random SEO folklore, anonymous X threads, vague agency opinions, or recycled “best practice” lists without data or clear SERP observation.
How contradictions are handled
If green and yellow appear to conflict, the framework should note the tension explicitly, explain why the conflict may exist, and then make a practical decision rule. Green sets the baseline. Yellow influences testing and implementation.

1. Title tag / title links

Official + field evidence
Title tags are a high-value SEO input, but the title Google actually displays in search results is not guaranteed to match the HTML <title> exactly. This section separates the official rules from the real-world rewrite behaviour.
Google Search impact
Pending — to be scored after the section evidence is complete.
LLM / AI search impact
Pending — We will score this topic separately for classic Google Search and LLM / AI search once the evidence is complete.
Official baseline
Google says title links are critical to helping users understand and choose results. It recommends a title on every page, descriptive and concise text, and warns that Google may form the displayed title link from multiple sources, not just the HTML <title>.
Street-level takeaway
In practice, Google rewrites displayed titles often enough that title strategy cannot be reduced to “write a tag and you’re done.” Alignment across title, H1, and visible page topic matters because Google regularly reaches for those other signals.
Green — official findings
What Google officially says
1. Title links are important.
Google says title links are often the primary piece of information people use to decide which result to click.

2. Every page should have a title element.
Google explicitly recommends specifying a <title> for every page.

3. Descriptive and concise is the standard.
Google recommends descriptive, concise titles and warns against vague labels like “Home” or “Profile”.

4. There is no official title length limit.
Google says there is no limit on how long a <title> can be, although the displayed title link may be truncated to fit device width.

5. Google can build title links from multiple sources.
Google says it may use the <title>, the main visual title, heading elements such as H1, og:title, other prominent on-page text, and references to the page from elsewhere on the web.

6. Google has also published a historical usage benchmark.
In its September 2021 Search Central blog update, Google said HTML title elements were being used as the titles shown in search results for the vast majority of pages, around 87% at that time after system adjustments. This is useful context, but it is a dated snapshot rather than a permanent rule.

7. Rewrite triggers are documented.
Google calls out half-empty titles, obsolete titles, inaccurate titles, micro-boilerplate, repeated boilerplate, competing headings, and language/script mismatches as common reasons it may generate a different title link.

8. If crawling is blocked, titles can still be generated from elsewhere.
Google says blocked pages can still be indexed if discovered via links, and off-page anchor text may then be used to generate the title link. OpenAI similarly says public websites can appear in ChatGPT search, and content must not block OAI-SearchBot if you want it included in summaries and snippets.
Green working rules
Write for clarity first.
Use one clear page concept, not a stack of near-duplicate keyword variants.

Keep title and page in sync.
The title should match the page’s visible main topic and not promise something the page does not actually deliver.

Do not cling to a fake character rule.
Prioritise front-loaded meaning over arbitrary 55–60 character formulas.

Uniqueness matters.
Repeated or templated titles create rewrite risk and weaken page differentiation.

Crawling access still matters for LLM visibility.
For ChatGPT summaries/snippets, OpenAI says not to block OAI-SearchBot.
Do not overstate
Google does not publish a hard character limit for title tags, and it does not guarantee that your HTML title will be the exact title shown in Search. Any practical display-width guidance belongs in yellow, not green.
Yellow — field evidence deep dive
Finding 1 — rewrite rates vary, but rewrites are common
The exact percentage depends on sample, date, and method, which is why street evidence should be treated as observational rather than universal law. But the pattern is consistent: rewrites are common. Ahrefs published an earlier figure of 33.4% title rewrites and noted Google used the H1 50.76% of the time when it ignored the title tag. A Search Engine Land Q1 2025 study reported Google changed 76% of title tags. SISTRIX also documents that Google can replace titles when the page title does not match the query or when the title is too generic, too long, stuffed, or boilerplate-heavy. The percentages do not match, but the conclusion does: displayed titles are more fluid than many SEO checklists imply.
Finding 2 — H1 alignment matters in the real world
Search Engine Roundtable reported Semrush data showing the number of search results displaying the HTML title tag dropped 77% on average in that 2021 sample, and roughly 75% of those overrides used the H1 instead. Even if the sample was small and early, it strongly supports a practical rule: your H1 should not fight your title tag.
Finding 3 — many rewrites are subtle, not catastrophic
Field reporting also warns against overreacting. Search Engine Roundtable noted that not all overwrites are dramatic; some just remove a brand name or make a small wording adjustment. So “Google rewrote the title” does not always mean the title strategy failed. It often means Google preferred a tighter or more query-matched version for the SERP.
Finding 4 — titles are getting even more dynamic
In March 2026, Search Engine Land reported Google had confirmed a small, narrow test of AI-generated headline rewrites in traditional Search. That matters because it pushes yellow evidence one step further: not only can Google select from multiple existing title sources, it may also experiment with generating alternative wording to better match queries and engagement patterns.
Yellow practical rules
1. Treat title tags as inputs, not guarantees.
You still optimise them, but you should expect Google to judge whether they deserve to be shown as-is.

2. Keep the title, H1, and first visible page framing closely aligned.
This does not guarantee no rewrites, but it reduces conflict between the page’s main signals.

3. Watch live SERPs, not just source code.
The HTML title is only part of the story. Real optimisation means checking what Google is actually displaying.

4. Don’t obsess over a single rewrite percentage.
The studies disagree on the number, but they agree on the behaviour. Use them directionally, not dogmatically.

5. Build titles that still make sense if the brand is removed.
A lot of real-world rewrites compress or simplify. The page topic should still be unmistakable without relying on the suffix.
Confidence and limits
Yellow evidence is useful because it reflects the SERP users actually see. But it is not doctrine. The samples differ, Google’s systems change, and some studies focus on matching behaviour rather than ranking impact. So yellow should guide implementation priorities and testing, not replace Google’s official guidance.
Green vs yellow tension
Where the tension appears:
Google’s official material says title elements are used for the vast majority of pages and gives detailed best practices for influencing title links. Yellow studies often report much higher visible rewrite rates in sampled SERPs.

Why this is not a clean contradiction:
The numbers are measuring different things across different periods. Google’s 87% figure was a historical system snapshot from 2021. Field studies often count any displayed wording change across their chosen query sets, devices, or result types.

Decision rule:
Treat the HTML title as a high-value input, not a guaranteed output. Write it carefully, align it with the H1 and visible page framing, and then verify what Google actually displays.
Best current synthesis
The safest synthesis is: Google’s official guidance tells you how to write strong title tags; field evidence tells you why alignment and live SERP checking matter. So title tags are still worth serious effort, but they should be written as part of a broader title system that includes the H1, visible page framing, and how the page is referenced internally and externally.
Open question for later ranking
Once we deep-dive intro/body copy and heading structure, we can decide whether title tags sit above or below visible page copy in overall SEO importance. Right now, title tags clearly belong in the top tier, but not as a simple standalone mechanic.

2. H1 / main page heading

Official + field evidence
H1s matter most as a page-understanding and alignment signal. Official guidance stresses semantics and clarity. Field evidence suggests H1s usually do not make or break rankings on their own, but they still matter enough that missing, messy, or conflicting H1s are worth cleaning up.
Google Search impact
Pending — to be scored after the section evidence is complete.
LLM / AI search impact
Pending — We will score this topic separately for classic Google Search and LLM / AI search once the evidence is complete.
Official baseline
Google treats heading elements as part of how it understands page structure, content priorities, and the main visible title of a page. The emphasis in the official material is clarity, semantics, and page understanding — not rigid matching rules.
Street-level takeaway
In practice, H1s behave more like a strong clarity signal than a heavyweight standalone ranking lever. Most reputable field sources do not treat H1s as make-or-break, but they do treat missing, weak, or conflicting H1s as a meaningful quality issue.
Green — official findings
What Google officially says
1. Headings help Google understand page context and priorities.
Google has said that one way its algorithms determine the context of content on a page is by looking at the page's headings, and that semantic use of h1, h2, and h3 helps it understand the priorities of a site's content.

2. Heading elements should be used semantically.
Google explicitly recommends using heading elements for headings, instead of styling paragraph text to look like headings. In other words, the markup should reflect the role the text plays on the page.

3. Headings should create hierarchy and make navigation easier.
Google says to use heading elements to emphasize important text and help create a hierarchical structure, making it easier for users and search engines to navigate through the document.

4. Clear and meaningful headings matter for visibility.
In Google's mobile-first indexing guidance, it says missing meaningful headings may negatively affect visibility in Search because Google might not be able to fully understand the page. Google also recommends using the same clear and meaningful headings on mobile as on desktop.

5. The H1 does not need to be an exact match to the title tag or URL.
In Google SEO Office Hours, Google said the URL, page title, and H1 do not need to be exactly the same. It also separately said title tags do not need to match the H1 exactly — you should do whatever makes sense from a user's perspective.

6. The main visible heading can influence title links.
Google says heading elements such as h1 can be used as a source when generating title links. It also warns that if a page has multiple large, prominent headings and it isn't clear which text is the main title, Google may use the first heading as the title link.

7. Text should exist as HTML, not only inside images.
Google has said that placing headline text only inside the image file should be avoided because it makes it harder for search engines and some users to recognize. If the headline is visually integrated with imagery, the text should still exist as HTML.
Green working rules
Use a real heading element for the page headline.
Do not fake headings with styled div or p tags when the text is functioning as the main page heading.

Make the H1 clear and specific.
The main heading should tell users and Google what the page is actually about, not just gesture vaguely at the topic.

Keep heading hierarchy logical.
Use headings to reflect structure and emphasis, not just visual styling choices.

Do not force exact-match duplication.
The H1 can overlap with the title tag and URL, but it does not need to be word-for-word identical.

Avoid ambiguous competing headlines.
If multiple headings look equally important, Google may not know which one represents the page's main title.

Preserve meaningful headings on mobile.
The mobile version still needs the same clear heading signal because that version is used for indexing.
What Google does not explicitly require here
In this official-source pass, Google does not state that the H1 must exactly match the title tag, and it does not state that the URL, page title, and H1 must be identical. The official emphasis is on clarity, semantics, and user sense-making rather than rigid duplication rules.
Yellow — field evidence deep dive
Finding 1 — reputable field sources do not treat H1 as a critical standalone ranking lever
Search Engine Land reported that Google’s systems are not tripped up by multiple H1s and that rankings can work fine with or without them. Search Engine Roundtable separately reported John Mueller saying H1 tags are not critical for search ranking, and that headings generally won’t make or break rankings. Street-level consensus from credible reporters is therefore much softer than old-school SEO folklore: H1s matter, but usually as a structure and clarity signal rather than a magic ranking switch.
Finding 2 — the industry still treats H1 problems as worth fixing
Semrush’s on-site SEO study reported that 20% of analyzed sites had multiple H1 tags, 20% were missing H1 tags, and 15% had duplicate information in the title tag and H1. Semrush’s current Site Audit issue sets and API still track missing H1, multiple H1, and duplicate H1/title patterns. That suggests a practical industry view: H1 issues may not be catastrophic, but they are common enough and meaningful enough to be part of routine quality control.
Finding 3 — one clear H1 is still the cleanest operating standard
Field guidance is more relaxed than the old “exactly one H1 or you’re doomed” rule, but it still lands in a practical place: one clear H1 is cleaner and easier to maintain. Search Engine Land’s HTML/SEO guidance notes that multiple H1s or no H1s will not trip up Google, but clear semantic headings still help search engines understand pages and help users navigate them. In practice, teams keep aiming for one obvious main heading because it reduces ambiguity and makes audits easier.
Finding 4 — H1/title alignment matters because Google may lean on the H1 elsewhere
Ahrefs’ keyword optimisation guidance says the H1 should be consistent with the title tag, even if the wording is slightly different. Search Engine Land’s title tag guide makes the same street-level point from a different angle: when Google pulls a replacement title link, one of the best places it can turn is the H1. So even if H1 is not a major standalone ranking factor, it becomes strategically important because it supports title coherence, page understanding, and how the page may be represented in search.
Yellow practical rules
1. Default to one clear H1 per page.
Not because Google demands a perfect one-H1 rule, but because one obvious main heading is the cleanest signal for users, audits, and templates.

2. Don’t panic about multiple H1 edge cases.
If a CMS, accessibility pattern, or component library creates more than one H1, it is usually not a crisis. Check whether the page is still clear and coherent.

3. Fix missing, empty, or vague H1s.
These are low-drama, worthwhile improvements because they strengthen page understanding and reduce ambiguity.

4. Keep H1 and title aligned enough to tell the same story.
They do not need to match exactly, but they should clearly point to the same page topic.

5. Treat H1 as a high-value clarity signal, not a silver bullet.
It is worth getting right, but it usually sits below broader content quality and indexability issues in priority.
Confidence and limits
Yellow evidence for H1 is less study-heavy than title-tag evidence. A lot of the credible material is field reporting, practitioner consensus, and audit data rather than controlled experiments showing direct ranking lift. That makes the practical conclusion fairly stable, but it should still be treated as directional rather than mathematically precise.
Blue — synthesis / decision rule
Official guidance and field evidence point to the same practical conclusion: H1s are important enough to get right, but not important enough to obsess over as a standalone ranking trick. Use one clear, descriptive H1 when possible. Keep it semantically correct and aligned with the title and page topic. Fix obvious H1 problems, but do not treat them as the first thing to worry about if deeper content, indexing, or intent-matching problems are unresolved.
Blue — green vs yellow tension
Where the tension appears:
Green guidance stresses that headings help Google understand the page, while yellow reporting repeatedly says H1s are not critical and won’t make or break rankings.

Why this is not a real contradiction:
These statements are talking about different levels of influence. H1 can be a meaningful page-understanding signal without being a dominant independent ranking factor.

Decision rule:
Always give the page a clear H1. Just don’t mistake that for the main reason the page will win or lose in search.

3. Meta description / snippets

Official + field evidence
Meta descriptions matter most as snippet suggestions and click-shaping copy, not as guaranteed SERP text. The official material frames them as optional snippet inputs and control tools. Field evidence will help us judge how much they influence click behaviour in practice and how often Google ignores or rewrites them.
Google Search impact
Pending — to be scored after the section evidence is complete.
LLM / AI search impact
Pending — We will score this topic separately for classic Google Search and LLM / AI search once the evidence is complete.
Official baseline
Google treats the meta description as a possible snippet source, not a guaranteed output. The official emphasis is on helping Google show a useful preview to users, while giving site owners ways to improve, limit, or suppress snippet text when needed.
Street-level takeaway
In practice, meta descriptions are worth writing for important pages, but mainly as click-shaping copy rather than reliable SERP text. Google rewrites or replaces them so often that the real job is to improve the chances of showing a strong snippet when Google does use them.
Green — official findings
What Google officially says
1. Snippets are primarily generated from page content.
Google says it primarily uses the content on the page itself to automatically determine the appropriate snippet. It may also use the meta description element when that gives users a more accurate description than content taken directly from the page.

2. Meta descriptions are suggestions, not guarantees.
Google says it will sometimes use the <meta name="description"> tag, and in Google SEO Office Hours it has said it is not guaranteed to use the particular meta description you wrote for a page. Snippets are auto-generated and can vary by query.

3. The meta description is meant to summarise and interest.
Google describes a meta description as a short, relevant summary of what a page is about. It says they are like a pitch that helps convince the user the page is exactly what they are looking for.

4. There is no official meta description length limit.
Google says there is no limit on how long a meta description can be, but snippets are truncated as needed, typically to fit device width.

5. Unique, page-specific descriptions are recommended.
Google says identical or very similar descriptions across pages are not helpful when individual pages appear in search results. It recommends using page-level descriptions wherever possible, while site-level descriptions may be acceptable for the home page or other aggregation pages.

6. Prioritisation is acceptable if writing every description is impractical.
Google says if you do not have time to create a description for every page, prioritise critical URLs such as the home page and popular pages.

7. Relevant page details can be included in the description.
Google says the meta description does not have to be limited to a plain sentence and can include useful details such as author, publication date, price, manufacturer, or other page-specific information that may not otherwise appear in the snippet.

8. Programmatic generation is allowed and encouraged in the right context.
For large database-driven sites, Google says programmatic generation of descriptions can be appropriate and is encouraged, as long as the descriptions are human-readable, diverse, and based on page-specific data.

9. Keyword strings and low-quality descriptions are discouraged.
Google says descriptions made of long strings of keywords do not give users a clear idea of the page's content and are less likely to be displayed as snippets. Its official blog also says empty, duplicate, off-topic, low-quality, or spammy meta descriptions are often ignored.

10. Google gives snippet control options beyond the meta description itself.
Google says you can prevent a snippet entirely with nosnippet, limit snippet length with max-snippet, and block specific sections from appearing in snippets with data-nosnippet.
Green working rules
Write descriptions as snippet candidates, not ranking levers.
The official goal is to help Google show a good preview and help users decide whether to click.

Keep descriptions specific to the actual page.
Recycled boilerplate across many pages weakens usefulness and makes Google less likely to rely on the tag.

Summarise the page clearly and concretely.
The description should help a user understand what is on the page, not just echo keyword variants.

Include genuinely relevant details where helpful.
Product, article, and listing pages can use the description to pull together key facts that may otherwise be scattered on the page.

Programmatic does not mean robotic.
If descriptions are generated at scale, they still need to be readable, diverse, and page-specific.

Use snippet controls intentionally.
If a page should not show a snippet, or only limited preview text, Google provides official controls for that.
What Google does not explicitly guarantee here
In this official-source pass, Google does not guarantee that your written meta description will be shown, and it does not give a hard character limit. It also frames the description as a snippet and click-preview aid rather than a guaranteed ranking mechanism. An older official Google blog post states that description meta tags have no effect on rankings and can instead improve clickthrough by producing a better snippet.
Yellow — field evidence deep dive
Finding 1 — Google rewrites meta descriptions most of the time
The strongest field-study signal here is that Google often does not show the written meta description. Ahrefs found Google rewrote meta descriptions 62.78% of the time in its study, which means the hardcoded description appeared only about 37% of the time on average. That lines up with the lived SEO reality that snippet writing matters, but control over the final display is limited.
Finding 2 — Meta descriptions still matter because CTR is the real battleground
Search Engine Land and SISTRIX both frame title-and-description optimisation as a major click lever inside the SERP. The practical logic is simple: once a page is ranking, the snippet is one of the biggest variables affecting whether the searcher clicks your result or someone else's. So meta descriptions matter less as ranking mechanics and more as conversion copy inside Google.
Finding 3 — Query intent changes how often Google keeps or replaces the description
Ahrefs found Google was slightly more likely to rewrite descriptions for long-tail queries than for fat-head terms, at 65.62% versus 59.65%. That supports the practical view that the more query-specific the search, the more likely Google is to build a custom snippet from page content rather than rely on the generic written description.
Finding 4 — Length best practice exists, but it does not control rewrite risk
Field evidence does not support the idea that simply keeping a description within a traditional character target makes Google keep it. In the Ahrefs study, rewrite rates were very similar for descriptions that were too long and those that were not. Search Engine Land's snippet coverage also reinforces that Google changes displayed snippet lengths over time, which is why rigid character-count folklore should be treated cautiously.
Finding 5 — Missing descriptions are common, but high-value pages still benefit most
Ahrefs found 25.02% of top-ranking pages had no meta description at all, which is another reminder that the tag is not a prerequisite for ranking. But the same study still recommends prioritising descriptions for pages that already get traffic, are designed to rank, or carry strong click value. That is a very practical rule for real sites: coverage does not need to be universal, but important URLs deserve better snippet copy.
Yellow practical rules
1. Write meta descriptions for valuable pages first.
Home pages, product pages, service pages, and articles with real impression volume deserve attention before low-value URLs.

2. Treat them as SERP ad copy, not page copy.
Their practical job is to make the result clearer and more clickable when Google chooses to show them.

3. Do not over-optimise around a hard character number.
Aim for concise, informative copy, but do not mistake old length formulas for control over the final snippet.

4. Expect Google to replace the description for many query variants.
Strong page content still matters because Google often pulls snippet text from the body instead.

5. Avoid boilerplate and lazy automation.
If you generate descriptions at scale, they still need to be distinct enough to compete when they do show.
Blue — synthesis / contradictions / decision rule
Blue synthesis
The strongest synthesis is: meta descriptions are not a strong SEO ranking lever, but they are still a meaningful SERP performance lever. They help most at the point where ranking has already been won and the next fight is click-through. That means they belong below title tags and below core visible page content in most SEO hierarchies, but above “nice to have” details on pages that actually matter.
Blue contradiction / tension
There is no major contradiction between green and yellow here. Google says meta descriptions are optional snippet suggestions and not ranking factors; the field evidence mostly reinforces that. The only tension is one of emphasis: the official material can make them sound secondary, while field evidence shows they still matter for CTR and SERP competitiveness on important pages.
Decision rule
Write meta descriptions for pages where impressions and clicks matter, write them to improve clarity and click appeal, and do not judge them by whether Google always shows them verbatim. Judge them by whether they improve the page's ability to win the click when they are shown.
Likely scoring direction
Google Search impact: mid-tier to high-mid-tier, mainly through snippet quality and click behaviour rather than ranking.

LLM / AI search impact: low, except where snippet controls or description text influence how pages are previewed in traditional search surfaces that feed discovery.

4. Intro copy / first visible body content

Official + field evidence
Google does not give the first paragraph a special standalone ranking rule. But its official guidance repeatedly points to clear, useful, visible text near the top of the page as a major part of helping Search understand the page, generate snippets, and satisfy users. This section focuses on what Google officially says about opening copy, before we layer in field evidence.
Google Search impact
Likely high. This section now has both green and yellow evidence, and it is shaping up as one of the stronger on-page elements for classic Google SEO.
LLM / AI search impact
Likely very high. This section now has both green and yellow evidence, and passage-level retrieval makes visible body copy especially important for AI search.
Official baseline
Google’s official material points to visible body copy as a core understanding layer. The opening copy matters not because Google gives the “first paragraph” a magic bonus, but because prominent, helpful, crawlable text helps Google understand what the page is about, helps users confirm they are in the right place, and can directly influence snippets and AI search visibility.
Yellow status
Field evidence now reinforces the official view, with one important nuance: strong opening copy matters most when it is extractable and useful, but it is not the only passage on the page that can drive relevance.
Green — official findings
What Google officially says
1. Google wants people-first content, not search-engine-first copy.
Google says content should be created primarily to help people, not primarily to manipulate search rankings. That directly affects how opening copy should be written: useful, clear, and genuinely informative rather than robotic or keyword-stuffed.

2. Use the words people would use, and place them in prominent locations.
Google Search Essentials says to use the words people would use to look for your content and place those words in prominent locations on the page, such as the title and main heading, along with other descriptive locations. That supports strong, query-aligned opening copy even though Google does not single out “sentence one” as a formal rule.

3. Important content should exist in textual form.
In Google’s AI features guidance, Google says important content should be in text on the page, because text is easier for its systems to process than content embedded only in images or videos. That makes visible written copy near the top of the page especially valuable for both classic Search understanding and AI search surfaces.

4. Google primarily generates snippets from page content.
Google says snippets are primarily created from the actual content on the page. This means visible body copy can directly affect how the page is previewed in search results, especially when Google does not use the meta description.

5. Helpful content should provide substantial, original value.
Google’s helpful content guidance asks whether the page provides original information, reporting, research, or analysis; whether it offers substantial value compared with other pages in search results; and whether it leaves the reader feeling they learned enough to achieve their purpose. Opening copy should set that standard immediately rather than wasting the top of the page on filler.

6. Clear structure and meaningful headings support understanding.
Google’s SEO Starter Guide and mobile-first indexing guidance both reinforce that content should be easy to read and well organised, and that meaningful headings help users and Google understand the page. Intro copy works best when it clearly introduces the topic and is supported by descriptive headings rather than vague generic openings.

7. AI search guidance points toward unique, non-commodity opening copy.
In Google’s AI search guidance, Google says to focus on unique, non-commodity content that visitors from Search will find helpful and satisfying. That raises the bar for opening copy: generic templated intros are less aligned with what Google says works well in AI search experiences.
Green working rules
Open with clarity, not scene-setting fluff.
The first visible copy should help a user and Google understand what the page offers in plain language.

Match the page to searcher vocabulary naturally.
Use the language real users would search for, but write it in a way that still sounds human and specific to the page.

Keep key meaning in crawlable HTML text.
Do not rely on images, sliders, or decorative hero treatments to communicate the page’s main point.

Make the top of the page earn its space.
The opening copy should carry real informational value, not just repeat the title in slightly different words.

Let intro copy reinforce the title and H1.
The title, H1, and first visible body copy should all point to the same main page concept without sounding repetitive.

Favour specific, helpful language over templated SEO prose.
Google’s people-first and AI-search guidance both push toward useful, distinctive copy rather than interchangeable intros.
What Google does not explicitly guarantee here
Google does not publish a rule saying the first paragraph or first sentence gets a special ranking bonus. It also does not require exact-keyword usage in the opening sentence. The official signal is broader: prominent, helpful, understandable, text-based page content written for people.
Yellow — field evidence deep dive
Finding 1 — answer-first copy tends to punch above its weight
Field guidance around featured snippets and AI retrieval keeps pointing in the same direction: concise, directly phrased answer blocks are easier for search systems to extract and reuse. Search Engine Land's featured snippet guidance explicitly recommends defining the topic concisely and answering the question directly. Ahrefs' featured snippet work reinforces that paragraph snippets are one of the most common snippet formats and that Google extracts them from already-ranking pages. In practice, this gives strong opening copy and strong section intros an outsized chance of being surfaced.
Finding 2 — opening copy matters, but it is not the only copy that matters
Passage-ranking discussion in the field cuts against the simplistic "all importance sits in the first paragraph" idea. Search Engine Roundtable's coverage of John Mueller's explanation of passage ranking says Google is not ranking an isolated paragraph on its own, but can rank a page because a relevant passage deeper in the page matches the query. That means good intro copy helps, but buried useful sections can still carry a page into relevance.
Finding 3 — AI search pushes the value of extractable opening copy even higher
Recent Search Engine Land coverage on AI search is very consistent here: answer-first, information-rich, well-structured passages are more likely to be surfaced, reused, and cited. That does not mean every page must open with a dictionary definition, but it does mean the first visible copy block often has a disproportionate opportunity to become the page's most reusable passage — especially when it cleanly states what the page is about and who or what it serves.
Finding 4 — structure beats waffle
Field guidance for both classic snippets and AI search keeps rewarding clear, self-contained blocks over fluffy scene-setting intros. The practical pattern is not "keyword in sentence one at all costs." It is "make the early copy easy to interpret, easy to extract, and easy to trust." Pages that bury the actual answer below generic throat-clearing are more likely to waste the top of the page than benefit from it.
Yellow practical rules
1. Make the first visible block do real work.
It should define, answer, or clarify the page topic quickly — not just repeat the H1 with filler around it.

2. Write opening copy so it can stand alone as an extractable passage.
A strong intro should still make sense when lifted into a snippet, citation, or AI answer context.

3. Do not confuse "answer-first" with "one-paragraph SEO."
Passage ranking means later sections can still become the page's relevance driver for specific queries.

4. Use the same discipline in later sections.
The intro is important, but so are sub-sections that answer narrower intents cleanly.

5. Avoid decorative opening copy.
Generic brand fluff or scene-setting that delays the actual point tends to underperform both for extractability and for user confirmation.
Confidence and limits
Yellow evidence here is stronger on behaviour than on a single clean statistic. There is no one famous industry number for "how much the first paragraph matters." The case is cumulative: featured snippets favour concise answers, passage ranking allows deeper sections to surface, and AI-search practitioners keep observing that clean, extractable passage design works better than vague top-of-page prose.
Blue — synthesis / contradictions / decision rule
Blue synthesis
The strongest synthesis is: intro copy does not get a magical algorithmic bonus, but it often has outsized practical influence because it is the page's first chance to establish topic clarity, provide extractable answer text, and confirm relevance for both users and machines. In other words, its importance is operational rather than mystical.
Blue contradiction / tension
The apparent tension is this: green says Google does not give the first paragraph a special rule, while yellow says answer-first opening copy often performs especially well. Those ideas can both be true. Google is evaluating the whole page, but the opening copy frequently becomes the most visible and reusable passage — unless a stronger section deeper in the page better matches the query.
Decision rule
Treat the first visible body block as high-value real estate. Write it to clearly explain the page in human language, make it strong enough to stand alone as a snippet or citation, and then support it with equally clean, intent-specific sections below. Do not force every page into the same exact intro formula, but do make sure the top of the page earns its prominence.
Likely scoring direction
Google Search impact: likely high, because visible copy helps with topic clarity, snippet extraction, and user confirmation even if it is not a standalone ranking switch.

LLM / AI search impact: likely very high, because extractable, answer-first passages are central to how AI systems surface and reuse content.
Sources

5. Heading structure (H2 / H3)

Official + field evidence
Heading structure is one of those topics where official guidance and field evidence mostly agree on the big picture, but differ on emphasis. Google treats H2/H3 structure as a page-understanding aid. Field evidence suggests it is rarely a make-or-break ranking lever on its own, but often becomes materially helpful when it improves extractability, passage matching, snippet eligibility, and machine-readable structure.
Google Search impact
Pending final score — likely above H1 structure in practical usefulness for long-form content, but still below core content quality and page intent.
LLM / AI search impact
Pending final score — likely stronger than its classic SEO reputation suggests because clear section headings help systems map subtopics, passages, and reusable chunks.
Official baseline
Google's official material treats supporting headings as a page-structure and understanding signal. H2 and H3 headings help show what each section is about, how sections relate to each other, and what the priorities of the page are. The official emphasis is less about rigid formulas and more about meaningful, semantic structure that helps both users and Google understand the document.
Street-level takeaway
In practice, H2/H3 structure usually will not rescue weak content or overpower stronger relevance signals. But it consistently shows up as part of content that earns featured snippets, People Also Ask visibility, passage-level retrieval, and AI-friendly extractability. So while headings may not be a starring ranking factor, they often help good content become easier for search systems to use.
Green — official findings
What Google officially says
1. Google uses headings to understand context and priorities.
In Google's own site-clinic guidance, it says one way its algorithms determine the context of content on a page is by looking at the page's headings, and that semantic markup such as h1, h2, and h3 helps Google understand the priorities of the content.

2. Semantic HTML should be used according to its purpose.
Google says semantic markup means using markup according to its meaning and purpose, including using heading elements h1 to h6 to mark up headings. That supports using real heading elements for section headings rather than styling generic text blocks to look like headings.

3. Clear and meaningful headings matter on mobile too.
Google's mobile-first indexing guidance says to use the same clear and meaningful headings on mobile as on desktop, and warns that missing meaningful headings may negatively affect visibility in Search because Google might not be able to fully understand the page.

4. Correct heading markup helps Google understand content and context.
In Google SEO Office Hours, Google said using semantic HTML correctly can help search engines better understand the content of a page and its context. It specifically used headings as an example, saying they are a clear sign that content fits under that heading.

5. Heading help is real, but not magical.
In that same Office Hours guidance, Google also said headings are not a secret path to number one rankings. Their value is interpretive: if Google has trouble understanding what a section means, headings provide a clear summary.

6. Visible heading text can feed other search systems too.
Google's title-generation guidance says it considers visible headline text, including text in H1 or other header tags, when generating title links. While that is not an H2/H3 ranking rule, it reinforces that visible structured heading text is part of how Google interprets and presents pages.
Green working rules
Build a real section hierarchy.
Use headings to divide the page into meaningful sections and subsections, not just to create bigger text.

Make section headings descriptive.
H2 and H3 headings should help a user understand what each section actually covers, not just add vague labels like “More” or “Details”.

Use semantic heading tags for headings.
If something functions as a section heading, mark it up as a heading rather than a styled paragraph or generic div.

Keep mobile and desktop heading structure aligned.
Do not simplify or strip meaningful headings on mobile in ways that make the page harder for Google to understand.

Think in summaries.
Each heading should summarise the section below it clearly enough that Google and users can anticipate what that section is about.

Use common sense over rigid obsession.
Google's own language here is pragmatic: heading structure matters, but it should serve clarity rather than become a mechanical box-ticking exercise.
What Google does not explicitly guarantee here
Google does not publish a rule saying a perfectly nested H2/H3 hierarchy creates a direct ranking boost on its own. It also does not treat heading structure as a substitute for helpful page content. The official signal is broader: headings help Google understand the page, especially when they are clear, semantic, and descriptive.
Yellow — field evidence deep dive
Finding 1 — heading clean-up rarely acts like a major standalone ranking lever
Field reporting around Google's public comments is pretty consistent here: heading structure matters, but usually not in a dramatic, isolated way. Search Engine Roundtable covered John Mueller saying headings are great for SEO and accessibility but are not going to make or break rankings, and later covered him saying fixing heading order is unlikely to visibly change ranking outcomes on its own. That supports a practical reading: poor heading structure is worth fixing, but it should not be mistaken for the main performance bottleneck on most pages.
Finding 2 — clear H2/H3 structure repeatedly shows up in snippet and answer extraction guidance
Search Engine Land's featured snippets guidance explicitly recommends a clear, matching H2 or H3 plus a short intro sentence beneath it, and its broader writing-for-SEO guidance says semantic headings help machines identify what each part of the page is about and whether it qualifies for featured snippets, voice answers, or AI-generated responses. This does not prove H2/H3 tags are strong ranking factors by themselves, but it does suggest they often help systems extract and reuse the right passage.
Finding 3 — heading structure helps chunking and query-to-section matching
Recent Search Engine Land guidance on query fan-out and content chunking argues that clear headings, lists, tables, and content chunks make it easier for AI systems to map sub-queries to specific sections of a page. That is especially relevant for long-form pages where different sections may answer different user intents. In other words, H2/H3 structure may be modest as a classic ranking tweak, but quite valuable as a retrieval and extractability aid.
Finding 4 — practitioners still use headings to reinforce topic coverage and readability
Ahrefs' on-page and checklist guidance treats H2/H3 headings as natural places to organise subtopics, improve skimmability, and include topic language where it genuinely belongs. Their advice is not that headings magically boost rankings, but that good heading structure makes pages easier to read, easier to audit, and easier for search systems to interpret section by section.
Yellow practical rules
1. Use H2s to define real sections, not decorative labels.
Each H2 should map to a meaningful subtopic or intent slice on the page.

2. Use H3s to break complex sections into extractable sub-units.
This is especially useful when one section covers steps, comparisons, FAQs, or multiple angles.

3. Write headings that match likely user questions or subtopics when natural.
This helps with skimming and can improve extractability for snippets, PAA, and AI answers.

4. Do not force keywords into every heading.
Heading structure works best when it clarifies the page rather than turning every subheading into an awkward keyword bucket.

5. Treat heading structure as a low-risk enhancement with compounding upside.
It improves readability, accessibility, auditing, and section-level machine understanding even when the ranking impact is hard to isolate.
Confidence and limits
Yellow evidence here is more about repeated practical observation than about one definitive numerical study. There is no clean “H2/H3 improves rankings by X%” dataset. The stronger pattern is indirect: clear heading structure repeatedly appears in guidance for snippet capture, content chunking, passage extraction, and AI-friendly page design.
Blue — synthesis / contradictions / decision rule
Blue synthesis
The strongest synthesis is: H2/H3 structure is usually not a major standalone ranking lever, but it is often a meaningful usability and extractability lever. That makes it more valuable than pure “nice to have” formatting, especially on pages where multiple subtopics, questions, or query variants need to be understood section by section.
Blue contradiction / tension
There is only a light tension here. Green says headings help Google understand sections and priorities. Yellow says heading fixes alone usually will not transform rankings. Those ideas fit together: headings matter most as supporting structure, not as a substitute for relevance, depth, or usefulness.
Decision rule
Use clear H2/H3 structure by default. It is low-risk, improves clarity for users, helps machines parse the page, and can increase the odds that good content gets extracted or matched to the right query. Even if the isolated ranking benefit is modest, it is usually worth doing because the downside is low and the structural upside is real.
Likely scoring direction
Google Search impact: mid-tier supporting factor — more important on long, multi-section pages than on short simple pages.

LLM / AI search impact: high-mid-tier supporting factor — clear headings help chunking, extractability, and subtopic retrieval even when they are not dominant ranking factors in classic SEO.

6. Content quality / people-first content

Official + field evidence
This section is about what Google officially means by high-quality content and what strong field evidence suggests actually works in practice. The core question is not whether the copy sounds polished. It is whether the page is genuinely useful, reliable, distinctive enough to add value, and satisfying enough that it deserves to compete.
Google Search impact
Likely top-tier. Content quality sits near the centre of Google's own guidance and of most real-world algorithm impact analysis.
LLM / AI search impact
Likely top-tier. Helpful, extractable, distinctive content is also one of the clearest predictors of whether AI systems can reuse, summarise, and trust a page.
Official baseline
Google repeatedly frames content quality around helpful, reliable, people-first content. It tells site owners to focus on content that satisfies users, adds meaningful value, and is written for people first rather than to manipulate rankings. Google also says the same core guidance carries into AI search features.
Yellow status
Field evidence strongly supports the official direction, with one important nuance: AI assistance itself is not the practical problem. Commodity, low-effort, weakly differentiated output is.
Green — official findings
What Google officially says
1. Create helpful, reliable, people-first content.
Google’s Search Essentials and helpful-content guidance put this at the centre of good SEO. It recommends content made primarily to benefit people, not content created mainly to attract search traffic.

2. Use the words real people use, in prominent places.
Google says to use the words people would use to look for your content and place those words in prominent locations on the page. That supports clarity and matching, but it is not a licence for keyword stuffing.

3. Quality means substance, originality, and satisfaction.
Google’s self-assessment questions ask whether the content provides original information, reporting, research, or analysis; whether it offers substantial, complete, or comprehensive value; and whether someone would leave feeling they learned enough to achieve their goal.

4. Search-engine-first content is explicitly discouraged.
Google warns against producing lots of content on many topics mainly because they seem to perform well in Search, summarising other sources without adding value, or writing to chase traffic rather than meet a user need.

5. Core update recovery guidance is mostly content guidance.
In Google’s core update documentation, the recommended response is not a technical trick. It is to assess the content honestly, improve it meaningfully, and focus on quality rather than short-term ranking workarounds.

6. Generative AI does not change the quality standard.
Google says AI-generated content is not automatically bad, but using AI or automation to generate many pages without adding value can violate its spam policies. The quality bar stays the same whether content is human-written, AI-assisted, or automated.
How this carries into AI search
1. Google says the same best practices apply to AI features.
Google’s AI features documentation says there are no additional special requirements to succeed in AI Overviews and AI Mode beyond the existing advice in Search Essentials.

2. Unique, non-commodity content matters more in AI search.
In Google’s 2025 AI search guidance, it says to focus on making unique, non-commodity content that visitors from Search will find helpful and satisfying. That strengthens the case that bland, templated copy is a weak long-term strategy even if it is technically optimised.

3. Important content should be present as text on the page.
Google says AI systems process textual content more easily than information locked inside images or other non-text formats. So quality is not just what you know; it is whether the useful part is actually accessible on the page.
Green working rules
Write to satisfy intent, not to simulate SEO.
The page should clearly help the person who arrived there and not read like it exists only to carry search terms.

Add real value beyond the obvious.
Don’t just restate what every other page says. Include specifics, distinctions, examples, useful detail, or firsthand knowledge where appropriate.

Avoid commodity, interchangeable copy.
If the same paragraph could sit on ten competitor pages with only the product name changed, it is probably too generic.

AI assistance is fine; unreviewed filler is not.
If AI helps draft or structure the content, the final page still needs accuracy, usefulness, and clear value for users.

Keep important value in plain text.
Critical selling points, differentiators, and explanations should not live only in images, sliders, or visual design elements.
Do not overstate
Google does not provide a fixed word count, a public “content quality score”, or a formula where adding more copy automatically improves rankings. “People-first” does not mean long-form by default; it means genuinely useful, reliable, and satisfying for the page’s purpose.
Yellow — field evidence deep dive
Finding 1 — the practical risk is low-value automation, not AI use itself
Field evidence is increasingly consistent on this point. Search Engine Land reported in 2025 that Google quality raters were being directed to treat some automated or generative-AI main content as Lowest quality when it is clearly low-effort or low-value. But Ahrefs' 2025 ranking study found that 86.5% of the top 20 ranking pages showed some level of AI-generated or AI-assisted content. The practical conclusion is not “AI is bad.” It is that cheap, interchangeable output is risky, while edited, useful AI-assisted content can rank perfectly well.
Finding 2 — content-score checklists are weak proxies for actual performance
Ahrefs' 2025 study comparing five popular content-optimisation scores found only weak correlations between those scores and rankings. That matters because it undercuts a common shortcut mindset: heavily tuned score-chasing copy may look "optimised" without actually being the strongest result. In practice, content quality appears broader than checklist saturation.
Finding 3 — helpful-content style updates often remove weak sites more than they promote perfect ones
SISTRIX's analysis of the 2023 Helpful Content Update noted that the pattern was more about removing visibility from weaker domains than producing obvious winners. That matches how many practitioners describe quality updates in the wild: thin, derivative, or search-engine-first content can drag a site down, while strong content quality is often more about earning resilience than getting a simple "boost".
Finding 4 — AI-era visibility still seems to reward clarity, review, and trust signals
Search Engine Land's practical AI-search coverage repeatedly points toward the same pattern: answerable, reviewed, trustworthy content tends to be more usable by search and AI systems than bloated filler. That does not turn quality into a measurable formula, but it reinforces the real-world importance of editorial review, factual confidence, and clear differentiation.
Yellow practical rules
1. Publish fewer weak pages, not more average pages.
Field evidence around helpful-content-style updates suggests that scaling mediocre content is a liability, not a moat.

2. Use AI as an accelerator, not as a substitute for judgment.
Drafting, structuring, and summarising are fine. Publishing bland, unreviewed output at scale is the real risk.

3. Stop treating optimisation scores as a finish line.
A page can score well in a tool and still be generic. Real competitiveness still depends on usefulness, specificity, and differentiation.

4. Prioritise trust and edit quality where the page carries commercial or informational weight.
The more valuable the page, the less acceptable it is for the content to feel templated, vague, or error-prone.

5. Distinctive content is defensive as well as offensive.
It does not just help a page win; it also helps a site avoid being lumped in with low-value, search-engine-first content.
Confidence and limits
Yellow evidence on content quality is strong directionally but rarely neat numerically. Many of the most credible observations come from update analysis, case studies, and broad ranking studies rather than controlled experiments. That means the practical conclusion is strong, even if the exact weight of each subfactor is hard to isolate.
Blue — synthesis / contradictions / decision rule
Blue synthesis
The strongest synthesis is: content quality is not just another on-page factor — it is one of the main conditions that determines whether the rest of the SEO work has anything worth amplifying. Titles, headings, links, and schema can sharpen a page, but weak, generic, or low-trust content limits what those signals can achieve.
Blue contradiction / tension
The biggest tension is this: Google warns against low-value automation, while yellow evidence shows that AI-assisted pages can rank extremely well. That is not a real contradiction once the quality standard is applied. The practical dividing line is not AI versus human; it is helpful versus commodity, reviewed versus unreviewed, and distinctive versus interchangeable.
Decision rule
Assume quality work is worth doing even when it is harder to quantify. If a choice improves usefulness, specificity, trust, and user satisfaction without creating spam or bloat, the blue summary should lean toward doing it. In practice that means editing harder, differentiating more clearly, and resisting the temptation to publish generic filler just because it is fast.
Likely scoring direction
Google Search impact: critical or near-critical. This is one of the strongest recurring themes in both official guidance and real-world update behaviour.

LLM / AI search impact: critical or near-critical. AI systems also depend heavily on content being clear, extractable, useful, and trustworthy enough to reuse.

8. URLs / slugs

Official + field evidence
URL work sits right in the middle of SEO myth and SEO hygiene. Official guidance is mostly about readable, stable, crawl-friendly structures. Field evidence helps answer the practical question: are clean slugs a material lever, or mostly something worth doing because they reduce friction and confusion?
Google Search impact
Likely supporting to important. URL structure matters most for clarity, crawl efficiency, duplication control, and long-term site hygiene. Exact slug wording looks low-weight compared with content and links.
LLM / AI search impact
Likely supporting. Clean, stable URLs help discoverability and site organisation, but there is little evidence that slug wording itself is a major AI-visibility lever.
Official baseline
Google recommends a simple URL structure that is intelligible to humans. In official guidance, good URLs are descriptive when possible, use the audience’s language, use hyphens to separate words, avoid unnecessary parameters, and minimise crawl problems caused by duplicate or exploding URL variations.
Street-level takeaway
Field evidence broadly agrees that URLs matter, but mostly as a low-risk support system rather than a powerful standalone ranking trick. Clean slugs are worth doing because they help clarity and reduce technical mess. Changing existing URLs purely for keyword gain usually is not worth the disruption.
Green — official findings
What Google officially says
1. Google recommends a simple, human-intelligible URL structure.
Google says a simple URL structure helps both Google Search and users better understand a site. It recommends URLs that are constructed logically and are intelligible to humans, not just to systems.

2. Use descriptive, readable words when possible.
Google explicitly recommends using readable words rather than long ID numbers in URLs when possible. This is presented as a clarity best practice, not as a promise of direct ranking gain.

3. Use your audience’s language.
Google says to use words in your audience’s language in the URL, and if applicable use transliterated words. This is a strong official hint that URLs should align with how real users think and search.

4. Use hyphens instead of underscores.
Google recommends using hyphens (-) rather than underscores (_) to separate words because hyphens help users and search engines identify concepts in the URL more easily.

5. Use as few parameters as you can.
Google recommends shortening URLs by trimming unnecessary parameters when possible, especially parameters that do not change content.

6. Case and duplication issues matter.
Google says URLs are case sensitive, and if your server treats uppercase and lowercase the same, convert all text to the same case so Google can more easily determine that the URLs reference the same page.

7. Overly complex URL patterns can waste crawl budget and create indexation issues.
Google warns that overly complex URLs with many parameters can create unnecessarily high numbers of URLs pointing to identical or similar content, which can waste crawl resources and make full indexing harder.

8. Some URL patterns are officially flagged as problematic.
Google calls out additive filters, irrelevant parameters, shopping sort parameters, session IDs, infinite calendars, and broken relative links as common URL-related problems. It recommends avoiding session IDs in URLs where possible and using robots.txt or other control methods to manage problem spaces.
Green working rules
Prefer short, readable slugs over system-looking URLs.
If a human can understand the page topic from the URL, that is usually closer to Google’s official preference than opaque IDs or bloated parameter strings.

Use words that match the audience and page topic naturally.
Do not force every keyword variation into the slug, but do make the page topic obvious.

Default to lowercase and hyphen-separated words.
That aligns with Google’s official guidance and reduces duplication or parsing mess.

Avoid unnecessary parameters and crawl traps.
Filtering, sorting, session IDs, calendars, and infinite combinations should be treated as technical risk areas rather than harmless URL details.

Stability matters.
Even though Google’s URL guidance is mostly about clarity and crawlability, unnecessary slug changes can create avoidable redirect and canonicalisation work later.
Do not overstate
Google does not say that putting an exact-match keyword in the slug will by itself produce strong ranking gains. The official guidance is much more about readability, logic, crawl control, and avoiding URL structures that create duplication or waste crawl resources.
Yellow — field evidence deep dive
Finding 1 — keywords in URLs look low-weight in practice
Field reporting keeps landing in the same place: words in URLs are not meaningless, but they are very light signals compared with the page’s actual content and links. Search Engine Roundtable’s January 2025 reporting on John Mueller said URLs provide only minimal additional signals for search engines. Earlier reporting on Mueller’s comments described words in URLs as a very lightweight factor, and one that becomes even less important once Google has indexed and understood the page.
Finding 2 — clean URLs are still widely treated as worthwhile
Serious practitioner guidance still recommends short, descriptive slugs because they improve readability, make pages easier to recognise and share, and reduce structural mess. Search Engine Land’s 2025 overview frames clean URLs as a fundamental support element even in AI-driven search, but not as a silver bullet. Ahrefs similarly treats the main keyword or topic in a slug as good practice without claiming a significant ranking boost.
Finding 3 — the bigger risk is changing URLs carelessly
Field evidence is much harsher on unnecessary URL changes than on imperfect slugs. Practitioners generally agree that changing established URLs purely to squeeze in a keyword can create redirect chains, lost equity, and measurement confusion that outweigh any small theoretical benefit. The practical lesson is that URL hygiene matters more at creation time than as a late-stage optimisation obsession.
Finding 4 — URL work compounds through site architecture
The most useful field pattern is not about one slug ranking better. It is about the way clean, consistent URLs make canonicalisation, internal linking, crawl paths, segmentation, and reporting easier to manage at scale. That means URL quality often helps indirectly by making the rest of the SEO system cleaner and more reliable.
Yellow practical rules
1. Use short, descriptive slugs when creating new pages.
This is usually the easy win: low risk, high clarity, little downside.

2. Do not chase URL rewrites for tiny keyword gains.
If a page is already live and indexed, changing the slug purely for SEO is often not worth the redirect and stability cost.

3. Prioritise architecture over wordsmithing.
Logical folders, stable paths, and duplicate-control matter more than squeezing one more keyword into a slug.

4. Treat ugly URLs as a hygiene problem first.
If a URL is unreadable, parameter-heavy, or duplication-prone, fix it because it creates friction and technical mess, not because it is a magic ranking unlock.

5. Make slugs useful even without brand or context.
The page topic should be obvious from the path itself, but still natural and concise.
Confidence and limits
Yellow evidence here is less about large causal studies and more about repeated, consistent practitioner observation. That makes the direction fairly trustworthy: URL keywords are light, readability and site hygiene matter, and reckless URL changes are often worse than imperfect slugs. But this is still not the kind of section where field data proves a huge direct ranking effect.
Green vs yellow tension
Where the tension appears:
Green guidance spends more time on readability, duplication control, and crawl efficiency than on ranking gains. Yellow evidence agrees with that, but many practitioners still instinctively treat keywords in slugs as more powerful than they really are.

Why this is not a real contradiction:
The official and field views mostly align. The difference is emphasis. Green explains what good URL structure looks like. Yellow clarifies that the biggest wins are architectural and hygienic, not direct ranking magic from the slug itself.

Decision rule:
For new URLs, do the clean, descriptive version by default. For existing URLs, only change them when the structural or user-facing benefit is clear enough to justify the disruption.
Best current synthesis
URLs matter most as a support system. They help users understand where they are, help search engines crawl and deduplicate more cleanly, and reduce technical mess across the site. That makes clean slugs worth doing, but mostly because they are low-risk, clarity-improving, and operationally smart — not because they are a heavyweight ranking lever by themselves.
Likely scoring direction
Google Search: Supporting to important, especially at site-architecture level.
LLM / AI search: Supporting.
Practical weight: Better to do clean URLs from the start than try to retrofit them later, but do not treat slug wording as a top-tier optimisation lever.
Sources

9. Canonicalisation

Official + field evidence
Canonicalisation is about telling Google which URL should be treated as the representative version when multiple URLs show duplicate or very similar content. Officially, this is less about chasing rankings and more about consolidating signals, reducing duplication confusion, and helping Google choose the version you actually want shown. Field evidence helps show how often canonical mistakes quietly create indexing, duplication, and signal-splitting problems in practice.
Google Search impact
Pending — to be scored after the section evidence is complete.
LLM / AI search impact
Pending — We will score this topic separately for classic Google Search and LLM / AI search once the evidence is complete.
Official baseline
Google says you can indicate a preferred canonical URL for duplicate or very similar pages, but none of the methods are strictly required. If you do not specify a canonical preference, Google will identify the version it considers objectively best for Search. The official purpose of canonicalisation is to consolidate duplicate signals and help Google choose the version you want represented.
Street-level takeaway
In practice, canonicalisation is rarely a growth trick by itself, but it is often the difference between a clean index and a messy one. Bad canonicals, parameter duplication, or conflicting signals can quietly waste crawl budget, split signals, and push Google toward the wrong URL.
Green — official findings
What Google officially says
1. You can specify a preferred canonical URL, but it is not mandatory.
Google says you can indicate your preferred canonical for duplicate or very similar pages, but none of the canonicalisation methods are required. If you do not specify one, Google will choose what it considers the best version to show in Search.

2. Canonical signals have different strengths.
Google explicitly says redirects are a strong signal that the target should become canonical, rel="canonical" link annotations are also a strong signal, and sitemap inclusion is a weak canonical signal.

3. Consistency matters more than mixing conflicting methods.
Google warns against specifying different canonical URLs for the same page via different methods, such as one URL in the sitemap and another via rel="canonical". Conflicting signals make canonicalisation less reliable.

4. Google recommends choosing one canonical implementation method and sticking to it.
For rel="canonical", Google supports both the HTML link element and the HTTP header method, but recommends choosing one and using it consistently because mixing both is more error-prone.

5. Some commonly misused tools are not canonical tools.
Google explicitly says not to use robots.txt for canonicalisation purposes, not to use the URL removal tool for canonicalisation, and not to rely on noindex to choose a canonical page within a site. It says rel="canonical" is the preferred solution for that use case.

6. Internal links and canonicals should align.
Google recommends linking internally to the canonical URL rather than a duplicate URL. Consistent internal linking helps Google understand your canonical preference.

7. Implementation details matter.
Google says the HTML canonical element must appear in the <head>, and recommends using absolute URLs rather than relative paths for canonical annotations. It also supports canonical HTTP headers for non-HTML files such as PDFs.

8. HTTPS preference is part of canonical handling.
Google says it prefers HTTPS pages over equivalent HTTP pages as canonical in many situations, and recommends reinforcing that preference with redirects and canonical annotations.
Green working rules
Choose a representative URL for duplicate or near-duplicate sets.
Do not leave duplicate URL families completely ambiguous if there is a clear preferred version.

Use redirects for retired URLs and rel="canonical" for duplicates that need to stay live.
That follows Google’s own signal hierarchy more closely than trying to solve everything with one tool.

Keep canonical signals aligned.
Sitemaps, internal links, canonicals, and protocol preferences should point in the same direction where possible.

Use absolute canonicals and valid placement.
Put canonical link elements in the <head> and use absolute URLs to reduce implementation mistakes.

Treat canonicalisation as technical clarity, not copywriting.
This is fundamentally a duplicate-control and signal-consolidation task.
Do not overstate
Google does not present canonicalisation as a direct ranking trick. The official value is that it helps consolidate duplicate or near-duplicate URLs, reduces mixed signals, and improves Google’s understanding of which version should represent the content. It is a technical control layer, not a substitute for strong content.
Yellow — field evidence deep dive
Finding 1 — canonical errors are usually quiet but can cause real indexing problems
The strongest field pattern is not that canonicals create ranking lifts, but that bad canonical setups can create surprisingly large indexing problems. Search Engine Land warns that multiple canonical tags, canonicals pointing to non-indexable pages, and other canonical conflicts can lead to indexing issues rather than clean consolidation. In other words, canonicalisation is often more defensive than offensive.
Finding 2 — Google frequently chooses a different canonical than the one site owners nominate
Ahrefs’ documentation around the Search Console status Duplicate, Google chose different canonical than user reinforces the practical reality that Google will override weak or unconvincing canonical preferences. Ahrefs also notes that Google uses many canonicalization signals together, which fits the idea that a lone canonical tag is not always enough if the rest of the site points elsewhere.
Finding 3 — parameter duplication and URL sprawl are common real-world causes
SISTRIX notes that URL parameters are often responsible for the majority of internal duplicates, and Search Engine Roundtable reported Google advice encouraging site owners to reduce duplication by serving the canonical URL directly wherever possible. In practice, canonicalisation problems often start upstream with messy URL generation rather than with the canonical tag itself.
Finding 4 — rel canonical is usually the safer duplicate-control tool than noindex
Search Engine Roundtable’s reporting on John Mueller’s comments reinforces the field-side rule that for duplicate content, rel="canonical" is generally the better signal than noindex. That matches the practical experience of many technical SEOs: noindex can remove a page from the index, but it does not express that two URLs should be consolidated into one representative version.
Yellow practical rules
1. Treat canonicalisation as cleanup with real consequences.
It may not feel glamorous, but bad canonical signals can quietly distort which URLs get indexed.

2. Fix duplication at the source where possible.
Parameter chaos, alternate paths, and needless duplicate URLs are often the bigger problem than the missing tag itself.

3. Make the whole site vote for the same URL.
Canonicals work better when internal links, redirects, sitemaps, and protocol choices all reinforce the same preferred version.

4. Do not rely on one lonely canonical tag to override the rest of the site.
If everything else points to a different version, Google may ignore the nominated canonical.

5. If duplicate families exist, it is usually better to clean them up than leave them ambiguous.
This is one of those low-risk technical fixes that is often worth doing even when the upside is mostly preventive.
Confidence and limits
Yellow evidence for canonicalisation is more operational than experimental. There are fewer neat “ranking lift” studies here, but there is strong cross-source agreement that wrong canonicals, duplication, and conflicting signals create crawl and index problems in the real world. That makes the practical direction fairly trustworthy even without a simple headline metric.
Green vs yellow tension
Where the tension appears:
Green guidance frames canonicalisation as a duplicate-control and signal-consolidation system rather than a direct ranking lever. Yellow evidence often sounds more urgent because canonical mistakes can cause real indexing and URL-selection problems.

Why this is not a real contradiction:
These views are compatible. Canonicalisation is not a classic “boost” factor, but it can still materially affect search performance by determining which version gets indexed and consolidated.

Decision rule:
If duplicate or near-duplicate URLs exist, do the canonical cleanup anyway. It is usually a low-risk technical fix that improves clarity for Google and reduces the chance of the wrong URL winning.
Best current synthesis
Canonicalisation is best thought of as index and signal hygiene with meaningful downside if neglected. It is not something to obsess over on perfectly simple sites with no duplication issues, but as soon as duplicate URL families, parameters, alternate versions, or migration leftovers exist, it becomes a very worthwhile layer to get right.
Likely scoring direction
Google Search: Important, and near-critical on sites with duplication, parameter, or migration complexity.
LLM / AI search: Supporting to important where duplicate URLs affect discoverability, indexing, or which URL becomes the representative source.
Practical weight: Usually worth doing when the risk exists, because the cost of ambiguity can be larger than the effort to clean it up.
Sources

10. Robots / noindex / crawl controls

Official + field evidence
This section covers the controls that decide whether crawlers can fetch a URL, whether engines can index it, and how much of it can be shown in results. This is one of the easiest places for site owners to get tripped up because robots.txt, noindex, nosnippet, and X-Robots-Tag do very different jobs.
Google Search impact
Likely critical. If these controls are wrong, important pages can vanish from Search or become unreadable to Google.
LLM / AI search impact
Likely critical. For ChatGPT search in particular, crawler access and readable noindex logic directly affect whether content can be surfaced, summarised, or reduced to link-only visibility.
Official baseline
Official documentation is very explicit here: robots.txt controls crawling, not indexing. noindex is the standard way to keep a page out of Google’s index, but crawlers must be allowed to fetch the page to see that instruction. OpenAI’s docs follow the same pattern: allow OAI-SearchBot if you want summaries and snippets, and use noindex if you want to prevent inclusion entirely.
Yellow status
Field evidence mostly reinforces the official guidance rather than contradicting it. The real-world lesson is that these controls are powerful, brittle, and easy to misconfigure at scale.
Green — official findings
What Google and OpenAI officially say
1. robots.txt controls crawling, not indexing.
Google says robots.txt is not a mechanism for keeping a page out of Search. A blocked URL can still be indexed if Google discovers it through links or other references.

2. Use noindex to prevent indexing.
Google says noindex is the directive that blocks a page from appearing in search results. But Google must be able to crawl the page to see that directive.

3. X-Robots-Tag extends these controls to non-HTML files.
Google says X-Robots-Tag is the correct mechanism for files like PDFs or images when you need indexing or snippet controls outside HTML pages.

4. Snippet controls are separate from indexing controls.
Google supports directives like nosnippet, max-snippet, and data-nosnippet to limit or suppress what can be shown in results without necessarily deindexing the page.

5. Don’t rely on JavaScript to undo noindex.
Google warns that when it sees noindex, it may skip rendering, so trying to remove that directive with JavaScript is unreliable.

6. OpenAI splits bot controls by use case.
OpenAI says OAI-SearchBot controls inclusion in ChatGPT search results, while GPTBot is separate. If a page is disallowed but discovered elsewhere, OpenAI may still show the link and page title only; use noindex to block that entirely.
Green working rules
Use robots.txt for crawl management, not deindexing.

Use noindex when you actually need a page out of Search.

Let bots fetch pages that need to carry noindex or snippet directives.

Keep bot controls simple and easy to audit.

For ChatGPT summaries/snippets, do not block OAI-SearchBot unless you intentionally want to opt out.
Yellow — field evidence deep dive
Finding 1 — robots mistakes cause outsized damage
Search Engine Land’s robots.txt guidance frames the file as a genuine SEO landmine because one misplaced character can block important sections or entire environments from crawling. Its SEO debugging coverage makes the same point more broadly: a single crawl-control error can create site-wide visibility problems that look like “ranking” issues but are really access issues.
Finding 2 — the robots.txt vs noindex confusion is still one of the most common technical errors
Field guides on indexability and crawlability keep repeating the same warning because it still causes real-world problems: if you block a page in robots.txt, Google may never see the noindex tag on that page. In practice, this means sites often accidentally preserve low-value URLs in the index or accidentally hide the wrong content from crawlers.
Finding 3 — simple, auditable controls beat clever setups
Practical debugging writeups tend to show that complexity creates risk. Large stacks of crawl rules, mixed staging/production rules, JS-based tag changes, and conflicting directives create more problems than they solve. The field lesson is not that these controls are mysterious; it is that they are easy to misuse.
Yellow practical rules
1. Keep robots.txt lean and intentional.
2. Use noindex for exclusion, not disallow.
3. Audit these controls any time visibility drops suddenly.
4. Treat ChatGPT/OAI-SearchBot access as a separate explicit decision, not an accidental side effect of old bot rules.
Blue synthesis / contradiction note
There is very little genuine contradiction here. Green and yellow both point the same way: these controls are foundational, and errors here can override or nullify a lot of good SEO work elsewhere. The practical move is to keep them simple, use the right control for the right job, and avoid “just in case” blocking that prevents engines from reading the page properly.
Blue decision rule
If a page should be searchable, don’t block the bot that needs to read it. If a page should not be searchable, use the directive built for exclusion. In contradiction cases, default to the cleaner, more readable control stack rather than the cleverer one.

11. XML sitemaps

Official + field evidence
Sitemaps are often either overrated as a ranking lever or underrated as technical infrastructure. Officially, they are a discovery and prioritisation hint, not a replacement for strong site architecture.
Google Search impact
Likely supporting to important. Sitemaps rarely make a weak site rank, but they can materially help discovery, crawl efficiency, and URL prioritisation.
LLM / AI search impact
Likely supporting. They help bots discover URLs, but they are secondary to crawl access, indexability, and the actual usefulness of the content.
Official baseline
Google says a sitemap is a file that tells search engines which pages and files you think are important, and it can help Google crawl the site more efficiently. But Google also says sitemap submission is only a hint, not a guarantee of crawl or indexation.
Yellow status
Field evidence mostly agrees: sitemaps are best understood as a safety net and efficiency layer, not as a substitute for good internal linking or a direct ranking mechanic.
Green — official findings
What Google officially says
1. Sitemaps help Google crawl more intelligently.
Google says sitemaps can help crawlers find pages, videos, images, and other files more efficiently.

2. Sitemaps are especially useful when discovery is hard.
Google highlights them as particularly useful for large sites, new sites, rapidly changing sites, pages not easily discovered through links, and non-textual content.

3. Only include canonical URLs you actually want shown in Search.
Google says the URLs in your sitemap should be the preferred canonical versions, not every duplicate path to the same content.

4. Submission is a hint, not a guarantee.
Google says submitting a sitemap does not guarantee the sitemap will be downloaded or that all listed URLs will be crawled or indexed.

5. Google doesn’t care about URL order in the sitemap.
It explicitly says the order of URLs in a sitemap does not matter.

6. CMS-generated sitemaps are usually fine.
Google recommends letting your CMS generate the sitemap when possible.
Green working rules
Include only index-worthy canonical URLs.

Use a sitemap as a backup discovery system, not your main architecture.

Keep it clean and automatically updated if possible.

Use sitemap indexes when the site is too large for one file.
Yellow — field evidence deep dive
Finding 1 — sitemaps are a safety net, not a substitute
Search Engine Land’s sitemap and site-architecture guidance repeatedly makes the same point: a sitemap can help crawlers find URLs that are newer, isolated, or harder to discover, but it is not a replacement for internal linking and sound architecture.
Finding 2 — sitemap quality matters more than sitemap existence
Field guides emphasise that bad sitemap hygiene creates noise: blocked URLs, noindexed URLs, duplicate URLs, redirected URLs, and parameter junk make the sitemap less useful. In practice, a selective sitemap tends to be more valuable than an indiscriminate one.
Finding 3 — strong sites can survive weak sitemaps, but weak discovery systems benefit more
The field pattern is that well-linked sites often get crawled anyway, while newer or technically messy sites benefit more from a good sitemap. That makes sitemaps more important as site complexity or discovery friction increases.
Yellow practical rules
1. Keep sitemaps clean.
2. Don’t list URLs you block, noindex, or canonicalise away.
3. Use them as a second-chance discovery path.
4. Expect crawl help, not ranking magic.
Blue synthesis / contradiction note
There is no major conflict between green and yellow here. Both say sitemaps help discovery and crawl efficiency, but neither supports treating them as a direct ranking trick. The practical move is simply to maintain a high-quality sitemap because it is low-risk and helps the site communicate its preferred URLs more clearly.
Blue decision rule
Use a sitemap if you care about discoverability, but don’t mistake it for architecture. Build strong internal linking first, then use the sitemap as reinforcement.

12. Structured data / schema

Official + field evidence
Structured data tends to get oversold as a ranking lever and undersold as a clarity layer. Officially, it helps Google understand content and may make pages eligible for rich results. In practice, it is best viewed as a visibility, interpretation, and enhancement system — not a guaranteed rankings boost.
Google Search impact
Likely important supporting signal. Structured data rarely replaces strong content or architecture, but it can materially improve understanding and rich-result eligibility.
LLM / AI search impact
Likely supporting to important. It is not a citation guarantee, but it can make entities, relationships, and page purpose easier for AI systems to interpret.
Official baseline
Google says structured data helps it understand the content of a page and may enable rich results when the markup follows the appropriate guidelines. Google is also explicit that rich results are not guaranteed.
Yellow status
Field evidence supports that official framing. Schema is usually most valuable for eligibility, clarity, and richer SERP presentation — not as a direct ranking shortcut.
Green — official findings
What Google officially says
1. Structured data helps Google understand content.
Google says it uses structured data to understand what the page contains and to gather information about the web and the world more broadly.

2. Structured data can make pages eligible for rich results.
The benefit is often richer presentation rather than a direct ranking boost.

3. Rich results are not guaranteed.
Google explicitly says it does not guarantee that structured-data features will appear in Search even when markup is valid.

4. Use supported types and follow the guidelines.
Only supported schema types tied to actual on-page content should be used, and pages must follow both syntax and policy guidelines.

5. Validate with Google’s tools.
Google recommends the Rich Results Test and Search Console reporting to validate, monitor, and debug markup.

6. Google includes its own case-study evidence for visibility gains.
In Google’s structured-data documentation, Rakuten and Nestlé reported stronger engagement and clickthrough on pages that surfaced with search features or rich results.
Green working rules
Mark up what is actually on the page.

Use supported schema types only where they genuinely fit the content.

Prefer clean JSON-LD when possible.

Validate and monitor rather than assuming markup is working.
Yellow — field evidence deep dive
Finding 1 — schema improves visibility more reliably than rankings
Search Engine Land’s structured-data coverage consistently treats schema as a way to unlock rich results, increase SERP real estate, and improve click potential. That matches what practitioners tend to see: better presentation and clearer interpretation more often than dramatic direct ranking lifts.
Finding 2 — even without a rich result, schema can still help machines interpret the page
Search Engine Land previously reported John Mueller saying that even when structured data does not surface as a rich result, Google’s systems still benefit from understanding the page better. That makes schema more than pure decorative markup.
Finding 3 — schema’s AI-search role is clarity, not magic
Recent Search Engine Land analysis on AI search says schema does not guarantee citations, but it can help systems understand entities and reduce ambiguity. That is directionally useful, but still a support layer rather than the whole answer.
Yellow practical rules
1. Implement the schema that fits the page, not the schema you wish the page qualified for.
2. Treat schema as a clarity and eligibility layer.
3. Do it anyway when relevant because it is low-risk and helps machines parse the page more cleanly.
4. Don’t expect schema alone to rescue weak content.
Blue synthesis / contradiction note
Green and yellow align pretty well. The only real tension is with overhyped SEO folklore: official docs and strong field evidence both say schema matters, but mostly for interpretation, eligibility, and richer search experiences rather than as a blunt ranking factor. The practical move is to implement relevant, policy-compliant schema because the upside is real and the downside is low when done properly.
Blue decision rule
If a schema type genuinely matches the page and is supported by Google, use it. It is usually better to add valid markup and not “need” it than to skip a clean enhancement that helps search and AI systems understand the page.

13. FAQ content & FAQ schema

Official + field evidence
This section combines two related but different things: the usefulness of FAQ content itself, and the narrower role of FAQPage structured data. The content format has broad value. The rich-result markup is now much more limited in Google than many old SEO guides imply.
Google Search impact
Likely supporting to important. Good FAQ content can satisfy long-tail questions and help with extractable answers, but FAQPage markup itself is now much more situational.
LLM / AI search impact
Likely important. Clear question-and-answer formatting is naturally reusable by AI systems even when FAQ rich results are unavailable.
Official baseline
Google’s FAQPage documentation now makes an important restriction explicit: FAQ rich results are only available for well-known authoritative government-focused or health-focused sites. But Google still recommends clear, visible FAQ content when it is genuinely part of the page, and its question/answer format remains easy for systems to understand.
Yellow status
Field evidence suggests the content format is often more useful than the markup itself. Real FAQs and answer-first sections can help with People Also Ask, extractable passages, and AI reuse even when FAQ rich-result eligibility is absent.
Green — official findings
What Google officially says
1. FAQPage is only for pages with one authoritative answer per question.
If users can submit multiple answers, Google says to use QAPage instead.

2. FAQ rich results are now limited.
Google’s current documentation says FAQ rich results are only available for well-known authoritative government or health sites.

3. FAQ content must be visible on the page.
Google says the question and answer must be present on the page, though answers may be hidden in expandable sections that users can open.

4. Mark up only one sitewide instance of repeated FAQs.
If the same question and answer appears across multiple pages, Google says to mark up only one instance.

5. Validation and monitoring still matter.
Google recommends the Rich Results Test, URL Inspection, Search Console monitoring, and sitemap submission after deployment.
Green working rules
Use FAQ markup only when it genuinely fits the page and the page is eligible.

Use visible question-and-answer content even when markup won’t earn a rich result.

Write direct, complete answers instead of padding.
Yellow — field evidence deep dive
Finding 1 — the content format outlived the markup hype
Field evidence since Google’s FAQ rich-result restriction has shifted attention back to the content itself. Practitioners still find value in clear question-led sections because they map naturally to user queries, People Also Ask opportunities, and answer extraction — even when FAQPage markup has limited visual payoff.
Finding 2 — question-led content is highly reusable in AI search
Recent Search Engine Land coverage on AI-driven local search argues that detailed FAQs built from real customer questions help prevent third-party sources from filling the gaps. That points to FAQ content as a high-utility format for LLM retrieval, even if the schema is no longer broadly rewarded.
Finding 3 — FAQ markup is now a situational enhancement, not a universal recommendation
The practical field view is that FAQ schema is no longer something to spray across every commercial page. The better use is to focus on authentic question coverage, then add markup only where it is supported and appropriate.
Yellow practical rules
1. Build FAQ sections from real questions.
2. Prioritise answer quality over markup quantity.
3. Don’t add FAQ schema by habit on ineligible sites.
4. Keep the question-and-answer pattern because it is still useful for users and AI extraction.
Blue synthesis / contradiction note
This is one of the clearest examples of why the document separates content from markup. Green says FAQPage markup has narrow eligibility. Yellow says the FAQ format itself still has wide practical value. So the right move is not to abandon FAQs — it is to stop treating FAQ schema as the whole strategy.
Blue decision rule
Write strong FAQs where real questions exist. Use markup only when it genuinely fits Google’s current rules. In other words: keep the content pattern, be selective with the schema.

14. Images / alt text

Official + field evidence
Image SEO is often treated as a side quest, but Google’s guidance gives it more weight than that. Images can be a discovery path, nearby text changes how Google interprets them, and alt text matters for both accessibility and search understanding. The practical question is how much of this is worth systematic effort on normal business pages.
Google Search impact
Likely supporting. Image work rarely carries a page by itself, but it improves discoverability, accessibility, image search potential, and machine understanding.
LLM / AI search impact
Likely supporting. Descriptive image context and alt text help systems interpret visual elements more reliably, though they are rarely the main citation driver.
Official baseline
Google says images can be how people find your site for the first time. It recommends high-quality images near relevant text, descriptive alt text, and avoiding keyword stuffing in alt attributes. It also says that when images act as links, the alt attribute is treated like anchor text.
Yellow status
Field evidence generally supports the official position: image optimisation is useful, but mostly as a support layer. Alt text and surrounding context help machines understand images; they are not a hidden shortcut to main-web ranking dominance.
Green — official findings
What Google officially says
1. Images are a discovery channel.
Google’s SEO Starter Guide says people can find your site through images and visual search.

2. Put high-quality images near relevant text.
Google says surrounding text helps it understand what the image is about and how it relates to the page.

3. Alt text should be descriptive and contextual.
Google says to write useful, information-rich alt text and avoid stuffing alt attributes with keywords.

4. Missing alt text and stuffed alt text are both bad.
Google shows both as poor examples in its documentation.

5. Image links use alt text as anchor text.
Google’s crawlable links guidance says that when an image is the link, the image’s alt attribute acts as the anchor text for that link.

6. Keep image handling consistent across desktop and mobile.
Google’s mobile-first guidance says mobile images should follow image best practices and keep the same alt text as desktop versions.
Green working rules
Use descriptive alt text where images carry meaning.

Place useful images near relevant copy.

Do not keyword stuff alt text.

If an image is a link, make sure the alt text would work as anchor text.
Yellow — field evidence deep dive
Finding 1 — image work is usually a support layer, not a primary ranking lever
Search Engine Land’s image-optimization guidance treats alt text, compression, filenames, context, and format choice as worthwhile SEO work, but not as a substitute for content, links, or indexability. That fits what practitioners typically see.
Finding 2 — context matters as much as the attribute itself
Field practice keeps coming back to the same pattern Google documents officially: isolated alt text is weak. Images work better when the file, caption, nearby text, and page topic all reinforce the same meaning.
Finding 3 — scale work can be useful, but automation still needs guardrails
Practical SEO guidance increasingly discusses AI-assisted alt text generation, but the useful version is descriptive and reviewed, not stuffed or generic. The field takeaway is to use automation to speed up good work, not to spray low-quality descriptors across the site.
Yellow practical rules
1. Optimise important images, not just every image mindlessly.
2. Treat alt text as descriptive context, not keyword inventory.
3. Keep image optimisation in proportion to the page’s real goals.
4. Do the basics anyway because they help accessibility as well as search.
Blue synthesis / contradiction note
There is no strong contradiction here. Green and yellow both say images and alt text matter, but mostly as a support layer. The practical move is to do image basics consistently because they are low-risk, help accessibility, and improve machine understanding, while recognising that image SEO alone rarely determines whether a page wins.
Blue decision rule
Optimise the images that matter, write alt text like a human description, and keep the surrounding page context strong. It is usually better to do this well than to skip it just because it is not the top-ranking lever.

15. Page experience / Core Web Vitals

Official + field evidence
Page experience is one of the easiest areas to either overhype or dismiss. Google’s official guidance is nuanced: Core Web Vitals are used by ranking systems, but they are not a standalone shortcut to the top. Beyond Core Web Vitals, other page experience aspects matter more for overall satisfaction than as direct ranking triggers.
Google Search impact
Likely important, but rarely decisive on its own. Page experience can influence outcomes when content quality and relevance are already competitive.
LLM / AI search impact
Likely supporting. It matters more indirectly through usability, renderability, and overall site quality than as a direct AI-specific signal.
Official baseline
Google says there is no single page experience signal. Core Web Vitals are used by ranking systems, but Google also says trying to get a perfect score just for SEO may not be the best use of time. Relevance still comes first.
Yellow status
Field evidence generally agrees that page experience matters, but usually as a competitive tiebreaker, quality threshold, or compounding support signal rather than the main reason a page ranks.
Green — official findings
What Google officially says
1. There is no single page experience signal.
Google says its ranking systems look at a variety of signals aligned with page experience.

2. Core Web Vitals are used in ranking systems.
Google explicitly says CWV are used by its ranking systems and recommends achieving good CWV for Search success and user experience.

3. Relevance still comes first.
Google says it always seeks to show the most relevant content, even if the page experience is sub-par.

4. Perfect scores are not the goal.
Google says chasing a perfect score purely for SEO may not be the best use of time.

5. Beyond CWV, other page experience factors mostly help indirectly.
Google says aspects like HTTPS, mobile usability, intrusive interstitials, and ad experience can make pages more satisfying even when they are not direct ranking triggers in the same way.

6. Page experience is often evaluated page by page.
Google says its core systems generally evaluate related aspects on a page-specific basis, though some sitewide assessments exist.
Green working rules
Get pages into “good” CWV territory where practical.

Focus on meaningful experience improvements, not cosmetic score chasing.

Keep the site secure, mobile-usable, and free of intrusive friction.
Yellow — field evidence deep dive
Finding 1 — CWV matters, but usually in context
Search Engine Land’s page-experience and Core Web Vitals coverage frames these metrics as important but not dominant. That matches what many practitioners see: performance improvements help more when the content is already relevant and there is meaningful competition among similar pages.
Finding 2 — page experience often acts like a threshold or tie-breaker
Field commentary increasingly treats poor performance as something that can hold pages back, especially on mobile, rather than something that automatically destroys rankings. In practical terms, page experience often behaves like a quality threshold rather than a primary relevance engine.
Finding 3 — teams still waste time chasing vanity scores
A recurring field warning is that SEO teams can overinvest in tool scores instead of fixing the bottlenecks that users actually feel. This lines up closely with Google’s own warning about not optimising for perfect scores just for SEO reasons.
Yellow practical rules
1. Fix bad experience issues that users actually feel.
2. Improve CWV until pages are healthy, not necessarily perfect.
3. Don’t expect speed work alone to outrank better content.
4. Still do the work, because poor experience can quietly hold otherwise good pages back.
Blue synthesis / contradiction note
Green and yellow align well here too. Page experience matters, but not as a standalone substitute for relevance and quality. The practical takeaway is to improve experience enough that it is no longer a handicap, then stop treating performance scores as the centre of the SEO strategy.
Blue decision rule
Do the meaningful page-experience work anyway. It is low-risk, user-positive, and often the kind of thing you only notice when it is bad. Just keep it in proportion to content, indexability, and architecture work.

16. Indexability & discoverability

Official + field evidence
This is the “nothing else matters if this fails” section. A page can have great content, titles, headings, and links, but if engines cannot find it, fetch it, render it, or decide it belongs in the index, the rest of the work has limited value.
Google Search impact
Likely critical. Discoverability, crawlability, renderability, and indexability are foundational preconditions for almost everything else in SEO.
LLM / AI search impact
Likely critical. AI search systems still depend on access, readable content, and discoverable public URLs before they can cite or summarise anything.
Official baseline
Google’s own developer guidance says all pages should be reachable from another findable page and that sitemaps help Googlebot crawl more intelligently. That means discoverability is not optional plumbing; it is part of whether the page gets a chance at all.
Yellow status
Field evidence reinforces this almost without contradiction. When pages are not indexed, most downstream SEO work becomes irrelevant until the discovery or indexing issue is solved.
Green — official findings
What Google officially says
1. Pages should be reachable through crawlable links.
Google’s developer guide says every page should be reachable from another findable page through crawlable linking.

2. Sitemaps help Googlebot crawl more intelligently.
Google recommends building and submitting a sitemap to help discovery and crawling.

3. Important content should be in text form on the page.
Google’s AI-features guidance says key content should be provided in text because text is easier for systems to process than information embedded only in visuals or interaction layers.

4. Blocking and rendering issues can prevent correct indexing.
Google’s crawling, JavaScript, robots, and mobile-first documentation all show that inaccessible or mismatched content can cause indexing and understanding failures.

5. Search Console is the official debugging surface.
Google points site owners to URL Inspection, Core Web Vitals, indexing reports, sitemap reports, and other Search Console tools to diagnose why content may not be surfacing.
Green working rules
Make every important page discoverable by links and sitemap.

Keep essential content accessible as HTML text.

Debug access and indexing issues before polishing secondary on-page factors.
Yellow — field evidence deep dive
Finding 1 — most “SEO failures” at this layer are really discoverability or indexing failures
Search Engine Land’s guides on crawlability, indexability, and technical SEO all frame this as foundational. Pages cannot rank if they never get a real chance to be crawled and indexed properly.
Finding 2 — indexability issues often masquerade as content or ranking issues
In practice, teams often spend time improving copy or metadata on URLs that remain “crawled currently not indexed,” blocked, canonicalised away, or hard to discover. The field lesson is that debugging comes before polishing.
Finding 3 — this is one of the few areas where strong wording is justified
Even cautious field sources use emphatic language here: without crawlability and indexability, nothing else matters much. That makes this one of the clearest critical buckets in the framework.
Yellow practical rules
1. Check discoverability and index status early.
2. Treat “not indexed” as a core problem, not a reporting detail.
3. Ensure important pages are linked, crawlable, renderable, and canonicalised correctly.
4. Don’t waste time polishing pages engines still can’t properly use.
Blue synthesis / contradiction note
There is essentially no contradiction between green and yellow here. Both say this layer is foundational. The only nuance is that indexing is not guaranteed even when a page is crawlable, which is why discoverability, quality, duplication control, and rendering all intersect in this section.
Blue decision rule
When in doubt, solve discoverability and indexability before trying to win with polish. This is one of the few sections where “better to do it and not need it” clearly applies because the cost of neglect is so high.

17. LLM discovery / ChatGPT visibility

Official + field evidence
This section is the dedicated AI-search layer. It does not replace the core SEO framework; it sits on top of it. Officially, ChatGPT search has its own crawler logic and publisher controls. In practice, the main difference is not that SEO stops mattering — it is that access, clean citation, structured meaning, and answer-ready content become easier to notice.
Google Search impact
Indirect only. Most of this section is not about Google rankings, though many of the same underlying best practices overlap.
LLM / AI search impact
Likely critical. This section is specifically about whether public content can be discovered, surfaced, cited, and linked in ChatGPT search contexts.
Official baseline
OpenAI says any public website can appear in ChatGPT search. To be included in summaries and snippets, the site must not block OAI-SearchBot. If a page is disallowed but known from elsewhere, OpenAI may still show the link and page title only. To prevent that entirely, OpenAI points publishers to noindex.
Yellow status
Field evidence suggests AI visibility is not the same thing as Google visibility. Strong Google rankings do not guarantee recommendation or citation in AI systems, and the traffic payoff from AI visibility is often smaller than the visibility itself.
Green — official findings
What OpenAI and Google officially say
1. Public sites can appear in ChatGPT search.
OpenAI’s publisher FAQ says any public website can appear in ChatGPT search.

2. OAI-SearchBot controls inclusion in summaries and snippets.
OpenAI says publishers who want content included in ChatGPT search summaries and snippets should not block OAI-SearchBot.

3. GPTBot and OAI-SearchBot are separate.
OpenAI’s crawler documentation says each setting is independent and serves different product functions.

4. noindex is the correct exclusion control.
OpenAI says if you want to stop a page from appearing in search responses entirely, use noindex. As with Google, the crawler must be able to access the page to read that directive.

5. Google says core SEO best practices still apply in AI features.
Google’s AI-features documentation says no special requirements are needed beyond the same strong fundamentals: helpful content, textual clarity, discoverability, good page experience, and proper access.
Green working rules
Allow OAI-SearchBot if you want ChatGPT search visibility.

Use noindex if you truly want to opt out.

Keep important content public, textual, and easy to cite.

Do not split AI-search work from core SEO fundamentals; layer it on top.
Yellow — field evidence deep dive
Finding 1 — AI visibility is not the same thing as Google visibility
Recent Search Engine Land reporting shows that brands or locations performing well in Google do not automatically appear in ChatGPT, Gemini, or Perplexity recommendations. That suggests AI visibility has overlap with SEO, but also its own selection patterns.
Finding 2 — AI bot activity is growing fast, but referral traffic remains relatively small
Search Engine Land’s reporting on TollBit data says AI bot traffic has risen sharply while referral traffic from AI systems still trails Google by a large margin. That means AI discovery matters strategically, but Google remains the much larger traffic driver for most sites today.
Finding 3 — AI visibility rewards detailed, answerable, machine-readable information
Field guidance on GEO/AI search increasingly emphasises clear entity information, clean site access, question-led content, and structured meaning. That is less a contradiction to SEO than an amplification of the parts SEO often under-prioritises.
Yellow practical rules
1. Don’t block OAI-SearchBot by accident.
2. Keep key information explicit and textual.
3. Expect AI visibility and AI traffic to be different things.
4. Build for citation clarity, not just blue-link clicks.
Blue synthesis / contradiction note
The main tension here is not between green and yellow; it is between expectation and reality. Officially, any public site can appear in ChatGPT search if access is allowed. Field evidence says that actual recommendation or citation is much more selective than simple eligibility, and the referral payoff is often smaller than people imagine. So AI visibility deserves attention, but not at the expense of core Google SEO.
Blue decision rule
Treat LLM discovery as a second scoring lens, not a separate universe. Do the low-risk access and clarity work now because it overlaps heavily with strong SEO, but keep expectations realistic about traffic and don’t abandon Google-first fundamentals.

Future ranking table

Blank for now
Classic SEO ranking
We will fill this after reviewing the evidence section by section.
LLM / AI visibility ranking
We will keep this separate from classic SEO and score it independently.