Entity Clarity
The disambiguation work that lets an AI search engine identify a brand as the same entity across mentions. The strategic layer schema for AI implements.
Read the entry →Stan Consulting · Marketing Atlas · Reference · AI Search
The schema.org JSON-LD markup decisions that improve AI search engine understanding of a page's entity, content, and authority. The same schema as SEO, applied with different priorities.
Section 02 · Quick definition
Schema for AI is the application of schema.org JSON-LD markup with priorities tuned for AI search retrieval. The vocabulary is the same vocabulary SEO has used for a decade. The priorities are different: Article with author, Organization with @id and sameAs, Person with stable bio URL, DefinedTerm over generic Thing, and SpeakableSpecification on definitional content. The goal is not a rich result in Google. The goal is a confident citation in an AI answer. The same JSON-LD block can be SEO-correct and AI-inert if the priorities are wrong.
Section 03 · Why it matters
Schema is the structural layer an AI retrieval agent uses to answer the question, what is this page, and can I cite it. A page with strong schema for AI gives the agent a clean answer in milliseconds: this is an Article, written by this Person, published by this Organization, about this DefinedTerm, with these speakable selectors. A page with weak or generic schema makes the agent infer everything from raw HTML, which is slower and less confident. The lower the agent's confidence, the lower the citation probability.
The metric matters because most operator sites finished schema work before AI search retrieval mattered. The Article tag is there. The Organization tag is there. The pieces that disambiguate entity, anchor authorship, and surface speakable content are usually missing. The fix is not new schema. The fix is reordering priorities and filling gaps.
The practical stake is that schema for AI is the technical work with the best dollar-for-dollar return on AI citation rate. The site already has the markup framework. The question is whether it carries the signals the retrieval layer actually needs.
Section 04 · How it works
An AI retrieval agent fetching a page parses the JSON-LD blocks in the head and body, builds an entity graph from the @type, @id, and sameAs values, and uses that graph to answer two questions about the page: which entity is this about, and how authoritative is this source. Strong schema produces clean answers to both. Weak schema forces the agent to fall back on raw text and document structure.
The agent reads every @type and @id in the JSON-LD. Organization @id, Person @id, Article @id, and any cross-referenced entities form a small graph. Pages with a complete graph (Organization with founder Person, Article with author Person, both Persons resolving to bio pages) score higher than pages with isolated nodes.
The agent prefers specific @type values over generic ones. DefinedTerm is more useful than Thing. SoftwareApplication is more useful than CreativeWork. ProfessionalService is more useful than LocalBusiness alone. The more specific the type, the more the agent learns from the markup without having to re-derive it from the page text.
The agent looks for SpeakableSpecification cssSelector values that mark which page elements are definitional. Pages that mark their H1 and definition card as speakable tell the agent what to quote when the answer needs a one-line description. Pages that omit speakable force the agent to choose what to quote, and the choice is often suboptimal.
The agent checks sameAs links from Organization and Person schema. Cross-references to Wikidata, LinkedIn, Crunchbase, Bloomberg, and other authoritative directories increase confidence. A Person @id with three sameAs links resolves more confidently than one with none. The verification compounds across the graph: a confident Person makes the Article more citable.
The four steps run on every page the agent retrieves. Improvements compound: a site that adds Person @id with sameAs across all author bios sees citation rates improve on subsequent retrievals across every page that names those authors.
Section 05 · Common misunderstandings
“Our schema validates, so it's done.”
Validation confirms the markup is parseable. It does not confirm the markup carries the signals AI retrieval needs. A page with valid Article, BreadcrumbList, and FAQ schema and no Person @id, no Organization sameAs, and no SpeakableSpecification validates cleanly and produces nothing useful for citation. Validation is the floor. Priority is the work.
“FAQ schema is the most important markup.”
FAQ schema was the highest-priority markup for SEO rich results in 2019. For AI search retrieval in 2026, Article with author and Organization with @id outrank it. FAQ markup still helps for specific question-answer surfaces, but the citation work has moved upstream. Operators still prioritizing FAQ over entity schema are working against the wrong scoring function.
“More schema is better schema.”
The opposite. AI retrieval prefers a focused, internally-consistent graph over a sprawling one. A page with five overlapping schema types and conflicting @id values scores lower than a page with three clean types and consistent @id. The work is to remove redundancy and tighten cross-references, not to add types.
“Speakable schema is for voice search and we don't care about Alexa.”
SpeakableSpecification was originally aimed at voice surfaces, but AI search engines now use the cssSelector values to choose what to quote in answer cards. A page that marks its H1 and definition body as speakable gives the model an explicit cue. The voice-surface origin is irrelevant; the AI-citation use is real.
“Generic schema types like Thing or CreativeWork are safe defaults.”
Safe in that they validate. Useless in that they communicate nothing the retrieval layer can use. AI retrieval prefers DefinedTerm to Thing, ProfessionalService to LocalBusiness, SoftwareApplication to CreativeWork. The specific types tell the model what kind of answer this page can support. The generic types tell it nothing.
Section 06 · Diagnostic questions
Does every key page carry Article, Person (author), and Organization (publisher) schema with @id values that resolve to a single canonical fragment URL across the site?
Do Organization and Person schema include sameAs cross-references to Wikidata, LinkedIn, Crunchbase, and other authoritative directories?
Are specific @type values used (DefinedTerm, ProfessionalService, SoftwareApplication) instead of generic ones (Thing, CreativeWork, LocalBusiness)?
Does definitional content carry SpeakableSpecification with cssSelector values that point at the H1 and the canonical definition body?
Does the @graph structure include all relevant entities in one block per page, or are there scattered separate JSON-LD blocks that fail to cross-reference?
Is BreadcrumbList present and accurate, with itemListElement positions that match the live navigation hierarchy?
Are Article entries published with datePublished and dateModified, and do those dates reflect actual editorial activity rather than automated stamps?
Section 07 · Related Atlas entries
Section 08 · Five Cents
Most pre-2024 schema implementations were technically correct and AI-inert. The markup signaled what to crawl, not what to trust. I have audited sites with valid Article, BreadcrumbList, FAQ, and Review schema across every page and zero Person @id, zero Organization sameAs, zero SpeakableSpecification. Validation passed. Citation never moved. The fix is not more markup. The fix is the thirty minutes spent adding @id cross-references to the existing Person and Organization blocks and the additional ten minutes adding sameAs to Wikidata. That hour shifts the page from technically present to citably present, which is the difference the retrieval layer actually scores on.
Stan · Marketing AtlasSection 09 · Sources