or Claude to “search the net,” it isn’t simply answering from its coaching knowledge. It’s calling a separate search system.
Most individuals know that half.
What’s much less clear is how a lot conventional serps matter and the way a lot has been constructed on prime of them.
All of it isn’t totally public, so I’m performing some psychological deduction right here. However we will use totally different hints from taking a look at bigger programs to construct a helpful psychological mannequin.
We’ll undergo question optimization, how serps are used for discovery, chunking content material, “on-the-fly” retrieval, and the way you might probably reverse-engineer a system like this to construct a “GEO [] scoring system.”
In case you’re conversant in RAG, a few of this might be repetition, however it could actually nonetheless be helpful to see how bigger programs cut up the pipeline right into a discovery part and a retrieval part.
In case you’re brief on time, you may learn the TL;DR.
TL;DR
Net search in these AI chatbots is probably going a two-part course of. The primary half leans on conventional serps to search out and rank candidate docs. Within the second half, they fetch the content material from these URLs and pull out probably the most related passages utilizing passage-level retrieval.
The large change (from conventional search engine optimization) is question rewriting and passage-level chunking, which let lower-ranked pages outrank greater ones if their particular paragraphs match the query higher.
The technical course of
The businesses behind Claude and ChatGPT aren’t totally clear about how their net search programs work inside the UI chat, however we will infer rather a lot by piecing issues collectively.
We all know they lean on serps to search out candidates, at this scale, it might be absurd to not. We additionally know that what the LLM truly sees are items of textual content (chunks or passages) when grounding their reply.
This strongly hints at some form of embedding-based retrieval over these chunks somewhat than over full pages.
This course of has a number of components, so we’ll undergo it step-by-step.
Question re-writing & fan-out
First, we’ll take a look at how the system cleans up human queries and expands them. We’ll cowl the rewrite step, the fan-out step, and why this issues for each engineering and search engine optimization.
I believe this half is perhaps probably the most clear, and the one most individuals appear to agree on on-line.
The question optimization step is about taking a human question and turning it into one thing extra exact. For instance, “please seek for these pink sneakers we talked about earlier” turns into “brown-red Nike sneakers.”
The fan-out step, alternatively, is about producing extra rewrites. So if a person asks about mountaineering routes close to me, the system would possibly attempt issues like “newbie hikes close to Stockholm,” “day hikes close to Stockholm public transport,” or “family-friendly trails close to Stockholm.”
That is totally different from simply utilizing synonyms, which conventional serps are already optimized for.
If that is the primary time you’re listening to about it and also you’re unconvinced, check out Google’s personal docs on AI query fan-out or do a little bit of digging round question rewriting.
To what extent this truly works, we will’t know. They might not fan it out that a lot and simply work with a single question, then ship extra ones down the pipeline if the outcomes are lackluster.
What we can say is that it’s in all probability not a giant mannequin doing this half. In case you take a look at the analysis, Ye et al. explicitly use an LLM to generate sturdy rewrites, then distill that right into a smaller rewriter to keep away from latency and price overhead.
As for what this a part of the pipeline means, for engineering, it simply means you need to clear up the messy human queries and switch them into one thing that has the next hit fee.
For the enterprise and search engine optimization folks on the market, it means these human queries you’ve been optimizing for are getting remodeled into extra robotic, document-shaped ones.
search engine optimization, as I perceive it, used to care rather a lot about matching the precise long-tail phrase in titles and headings. If somebody looked for “greatest trainers for dangerous knees,” you’d keep on with that precise string.
What it’s essential to care about now can be entities, attributes, and relationships.
So, if a person asks for “one thing for dry pores and skin,” the rewrites would possibly embody issues like “moisturizer,” “occlusive,” “humectant,” “ceramides,” “fragrance-free,” “keep away from alcohols” and never simply “how would I discover a good product for dry pores and skin.”
However let’s be clear so there’s no confusion: we will’t see the inner rewrites themselves, so these are simply examples.
In case you’re on this half, you may dig deeper. I guess there are many papers on the market on how to do that effectively.
Let’s transfer on to what these optimized queries are literally used for.
Utilizing serps (for doc stage discovery)
It’s fairly frequent data by now that, to get up-to-date solutions, most AI bots depend on conventional serps. That’s not the entire story, but it surely does reduce the net all the way down to one thing smaller to work with.

I’m assuming the total net is simply too massive, too noisy, and too fast-changing for an LLM pipeline to drag uncooked content material immediately. So through the use of already established serps, you get a option to slender the universe.
In case you take a look at bigger RAG pipelines that work with tens of millions of paperwork, they do one thing comparable. I.e. utilizing a filter of some type to determine which paperwork are essential and value additional processing.
For this half, we do have proof.
Each OpenAI and Anthropic have stated they use third-party serps like Bing and Courageous, alongside their very own crawlers.
Perplexity might have constructed out this half on their very own by now, however to start with, they might have achieved the identical.
We even have to think about that conventional serps like Google and Bing have already solved the toughest issues. They’re a longtime expertise that handles issues like language detection, authority scoring, area belief, spam filtering, recency, geo-biasing, personalization, and so forth.
Throwing all of that away to embed your complete net your self appears unlikely. So I’m guessing they lean on these programs as a substitute of rebuilding them.
Nevertheless, we don’t know what number of outcomes they really fetch per question, whether or not it’s simply the highest 20 or 30. One unofficial article in contrast citations from ChatGPT and Bing, seemed on the rating order, and located that some got here from as far down as twenty second place. If true, this means it’s essential to intention for top-20-ish visibility.
Moreover, we additionally don’t know what different metrics they use to determine what surfaces from there. This article argues that AI engines closely favor earned media somewhat than official websites or socials, so there’s extra happening.
Nonetheless, the search engine’s job (whether or not it’s totally third-party or a combination) is discovery. It ranks the URL primarily based on authority and key phrases. It would embody a snippet of data, however that alone gained’t be sufficient to reply the query.
If the mannequin relied solely on the snippet, plus the title and URL, it might possible hallucinate the small print. That’s not sufficient context.
So this pushes us towards a two-stage structure, the place a retrieval step is baked in — which we’ll get to quickly.
What does this imply when it comes to search engine optimization?
It means you continue to must rank excessive in conventional serps to be included in that preliminary batch of paperwork that will get processed. So, sure, basic search engine optimization nonetheless issues.
However it might additionally imply it’s essential to take into consideration potential new metrics they is perhaps utilizing to rank these outcomes.
This stage is all about narrowing the universe to a couple pages value digging into, utilizing established search tech plus inside knobs. Every part else (the “it returns passages of data” half) comes after this step, utilizing normal retrieval methods.
Crawl, chunk & retrieve
Now let’s transfer on to what occurs when the system has recognized a handful of attention-grabbing URLs.
As soon as a small set of URLs passes the primary filter, the pipeline is pretty easy: crawl the web page, break it into items, embed these items, retrieve those that match the question, after which re-rank them. That is what’s referred to as retrieval.

I name it on-the-fly right here as a result of the system solely embeds chunks as soon as a URL turns into a candidate, then it caches these embeddings for reuse. This half is perhaps new for those who’re already conversant in retrieval.
To crawl the web page, they use their very own crawlers. For OpenAI, that is OAI-SearchBot, which fetches the uncooked HTML so it may be processed. Crawlers don’t execute JavaScript. They depend on server-rendered HTML, so the identical search engine optimization guidelines apply: content material must be accessible.
As soon as the HTML is fetched, the content material needs to be became one thing searchable.
In case you’re new to this, it would really feel just like the AI “scans the doc,” however that’s not what occurs. Scanning total pages per question could be too gradual and too costly.
As an alternative, pages are cut up into passages, often guided by HTML construction: headings, paragraphs, lists, part breaks, that form of factor. These are referred to as chunks within the context of retrieval.
Every chunk turns into a small, self-contained unit. Token-wise, you may see from Perplexity UI citations that chunks are on the order of tens of tokens, perhaps round 150, not 1,000. That’s about 110–120 phrases.
After chunking, these items are embedded utilizing each sparse and dense vectors. This allows the system to run hybrid search and match a question each semantically and by key phrase.
In case you’re new to semantic search, in brief, it means the system searches for that means as a substitute of tangible phrases. So a question like “signs of iron deficiency” and “indicators your physique is low on iron” would nonetheless land close to one another in embedding house. You possibly can learn extra on embeddings right here for those who’re eager to be taught the way it works.
As soon as a preferred web page has been chunked and embedded, these embeddings are in all probability cached. Nobody is re-embedding the identical StackOverflow reply 1000’s of instances a day.
That is clearly why the system feels so quick, in all probability the new 95–98% of the net that really will get cited is already embedded, and cached aggressively.
We don’t know to what extent although and the way a lot they pre-embed to ensure the system runs quick for standard queries.
Now the system wants to determine which chunks matter. It makes use of the embeddings for every chunk of textual content to compute a rating for each semantic and key phrase matching.
It picks the chunks with the very best scores. This may be something from 10 to 50 top-scoring chunks.
From right here, most mature programs will use a re-ranker (cross-encoder) to course of these prime chunks once more, doing one other spherical of rating. That is the “repair the retrieval mess” stage, as a result of sadly retrieval isn’t at all times utterly dependable for lots of causes.
Though they are saying nothing about utilizing a cross-encoder, Perplexity is likely one of the few that paperwork their retrieval course of overtly.
Their Search API says they “divide paperwork into fine-grained items” and rating these items individually to allow them to return the “most related snippets already ranked.”
What does this all imply for search engine optimization? If the system is doing retrieval like this, your web page isn’t handled as one huge blob.
It’s damaged into items (typically paragraph or heading stage) and people items are what get scored. The total web page issues throughout discovery, however as soon as retrieval begins, it’s the chunks that matter.
Which means every chunk must reply the person’s query.
It additionally implies that in case your essential data isn’t contained inside a single chunk, the system can lose context. Retrieval isn’t magic. The mannequin by no means sees your full web page.
So now we’ve lined the retrieval stage: the place the system crawls pages, chops them into items, embeds these items, after which makes use of hybrid retrieval and re-ranking to drag out solely the passages that may reply the person’s query.
Doing one other move & handing over chunks to LLM
Now let’s transfer on to what occurs after the retrieval half, together with the “persevering with to go looking” characteristic and handing the chunks to the principle LLM.

As soon as the system has recognized a number of high-ranking chunks, it has to determine whether or not they’re adequate or if it must hold looking out. This choice is sort of actually made by a small controller mannequin, not the principle LLM.
I’m guessing right here, but when the fabric seems to be skinny or off-topic, it might run one other spherical of retrieval. If it seems to be stable, it could actually hand these chunks over to the LLM.
Sooner or later, that handoff occurs. The chosen passages, together with some metadata, are handed to the principle LLM.
The mannequin reads all of the supplied chunks and picks whichever one greatest helps the reply it desires to generate.
It doesn’t mechanically comply with the retriever’s order. So there’s no assure the LLM will use the “prime” chunk. It could choose a lower-ranked passage just because it’s clearer, extra self-contained, or nearer to the phrasing wanted for the reply.
So similar to us, it decides what to absorb and what to disregard. And even when your chunk scores the very best, there’s no assurance will probably be the primary one talked about.
What to assume about
This method isn’t actually a black field. It’s a system folks have constructed at hand the LLMs the suitable data to reply a person’s query.
It finds candidates, splits paperwork into items, searches and ranks these items, after which fingers them over to an LLM to summarize. So once we perceive how the system works, we will additionally work out what we’d like to consider when creating content material for it.
Conventional search engine optimization nonetheless issues rather a lot, as a result of this method leans on the previous one. Issues like having a correct sitemap, simply rendered content material, correct headings, area authority, and correct last-modified tags are all essential in your content material to be sorted appropriately.
As I identified, they could be mixing serps with their very own expertise to determine which URLs get picked, which is value holding in thoughts.
However I believe paragraph stage relevance is the brand new leverage level.
Perhaps this implies answer-in-one-chunk design will rule. (Simply don’t do it in a method that feels bizarre, perhaps a TL;DR.) And keep in mind to make use of the suitable vocabulary: entities, attributes, relationships, like we talked about within the question optimization part.
The right way to construct a “GEO Scoring System” (for enjoyable)
To determine how effectively your content material will do, we’ll need to simulate the hostile setting your content material will stay in. So let’s attempt to reverse engineer this pipeline.
Be aware, that is non-trivial, as we don’t know the inner metrics they use, so consider this as a blueprint.
The concept is to create a pipeline that may do question rewrite, discovery, retrieval, re-ranking and an LLM choose, after which see the place you find yourself in comparison with your rivals for various subjects.

You start with a number of subjects like “hybrid retrieval for enterprise RAG” or “LLM analysis with LLM-as-judge,” after which construct a system that generates pure queries round them.
Then you definitely move these queries by way of an LLM rewrite step, as a result of these programs typically reformulate the person question earlier than retrieval. These rewritten queries are what you truly push by way of the pipeline.
The primary examine is visibility. For every question, take a look at the highest 20–30 outcomes throughout Courageous, Google and Bing. Be aware whether or not your web page seems and the place it sits relative to rivals.
On the identical time, gather domain-level authority metrics (Moz DA, Ahrefs DR, and many others.) so you may fold that in later, since these programs in all probability nonetheless lean closely on these indicators.
In case your web page seems in these first outcomes, you progress on to the retrieval half.
Fetch your web page and the competing pages, clear the HTML, cut up them into chunks, embed these chunks, and construct a small hybrid retrieval setup that mixes semantic and key phrase matching. Add a re-ranking step.
Someplace right here you additionally inject the authority sign, as a result of higher-authority domains realistically get scored greater (regardless that we don’t know precisely how a lot).
Upon getting the highest chunks, you add the ultimate layer: an LLM-as-a-judge. Being within the prime 5 doesn’t assure quotation, so that you simulate the final step by handing the LLM a number of of the top-scored chunks (with some metadata) and see which one it cites first.
Once you run this in your pages and rivals, you see the place you win or lose: the search layer, the retrieval layer or the LLM layer.
Keep in mind that is nonetheless a tough sketch, but it surely offers you one thing to start out with if you wish to construct an analogous system.
This text targeted on the mechanics somewhat than the technique facet of search engine optimization/GEO, which I get gained’t be for everybody.
The purpose was to map the move from a person question to the ultimate reply and present that the AI search device isn’t some opaque drive.
Even when components of the system aren’t public, we will nonetheless infer an inexpensive sketch of what’s occurring. What’s clear thus far is that the AI net search doesn’t change conventional serps. It simply layers retrieval on prime of them.
Earlier than ending this, it’s value mentioning that the deeper analysis characteristic is totally different from the built-in search instruments, that are pretty restricted and low cost. Deep analysis possible leans on extra agentic search, which can be “scanning” the pages to a better extent.
This would possibly clarify why my very own content material from my web site reveals up in deep analysis regardless that it’s not optimized for the fundamental search layer, so it virtually by no means reveals up in primary AI search.
There’s nonetheless extra to determine earlier than saying what truly issues in observe. Right here I’ve principally gone by way of the technical pipeline but when that is new stuff I hoped it clarify it effectively.
Hopefully it was straightforward to learn. In case you loved it, be happy to share it or join with me on LinkedIn, Medium or by way of my site.
❤

