All articles
6 min read

40% More Leads Through LLM Optimization (AEO) in GitBook

A practical framework that transforms GitBook into an AEO/GEO channel. Content structure, JSON-LD, sitemap, edge refreshing on Vercel, public RAG, and attribution to CRM. Goal: +40% B2B leads.

Cover illustration for article: 40% More Leads Through LLM Optimization (AEO) in GitBook

Key takeaways

  • AEO is about optimizing for LLM responses, not just 'SEO for docs'.
  • GitBook + Vercel (edge, sub-second revalidations) accelerates indexing by LLM.
  • JSON-LD (TechArticle/HowTo/FAQ), clean sitemaps, and stable anchors enhance citation relevance.
  • Public RAG endpoint + Vercel AI Gateway (e.g., Grok 4.3) strengthens response sourcing.
  • CRM attribution: AEO UTM parameters, user question logging, self-report fields, and intent mapping.
  • A 6-week experimental plan delivers +40% MQL without product changes.

LLMs are emerging as a viable B2B acquisition channel: in Perplexity/ChatGPT, users ask 'how to do X?', and the responses link to documentation. Analytics (UTM=aeo, self-report) show transitions from these answers to demos/trials. If your documentation and changelogs in GitBook are not 'digestible', you're losing leads. Below is a framework to achieve AEO/GEO on GitBook + Vercel in 6 weeks and deliver ~40% more qualified leads.

AEO on GitBook: A Channel for Mature Queries, Not Just 'Visibility'

Answer Engine Optimization (AEO) involves structuring content so that LLMs can accurately cite your responses. It focuses on transactional intents ('how to integrate X with Y?') and tool recommendations, rather than general exposure.

Why now? GitBook hosted on Vercel supports thousands of services with sub-second updates. Perplexity/ChatGPT are growing as sources for tool recommendations. Vercel provides secure connections to Vercel Postgres via firewall/IP allowlist and convenient content revalidation mechanisms. For teams using GitBook, this means easier and safer deployment of their own /api/rag and storing pgvector without exposing the database to the public. The AI Gateway includes Grok 4.3 (context up to 1M tokens). These are excellent conditions for connecting a public RAG and quick refreshing.

The business goal is straightforward: more citations in LLM responses to buyer questions and a shorter path to demos. We measure success by the increase in MQL and conversion to pipeline, not just website traffic.

Content Structure and Signals for LLM (schema, JSON-LD, sitemap)

LLMs require clear units of knowledge. Create one topic per page, consistent H1/H2, short paragraphs, stable anchors, and canonical URLs. Write changelogs in 'job-to-be-done' language (what users can now do faster).

Add JSON-LD. Use TechArticle for conceptual pages, HowTo for instructions, FAQPage for questions, and SoftwareApplication for product pages (including name, integrations, and trial link). These types clearly communicate to crawlers and LLMs the intent they are dealing with: TechArticle signals expert content, HowTo outlines execution steps, FAQPage presents question-answer pairs, and SoftwareApplication defines the product entity and relationships. Each page should have a modification date and API version in the metadata.

Ensure you have maps: sitemap index (docs + changelog), update frequency, priorities. In robots.txt, include GPTBot, PerplexityBot, and Google-Extended. Avoid walls and paywalls on docs. These are signals that you are a credible source for LLMs.

Structured content + correct schemas and sitemaps is the most cost-effective way to ensure your responses are frequently included in LLM recommendations. Do your key pages already have JSON-LD and stable anchors?

  • One intent per page; names consistent with product and API.
  • 'Quick Start' section (5–7 steps) for each integration.
  • FAQ with 5–10 questions at the bottom of the page (FAQPage).
  • Changelog in Added/Changed/Fixed/Deprecated format + link to docs.
  • 'See also' links: 3–5 related topics (no carousels).

Fast Refreshing (Edge) and Public RAG on Vercel AI Gateway

Edge is a CDN layer executing logic close to the user and crawler. On Vercel, enable revalidation after a webhook from GitBook: publication → webhook → tag revalidate → refreshed HTML + sitemap in sub-seconds. Use ETag and s-maxage to ensure LLMs see fresh changes without full rebuilds.

A public RAG (Retrieval-Augmented Generation) is an endpoint that returns snippets of your docs as context for the model. Expose /api/rag?q=… with top-k=8, version filtering, and 'source' with canonical URL. Vectorize paragraphs and code; store embeddings in Postgres (pgvector). Vercel Postgres with IP allowlist/firewall simplifies operations.

In the generation layer, leverage Vercel AI Gateway. You get routing, caching, and observability 'out of the box'. Grok 4.3 with 1M tokens allows for long contexts of changelogs and entire integrations. A prompt is the command passed to the model; an agent is an autonomous AI component that performs steps based on a goal—it can call your RAG and return links to docs.

Add guardrails: request limiter, PII filter, domain whitelist in CORS, and logging of requests (without sensitive data) for analytics. In return, you gain stable citations and insights into real market questions.

  • Webhook from GitBook → revalidateTag in Vercel (ISR).
  • Sitemaps refreshed with the same webhook.
  • /api/rag returns: title, url, score, snippet, version.
  • Embeddings: OpenAI/Grok embeddings; storage: Postgres + pgvector.
  • AI Gateway: cache, fallback model, cost and accuracy metrics.

Attribution and a Plan for +40% Leads in 6 Weeks

AEO must integrate with CRM. Add UTM 'utm_channel=aeo' to CTAs in docs and changelogs—this way, clicks from documentation (often after recommendations from LLMs that do not pass the referrer) reach CRM with the correct channel instead of 'Direct'. This allows you to calculate MQL/pipeline from AEO and compare it with other sources. The lead form should capture: last page, anchor, query to /api/rag, and response ID. Add a self-report field 'Which AI did you use?' (Perplexity/ChatGPT/Others). Map the intent (e.g., 'Integration with Slack ETL') to the account and use case.

The experimental plan without product overhaul: week 0 – baseline MQL and indexing; week 1 – refactor top 10 pages + JSON-LD + FAQ; week 2 – dedicated changelog and sitemap-changelog; week 3 – launch /api/rag and AI Gateway; week 4 – CTA with UTM and fields in the form; week 5 – prompt tuning, filters; week 6 – impact assessment and rollout.

Measure: the number of citations in LLM responses (from RAG logs and self-reports), time from publication to the first LLM query, MQL from the AEO channel, conversion from AEO→demo→pipeline, and cost per MQL. If numbers do not increase after 3 weeks, check JSON-LD and title consistency.

  • Form: hidden fields question, answer_id, anchor, utm_channel=aeo.
  • Report: MQL from AEO vs other channels; cost/MQL.
  • Alert: no citations from RAG for 48h after release → check sitemap/robots.

AEO on GitBook is a quick path to purchase intents present in LLM responses. Combine structured content, fast refreshing on Vercel, public RAG, and CRM attribution, and you will build a repeatable channel with +40% MQL. Want to discuss implementation for your stack? Schedule a brief consultation—no sales pressure, with a concrete 6-week plan.

Frequently asked questions

Does AEO replace traditional SEO?

No. AEO complements SEO but operates on a different surface with different metrics. SEO measures: rankings, impressions, CTR, non-brand sessions, SERP share. AEO measures: share-of-answer (share in LLM responses), number of citations and clicks on source links, time to first citation after publication, MQL/SQL/pipe from CTAs in docs, and cost per MQL. The format and signals also differ: task-based docs/HowTo/FAQ and freshness vs domain authority and links. The channels reinforce each other, but you report them separately.

What types of JSON-LD are most important for documentation?

TechArticle for technical articles, HowTo for step-by-step instructions, FAQPage for Q&A sections, and SoftwareApplication for product and integration pages.

Will a public RAG not reveal too much knowledge to competitors?

RAG should only return snippets of content that are already public in your docs—there is no access to internal data or customer information. The competitive risk mainly involves easier access to already public snippets. Mitigate this through rate-limiting, PII/secrets redaction before indexing, version/range filtering, and request logging. The gains (citations, attribution, MQL) usually outweigh this risk.

How many resources does an SME need to implement this framework?

A team of 2–3 people for 4–6 weeks: tech writer/PM, developer (Vercel/DB), ops/analytics. You can start faster by focusing on the top 10 pages and one RAG endpoint.

How to measure the impact of AEO when the referrer is 'Direct'?

Take a straightforward approach: 1) add UTM to all CTAs in docs/changelog (e.g., utm_channel=aeo)—when a user comes from LLM and clicks the CTA, the 'AEO' channel will enter CRM even without a referrer; 2) self-report field 'Which AI did you use?' (Perplexity/ChatGPT); 3) hidden fields: last page/anchor and last question to /api/rag. In CRM, set a rule: if utm_channel=aeo or self-report indicates AI, assign the lead to AEO.

Let's talk
about your project

The consultation is free and no-strings-attached. We'll review your needs and I'll suggest concrete solutions.

Send a message

Briefly describe your problem — I'll get back to you with concrete suggestions.