RESEARCH

The Mathematics of LLM Optimization: How AI Readiness Is Scored

Most "AI readiness" tools are glorified checklists. Here's the quantitative methodology that separates a rigorous score from a random number generator.
By Faneros AI · March 2026 · 7 min read

Most "AI readiness" tools are glorified checklists. They scan your site, flag a few issues, and hand you a number. But how is that number actually calculated? What separates a rigorous scoring methodology from a random number generator with a progress bar?

At Faneros, we built a quantitative scoring engine grounded in weighted composite analysis — the same mathematical framework used in credit scoring, portfolio risk assessment, and medical diagnostic models.

The Core Formula

Every AI readiness score is fundamentally a weighted linear composite:

SGEO = Σi=1n λi · φ(ξi)    where Σλi = 1
φi) = min(max(τi−1 · Ri , 0), 100)    ∀ i ∈ {1,...,n}

Where SGEO is the composite score, λi is the weight assigned to variable i, φ is a bounded normalization function, ξi is the raw input signal, τi is the threshold parameter, and Ri is the measured response value.

The constraint Σλi = 1 ensures the composite score remains on a 0–100 scale. The normalization function φ clamps each variable's contribution to prevent any single catastrophic failure from producing a negative score.

Why Weighted Composites?

Not all variables matter equally. If an AI crawler cannot physically reach your website, it doesn't matter how beautiful your schema markup is.

A site that blocks all 8 major AI crawlers scores differently than a site with perfect crawler access but missing structured data. The weighting function reflects this hierarchy of dependencies.

The mathematical property we optimize for is monotonicity with diminishing returns: fixing the highest-weighted failure always produces the largest score improvement. This means the audit naturally prioritizes the most impactful changes.

The Variable Space

Our model evaluates these signal categories:

Access Layer

robots.txt Configuration — Whether the file explicitly allows or blocks AI crawlers. llms.txt Presence — The emerging standard for AI communication. XML Sitemap Accessibility — Whether a well-formed sitemap exists and is discoverable.

Comprehension Layer

JSON-LD Schema Depth — Not just presence but coverage of GEO-specific fields. Content Structure & Hierarchy — Proper heading tags, meta descriptions, semantic HTML. FAQ Schema Depth — AI platforms frequently cite FAQ structured data verbatim.

Trust Layer

SSL Certificate Validity — Baseline trust signal. Security Header Coverage — HSTS, CSP, X-Frame-Options. Page Speed (TTFB) — AI crawlers have time budgets under 200ms.

Dependency Chains

Some factors are prerequisites for others. If AI crawlers are blocked, schema quality is irrelevant — the crawler never sees it. Our model captures these through conditional weighting: downstream factors receive reduced weight when their prerequisite fails.

This prevents a common failure mode where a site scores 70% by having great schema, llms.txt, and content — while being completely invisible because crawlers are blocked. In our model, that site correctly scores much lower.

Grade Mapping

G = { A : S ≥ 90,   B : S ≥ 80,   C : S ≥ 70,   D : S ≥ 60,   F : S < 60 }

The practical implication: when Faneros tells you to fix your robots.txt before worrying about schema markup, it's not arbitrary — it's the mathematically optimal sequence for improving your visibility.

See Your AI Visibility Score

Faneros scans 7 AI platforms in 60 seconds. Find out if ChatGPT, Claude, and Perplexity can see your business.

Scan My Site →