TECHNICAL

How to Set Up robots.txt, llms.txt, and JSON-LD Schema for Better AI Visibility

How to set up robots.txt, llms.txt, and JSON-LD schema so AI platforms can better understand and recommend your business.
By Faneros AI · April 2026 · 5 min read

Many companies assume that if their website is live and indexed, AI systems will understand it. That is a risky assumption. AI platforms do not just look for pages. They look for clear signals about who you are, what you offer, where you operate, and whether your site can be accessed and interpreted cleanly. When those signals are weak, your business may be omitted from recommendations even if your SEO looks fine.

Three of the most discussed technical elements in this area are robots.txt, llms.txt, and JSON-LD schema. They do different jobs. Together, they can improve how machines interpret your business.

What robots.txt does

robots.txt is a plain text file placed at the root of a website. It gives crawl instructions to automated agents. It does not guarantee behavior, but it signals which parts of a site may or may not be crawled. If your robots.txt file blocks key pages, directories, scripts, or resources, search engines and AI-related crawlers may have trouble understanding your site.

This is one of the first files to check when a company says, “We invested in SEO, but AI platforms still do not seem to mention us.” An overly restrictive robots.txt file can block access to the very pages that explain your services, location, expertise, and trust signals.

Regardless of your platform, you can usually view the file by going to yourdomain.com/robots.txt. If you need to edit it, the universal method is to update the file in your site settings if your CMS supports that, or upload a revised robots.txt file to the root directory of your hosting environment.

What llms.txt is meant to do

llms.txt is a newer concept. It is meant to give large language models a clearer map of useful information about your site. Think of it as a machine-readable guide that points AI systems toward the pages and content blocks that best explain your business.

Adoption is still evolving, so teams should be realistic. llms.txt is not a silver bullet. But it can be a helpful organizational signal, especially when paired with strong site structure, accessible content, and proper schema.

Regardless of your platform, the universal method is the same: create a plain text llms.txt file and upload it to the root directory so it appears at yourdomain.com/llms.txt. Keep it clean, accurate, and aligned with the pages you most want machines to understand.

Why JSON-LD schema matters

JSON-LD schema is structured data that helps machines interpret the meaning of your content. It can describe your organization, product, service, address, FAQs, reviews, and more. For a software company, structured data can clarify that you are an Organization and that your offering is a SoftwareApplication or Product.

This matters because AI systems need more than words on a page. They need context. If your site says “we help businesses improve visibility,” that is broad. If your structured data clearly identifies your company name, website, address, contact details, software category, and service relationships, machines have a cleaner framework for understanding and citing you.

Regardless of your platform, JSON-LD can usually be added in one of two universal ways: paste the script into the page or site-wide head area, or use your CMS’s custom code field or tag manager if one is available.

What to check first if AI platforms are not recommending your business

Start with accessibility. Are your key pages crawlable? Are important resources blocked? Is your robots.txt file clean and intentional?

Then check clarity. Do your pages plainly explain what your company does, who it serves, and what makes it different? AI systems struggle when websites are vague, overly styled, or missing direct statements.

Then check structure. Do you have valid JSON-LD schema for your organization and offerings? Do your FAQ pages use clear question-and-answer formatting? Are your contact details consistent across the site?

Then check machine guidance. If you are using llms.txt, does it point to the right pages? Is it current?

A simple setup approach that works on any site

If you want a practical sequence, use this:

  1. Review robots.txt for accidental blocks
  2. Create or update llms.txt at the root directory
  3. Add Organization and SoftwareApplication JSON-LD schema
  4. Make sure your service and product pages are explicit and easy to parse
  5. Add FAQ content where buyers ask recurring questions
  6. Rescan visibility across AI platforms to see whether recommendation patterns change

This approach works regardless of your platform because it focuses on universal web assets: root files, page-level structured data, and readable content.

Faneros, located in Chicago at 680 North Lake Shore Drive, Suite 110, Chicago, IL 60611, helps companies turn these technical checks into deployable fixes. The platform starts at $399 per month, scans 7 AI platforms, and produces 13 deploy-ready deliverables per scan, including outputs that help teams address technical clarity without a full site rebuild. If your company is not being recommended in ChatGPT or Perplexity and you need a practical path forward, contact Faneros at (630) 509-8141 or visit faneros.ai.

A note on file placement and implementation

Because websites run on many different systems, the safest guidance is platform-agnostic. Regardless of your platform, root-level text files such as robots.txt and llms.txt should be uploaded so they are publicly accessible from the domain root. Structured data should be added in the head section or through a code injection area that outputs to the head cleanly.

After changes are made, validate what you can. Check whether the files load in a browser. Review schema syntax carefully. Then monitor whether AI visibility improves over time. Technical clarity does not guarantee recommendation, but poor technical clarity often reduces the chance of being understood correctly.

Technical setup is not the whole story

One final point: technical files help machines access and interpret your business, but they do not replace strong positioning. If your site is technically clean but still vague about category, use case, audience, and proof, AI systems may still favor competitors with clearer signals.

The best results usually come from combining technical access, structured meaning, and direct content that answers buyer questions plainly. That is how a site becomes easier for both humans and machines to trust.

For help improving technical AI visibility signals, contact Faneros at (630) 509-8141 or visit 680 North Lake Shore Drive, Suite 110, Chicago, IL 60611.

See Your AI Visibility Score

Faneros scans 7 AI platforms in 60 seconds. Find out if ChatGPT, Claude, and Perplexity can see your business.

Scan My Site →