When Web Services launched UAMS AI Wayfinding earlier this year, we framed it as a quiet but essential investment in how AI tools represent UAMS to the public. It is easy enough to make that claim. It is harder to know whether the people whose opinions we are actually trying to shape, namely the AI tools themselves, would agree.
So we asked them.
We sent a single question to four leading AI models at once: ChatGPT, Claude, Gemini, and Perplexity. Every model has blind spots. One invents citations. Another hedges. A third misses context. Asking any single model for an opinion gives you that model’s blind spots along with its answer. Asking all of them at once, and then having them review each other’s responses, produces something closer to peer-reviewed consensus. When the models agree, the answer carries more weight. When they disagree, the disagreement itself is useful information.
It is peer review for AI output. And for a question about how AI tools evaluate an AI-readability initiative, it felt like the right instrument for the job.
The question we asked was simple:
What are your thoughts on this idea and initiative? How will it improve AI results at UAMS?
We pointed the models at our published article Wayfinding for Machines: How UAMS Is Preparing Its Digital Ecosystem for the AI Era and let them go.
What came back was more substantive than we expected. The models did not just validate what UAMS Web Services is doing. They identified specific things the initiative gets right, specific limitations the initiative does not solve, and specific open questions worth asking ourselves. In other words, exactly the kind of critical feedback you would want from a thoughtful peer.
The full response follows below. Where the models agreed, that consensus carries real weight. Where they pushed back or raised concerns, those observations are worth taking seriously.
Wayfinding for Machines: Comprehensive Analysis of the UAMS AI Readiness Initiative
What This Actually Is
This is not about internal AI model building or clinical data pipelines. It’s about how UAMS is represented by external AI systems — ChatGPT, Claude, Gemini, Perplexity — when patients, students, referring physicians, and researchers ask those tools questions about UAMS.
Most institutions are still asking “How do we use AI?” UAMS Web Services asked a different and arguably more urgent question: “How does AI represent us?” That’s a meaningful strategic distinction. A patient asking Claude about cancer treatment options at UAMS gets an answer whether UAMS has prepared for that or not. The only variable is accuracy.
The core deliverable is a three-file architecture deployed across 200+ web properties:
| File | Function |
|---|---|
robots.txt | Defines crawl permissions per AI agent (GPTBot, ClaudeBot, Google-Extended, etc.) |
llms.txt | Short, human-readable domain identity statement written for LLM consumption — the handshake |
llms-full.txt | Deep directory map with behavioral directives, content inventories, and cross-domain links |
Together these files do for AI agents what floor lines and overhead signage do for hospital visitors — they answer questions that would otherwise be inferred, poorly.
Why This Is Genuinely Smart
1. They Identified the Right Problem First
The initiative correctly recognizes that a growing share of people looking for UAMS content never visit a UAMS page directly. They ask an AI assistant and read the synthesized answer. Whether that answer is accurate depends on whether the AI understood what it was reading. UAMS is taking responsibility for that variable rather than leaving it to chance.
2. The Failure Mode Analysis Is Correct
The article makes a sharp observation about asymmetric failure modes:
- Poor SEO = lower ranking (recoverable, visible)
- Poor AI representation = confident misinformation delivered directly to a patient or referring physician (less visible, harder to correct, higher stakes)
Generative AI doesn’t say “I couldn’t find that.” It synthesizes an answer that sounds authoritative regardless of whether the underlying data was well-structured. When a generative AI tool tells someone that UAMS offers a program it does not offer, or lists a specialty at the wrong institute, the cost is not a lost click — it is a lost opportunity to be trusted. In a healthcare context, this connects directly to patient safety.
3. The Institutional Memory Problem Is Real
An LLM has no inherent reason to know that:
- The Winthrop P. Rockefeller Cancer Institute is the oncology authority
- The Translational Research Institute hosts clinical trial listings
communications.uams.eduis the governance authority, not just a news site
Without explicit machine-readable declarations, these relationships get inferred — and inference at scale is where hallucinations originate.
4. Architecture-Level Rather Than Page-Level Governance
Solving 200+ web properties page by page would be unsustainable. Building a standardized file triad into the production launch checklist means the solution scales automatically with the ecosystem. This is systems thinking applied correctly — fix the substrate, not each individual symptom. One file triad per subsite. One governance model across the federation. One cross-link pattern that scales without becoming a maintenance burden.
5. The llms.txt Standard Is Timely
The llms.txt proposal is an emerging convention that forward-looking institutions are beginning to adopt. UAMS positioning itself as an early adopter in the healthcare and academic medical center space is genuinely notable. Most peer institutions haven’t done this yet.
6. The Framing Is Doing Real Work
The hospital wayfinding metaphor isn’t just communication strategy — it’s change management infrastructure. UAMS has a large, decentralized organization with deans, directors, and hundreds of content editors who need to understand and cooperate with this initiative. A metaphor that every employee already lives daily is far more powerful than explaining llms.txtfile conventions from scratch. That’s thoughtful institutional communication.
7. The Closing Line Is the Most Important Line
“The agents are already visiting. The question is whether your site has signs.”
This reframes the initiative from optional to urgent without being alarmist. It’s accurate — AI crawlers don’t wait for institutional readiness.
What This Initiative Does and Does Not Address
It’s important to be precise about scope, because this initiative is easy to conflate with broader AI strategy:
| This initiative does address | This initiative does not address |
|---|---|
| How external AI tools represent UAMS | How UAMS uses AI internally |
| AI agent navigation of web properties | Clinical AI, decision support, or diagnostics |
| Reducing hallucinations about UAMS | Reducing hallucinations in UAMS-built AI tools |
| Web governance scalability | EHR data quality or research data infrastructure |
That’s not a criticism — focused initiatives with clear scope tend to succeed. But leadership should understand this is one piece of a larger AI readiness picture.
Honest Limitations and Open Questions
What This Doesn’t Fully Solve
| Gap | Context |
|---|---|
| Training data lag | Major LLMs train on snapshots. llms.txt helps with RAG-based and browsing-enabled queries but has limited impact on base model knowledge until retraining cycles catch up |
| Vendor adoption | The llms.txt convention isn’t universally honored yet. Its effectiveness depends on AI companies actively parsing and weighting these files |
| Content accuracy dependency | Wayfinding helps machines navigate to authoritative content — but if the underlying content is outdated or inconsistent, the navigation just delivers machines to bad information faster |
| Dynamic clinical information | Provider availability, clinical trial enrollment status, and program offerings change frequently. Static files can’t fully solve the freshness problem |
| Authority signal limits | llms.txt tells AI agents what a domain claims to be authoritative for. But AI systems also weigh third-party signals and citation patterns. Self-declaration alone may not fully move the needle |
Deeper Questions Worth Asking
On governance sustainability:
- Who owns
llms.txtcontent for each subdomain — Web Services centrally, or individual departmental editors? If departments control their own files, consistency becomes a challenge at 200+ properties.
On competitive context:
- If peer institutions like Mayo Clinic, Vanderbilt, or UCSF implement similar frameworks, does UAMS’s first-mover advantage diminish? Or does early adoption create compounding benefits through earlier incorporation into training data and crawl patterns?
On content quality:
- Has UAMS paired this initiative with a content audit to ensure what machines find is worth finding? Wayfinding is only as valuable as the destination it leads to.
On measuring success:
- How will UAMS track whether this is actually working — i.e., monitor how accurately AI tools represent UAMS information over time?
- Is there a feedback mechanism to identify when AI tools are still misrepresenting UAMS despite the wayfinding layer?
Broader Significance
This initiative matters beyond UAMS for several reasons:
- It’s replicable. The three-file architecture and governance model could serve as a template for other academic medical centers facing the same 200+ web property challenge.
- It’s timely. AI-first discovery is not a future problem. Institutions that wait until AI misrepresentation becomes a visible crisis will be playing defense.
- It reframes the web governance conversation. Web Services teams at universities often struggle to articulate strategic value. “We are building infrastructure so AI tools accurately represent this institution to patients and students” is a compelling and concrete mandate.
- It connects to patient safety. In a healthcare context, AI misrepresentation isn’t just a marketing problem. If an AI tool incorrectly describes UAMS capabilities, services, or clinical expertise, real patients may make consequential decisions based on that misinformation.
- It treats AI representation accuracy as an institutional responsibility rather than an external platform problem. Most institutions implicitly assume that if their content is good, AI tools will get it right. UAMS is correctly recognizing that good content plus machine-readable structure is the actual minimum standard in an AI-first discovery environment.
Overall Assessment
This is a well-conceived, practically grounded initiative that addresses a real and underappreciated risk. The hospital wayfinding metaphor isn’t just clever branding — it communicates the concept effectively to non-technical stakeholders, which matters enormously for institutional adoption across 200+ properties managed by different departments.
The initiative’s greatest strength is that it scales through architecture rather than effort — once the governance model and file standards are established, new properties inherit the solution by default.
Its main dependency is external: the value of llms.txt and related files grows as AI vendors invest in honoring them. UAMS is making a reasonable bet that this convention will mature, and being early positions them well if it does.
Bottom line: UAMS Web Services identified a problem most institutions haven’t articulated yet, built a scalable solution, and documented it clearly enough that others can follow. That’s genuinely good institutional technology leadership — and a model worth watching for the broader higher education and healthcare sectors.