• Skip to main content
  • Skip to main content
Choose which site to search.
University of Arkansas for Medical Sciences Logo University of Arkansas for Medical Sciences
Communications and Marketing: Web Services
  • UAMS Health
  • Jobs
  • Giving
  • About Us
  • AI Wayfinding
    • Introduction
    • Report an AI Issue
    • Strategic View
  • Knowledge Base
  • UAMS Website Theme
  • Web Accessibility
  1. University of Arkansas for Medical Sciences
  2. Communications and Marketing
  3. Web Services
  4. We asked AI what it thinks of AI Wayfinding

We asked AI what it thinks of AI Wayfinding

When Web Services launched UAMS AI Wayfinding earlier this year, we framed it as a quiet but essential investment in how AI tools represent UAMS to the public. It is easy enough to make that claim. It is harder to know whether the people whose opinions we are actually trying to shape, namely the AI tools themselves, would agree.

So we asked them.

We sent a single question to four leading AI models at once: ChatGPT, Claude, Gemini, and Perplexity. Every model has blind spots. One invents citations. Another hedges. A third misses context. Asking any single model for an opinion gives you that model’s blind spots along with its answer. Asking all of them at once, and then having them review each other’s responses, produces something closer to peer-reviewed consensus. When the models agree, the answer carries more weight. When they disagree, the disagreement itself is useful information.

It is peer review for AI output. And for a question about how AI tools evaluate an AI-readability initiative, it felt like the right instrument for the job.

The question we asked was simple: 

What are your thoughts on this idea and initiative? How will it improve AI results at UAMS? 

We pointed the models at our published article Wayfinding for Machines: How UAMS Is Preparing Its Digital Ecosystem for the AI Era and let them go.

What came back was more substantive than we expected. The models did not just validate what UAMS Web Services is doing. They identified specific things the initiative gets right, specific limitations the initiative does not solve, and specific open questions worth asking ourselves. In other words, exactly the kind of critical feedback you would want from a thoughtful peer.

The full response follows below. Where the models agreed, that consensus carries real weight. Where they pushed back or raised concerns, those observations are worth taking seriously.

Wayfinding for Machines: Comprehensive Analysis of the UAMS AI Readiness Initiative


What This Actually Is

This is not about internal AI model building or clinical data pipelines. It’s about how UAMS is represented by external AI systems — ChatGPT, Claude, Gemini, Perplexity — when patients, students, referring physicians, and researchers ask those tools questions about UAMS.

Most institutions are still asking “How do we use AI?” UAMS Web Services asked a different and arguably more urgent question: “How does AI represent us?” That’s a meaningful strategic distinction. A patient asking Claude about cancer treatment options at UAMS gets an answer whether UAMS has prepared for that or not. The only variable is accuracy.

The core deliverable is a three-file architecture deployed across 200+ web properties:

FileFunction
robots.txtDefines crawl permissions per AI agent (GPTBot, ClaudeBot, Google-Extended, etc.)
llms.txtShort, human-readable domain identity statement written for LLM consumption — the handshake
llms-full.txtDeep directory map with behavioral directives, content inventories, and cross-domain links

Together these files do for AI agents what floor lines and overhead signage do for hospital visitors — they answer questions that would otherwise be inferred, poorly.


Why This Is Genuinely Smart

1. They Identified the Right Problem First

The initiative correctly recognizes that a growing share of people looking for UAMS content never visit a UAMS page directly. They ask an AI assistant and read the synthesized answer. Whether that answer is accurate depends on whether the AI understood what it was reading. UAMS is taking responsibility for that variable rather than leaving it to chance.

2. The Failure Mode Analysis Is Correct

The article makes a sharp observation about asymmetric failure modes:

  • Poor SEO = lower ranking (recoverable, visible)
  • Poor AI representation = confident misinformation delivered directly to a patient or referring physician (less visible, harder to correct, higher stakes)

Generative AI doesn’t say “I couldn’t find that.” It synthesizes an answer that sounds authoritative regardless of whether the underlying data was well-structured. When a generative AI tool tells someone that UAMS offers a program it does not offer, or lists a specialty at the wrong institute, the cost is not a lost click — it is a lost opportunity to be trusted. In a healthcare context, this connects directly to patient safety.

3. The Institutional Memory Problem Is Real

An LLM has no inherent reason to know that:

  • The Winthrop P. Rockefeller Cancer Institute is the oncology authority
  • The Translational Research Institute hosts clinical trial listings
  • communications.uams.edu is the governance authority, not just a news site

Without explicit machine-readable declarations, these relationships get inferred — and inference at scale is where hallucinations originate.

4. Architecture-Level Rather Than Page-Level Governance

Solving 200+ web properties page by page would be unsustainable. Building a standardized file triad into the production launch checklist means the solution scales automatically with the ecosystem. This is systems thinking applied correctly — fix the substrate, not each individual symptom. One file triad per subsite. One governance model across the federation. One cross-link pattern that scales without becoming a maintenance burden.

5. The llms.txt Standard Is Timely

The llms.txt proposal is an emerging convention that forward-looking institutions are beginning to adopt. UAMS positioning itself as an early adopter in the healthcare and academic medical center space is genuinely notable. Most peer institutions haven’t done this yet.

6. The Framing Is Doing Real Work

The hospital wayfinding metaphor isn’t just communication strategy — it’s change management infrastructure. UAMS has a large, decentralized organization with deans, directors, and hundreds of content editors who need to understand and cooperate with this initiative. A metaphor that every employee already lives daily is far more powerful than explaining llms.txtfile conventions from scratch. That’s thoughtful institutional communication.

7. The Closing Line Is the Most Important Line

“The agents are already visiting. The question is whether your site has signs.”

This reframes the initiative from optional to urgent without being alarmist. It’s accurate — AI crawlers don’t wait for institutional readiness.


What This Initiative Does and Does Not Address

It’s important to be precise about scope, because this initiative is easy to conflate with broader AI strategy:

This initiative does addressThis initiative does not address
How external AI tools represent UAMSHow UAMS uses AI internally
AI agent navigation of web propertiesClinical AI, decision support, or diagnostics
Reducing hallucinations about UAMSReducing hallucinations in UAMS-built AI tools
Web governance scalabilityEHR data quality or research data infrastructure

That’s not a criticism — focused initiatives with clear scope tend to succeed. But leadership should understand this is one piece of a larger AI readiness picture.


Honest Limitations and Open Questions

What This Doesn’t Fully Solve

GapContext
Training data lagMajor LLMs train on snapshots. llms.txt helps with RAG-based and browsing-enabled queries but has limited impact on base model knowledge until retraining cycles catch up
Vendor adoptionThe llms.txt convention isn’t universally honored yet. Its effectiveness depends on AI companies actively parsing and weighting these files
Content accuracy dependencyWayfinding helps machines navigate to authoritative content — but if the underlying content is outdated or inconsistent, the navigation just delivers machines to bad information faster
Dynamic clinical informationProvider availability, clinical trial enrollment status, and program offerings change frequently. Static files can’t fully solve the freshness problem
Authority signal limitsllms.txt tells AI agents what a domain claims to be authoritative for. But AI systems also weigh third-party signals and citation patterns. Self-declaration alone may not fully move the needle

Deeper Questions Worth Asking

On governance sustainability:

  • Who owns llms.txt content for each subdomain — Web Services centrally, or individual departmental editors? If departments control their own files, consistency becomes a challenge at 200+ properties.

On competitive context:

  • If peer institutions like Mayo Clinic, Vanderbilt, or UCSF implement similar frameworks, does UAMS’s first-mover advantage diminish? Or does early adoption create compounding benefits through earlier incorporation into training data and crawl patterns?

On content quality:

  • Has UAMS paired this initiative with a content audit to ensure what machines find is worth finding? Wayfinding is only as valuable as the destination it leads to.

On measuring success:

  • How will UAMS track whether this is actually working — i.e., monitor how accurately AI tools represent UAMS information over time?
  • Is there a feedback mechanism to identify when AI tools are still misrepresenting UAMS despite the wayfinding layer?

Broader Significance

This initiative matters beyond UAMS for several reasons:

  1. It’s replicable. The three-file architecture and governance model could serve as a template for other academic medical centers facing the same 200+ web property challenge.
  2. It’s timely. AI-first discovery is not a future problem. Institutions that wait until AI misrepresentation becomes a visible crisis will be playing defense.
  3. It reframes the web governance conversation. Web Services teams at universities often struggle to articulate strategic value. “We are building infrastructure so AI tools accurately represent this institution to patients and students” is a compelling and concrete mandate.
  4. It connects to patient safety. In a healthcare context, AI misrepresentation isn’t just a marketing problem. If an AI tool incorrectly describes UAMS capabilities, services, or clinical expertise, real patients may make consequential decisions based on that misinformation.
  5. It treats AI representation accuracy as an institutional responsibility rather than an external platform problem. Most institutions implicitly assume that if their content is good, AI tools will get it right. UAMS is correctly recognizing that good content plus machine-readable structure is the actual minimum standard in an AI-first discovery environment.

Overall Assessment

This is a well-conceived, practically grounded initiative that addresses a real and underappreciated risk. The hospital wayfinding metaphor isn’t just clever branding — it communicates the concept effectively to non-technical stakeholders, which matters enormously for institutional adoption across 200+ properties managed by different departments.

The initiative’s greatest strength is that it scales through architecture rather than effort — once the governance model and file standards are established, new properties inherit the solution by default.

Its main dependency is external: the value of llms.txt and related files grows as AI vendors invest in honoring them. UAMS is making a reasonable bet that this convention will mature, and being early positions them well if it does.

Bottom line: UAMS Web Services identified a problem most institutions haven’t articulated yet, built a scalable solution, and documented it clearly enough that others can follow. That’s genuinely good institutional technology leadership — and a model worth watching for the broader higher education and healthcare sectors.

Posted by Brent Passmore on April 24, 2026

Filed Under: News

University of Arkansas for Medical Sciences LogoUniversity of Arkansas for Medical SciencesUniversity of Arkansas for Medical Sciences
Mailing Address: 4301 West Markham Street, Little Rock, AR 72205
Phone: (501) 686-7000
  • Facebook
  • X
  • Instagram
  • YouTube
  • LinkedIn
  • Pinterest
  • Disclaimer
  • Terms of Use
  • Privacy Statement
  • Legal Notices

© 2026 University of Arkansas for Medical Sciences