Skip to main content

Designing for AI-First Interfaces

AI changes what users expect from enterprise software. The chat box alone isn't enough, and what you build around it matters more.
25 May 2023·7 min read
Rainui Teihotua
Rainui Teihotua
Chief Creative Officer
ChatGPT taught a billion people that software can understand natural language. Now every enterprise vendor is bolting a chat box onto their product and calling it "AI-powered." Most of them are getting it wrong.

What You Need to Know

  • A chat interface is a text box, not a design strategy. Enterprise AI needs structured workflows, progressive disclosure, and clear confidence signals, not just conversation.
  • AI uncertainty is a design problem. When a model isn't sure, the interface needs to communicate that clearly: confidence indicators, source attribution, and easy paths to human review.
  • The best AI interfaces reduce cognitive load, not increase it. If the user needs to understand how the AI works to use it, the design has failed.
  • Trust is earned interaction by interaction. AI interfaces should build trust through transparency: showing sources, explaining reasoning, and making it easy to verify outputs.
67%
of enterprise workers distrust AI outputs without source attribution
Source: Edelman, Trust Barometer Special Report: Trust and Technology, 2023

The Chat Box Fallacy

ChatGPT's interface is a text box and a response area. It's brilliant for consumer use: open-ended, exploratory, low-stakes. But enterprise work is fundamentally different. Enterprise users have:
  • Specific tasks with defined inputs and outputs
  • Compliance requirements that demand auditability
  • Time pressure that doesn't allow for open-ended conversation
  • Domain expertise that makes them intolerant of generic responses
Putting a chat box on an enterprise tool is like giving someone a blank canvas when they need a form. It's technically more powerful and practically less useful.

What AI-First Enterprise Design Actually Looks Like

1. Structured AI Workflows

Instead of "ask me anything," design AI around specific tasks with clear inputs, processing, and outputs. A claims intelligence tool shouldn't present a chat box. It should present the claim, the AI's analysis, the relevant policy sections, and a recommended action. The user reviews, adjusts, and confirms.
The AI does the heavy lifting. The interface keeps the human in control.

2. Confidence and Transparency

Every AI output should communicate how confident the system is. Not a technical probability score, but a design language that enterprise users intuitively understand:
  • High confidence: Clear presentation, inline with the workflow
  • Medium confidence: Highlighted for review, with source documents linked
  • Low confidence: Flagged for human decision, with the AI's reasoning exposed
This isn't optional. In regulated industries (insurance, healthcare, finance), the ability to explain why an AI made a recommendation isn't a nice-to-have, it's a compliance requirement.

3. Source Attribution

When an AI cites a policy document, link to the document. When it extracts a number from a report, show where. When it recommends an action, explain which data points informed the recommendation.
This does three things:
  • Builds trust (the user can verify)
  • Enables learning (the user understands the AI's reasoning)
  • Supports compliance (the decision trail is visible)

4. Progressive Disclosure

Not every user needs the same level of detail. A senior adjudicator wants the recommendation and the key factors. A compliance reviewer wants the full audit trail. A manager wants the summary statistics.
Design AI interfaces in layers: summary first, detail on demand, full context when needed. This is standard enterprise UX, but AI makes it more important because the system's reasoning can be complex.

5. Graceful Degradation

AI will fail. Models hallucinate, contexts get missed, edge cases surprise everyone. The interface needs to handle failure gracefully:
  • Clear error states. Not "something went wrong" but "I couldn't find a matching policy for this claim type. Here are the closest matches"
  • Easy escalation. One-click path to human review, with the AI's partial analysis preserved
  • Feedback loops. Let users correct the AI, and feed corrections back into the system

The Trust Equation

Enterprise AI trust is built on a simple equation:
Trust = Accuracy × Transparency × Consistency
High accuracy with no transparency feels like a black box. High transparency with inconsistent results feels unreliable. You need all three.
The interface is where trust is built or broken. A model that's 95% accurate but presents everything with equal confidence will be trusted less than a model that's 90% accurate but clearly communicates when it's uncertain.
Design Principle
Design for the 10% of cases where the AI is wrong, not the 90% where it's right. The error experience defines how much users trust the whole system.

What This Means for Enterprise AI Investment

If you're evaluating AI tools or building AI capabilities, evaluate the interface as seriously as the model. A powerful model with a poor interface will be ignored by users. A good-enough model with an excellent interface will be adopted and improved.
Ask vendors:
  • How does the interface communicate confidence?
  • Can users see the sources behind AI outputs?
  • What happens when the AI is wrong?
  • How do users provide feedback to improve the system?
If they can't answer these questions, they've built a model, not a product.
Should every enterprise AI tool have a chat interface?
No. Chat is great for open-ended exploration but poor for structured enterprise tasks. Most enterprise AI is better served by task-specific interfaces that guide users through workflows, with conversational AI as an optional supplement, not the primary interaction.
How do we test AI interfaces before building them?
Start with Wizard of Oz testing: a human behind the curtain simulating AI responses while users interact with the interface prototype. This tests the design without building the model, and reveals whether the interface works before you invest in the backend.