Skip to main content

The Agentic Future

AI agents that take actions, not just generate text. Early days, but the implications for enterprise are significant.
20 October 2023·6 min read
Mak Khan
Mak Khan
Chief AI Officer
The AI conversation has been about generation - text, images, code. But something more interesting is emerging. AI systems that don't just produce outputs but take actions. Agents that can reason, plan, use tools, and execute multi-step tasks. We're early. But the enterprise implications are worth thinking about now.

What Are AI Agents?

Current AI tools are reactive. You give them a prompt, they give you an output. The interaction is one-turn: input, output, done.
AI agents are different. An agent receives a goal, breaks it into steps, uses tools to execute those steps, evaluates the results, and adjusts its approach. The interaction is multi-turn and autonomous. The agent does work, not just generation.
A simple example: instead of "summarise this document" (generation), an agent could handle "find all documents related to last quarter's claims, summarise the trends, compare to the previous quarter, and draft a report for the risk committee." That's research, analysis, synthesis, and writing - multiple steps, multiple tools, one goal.

Where This Is Today

Let me be clear: we're early. The current generation of agents is impressive in demos and fragile in practice. They work well on constrained tasks with well-defined tools and clear success criteria. They struggle with ambiguity, long-horizon planning, and error recovery.
What exists today:
Code agents. Systems like GitHub Copilot's upcoming workspace features that can modify multiple files, run tests, and iterate on code changes. These are the most mature agent applications because code provides clear feedback signals - does it compile? Do the tests pass?
Research agents. Systems that can search multiple sources, synthesise information, and produce structured reports. AutoGPT and similar projects demonstrated the concept (if not the reliability). More constrained versions - agents that search a specific knowledge base and compile findings - are genuinely useful.
Process agents. Early experiments with agents that can navigate software interfaces, fill out forms, and complete administrative workflows. These are fragile but the trajectory is clear.
<5%
of enterprise AI deployments involve agentic capabilities in late 2023, with significant growth expected
Source: Gartner, Emerging Technology Survey, Q3 2023

The Technical Architecture

Agents combine several capabilities we already understand:
LLM as reasoning engine. The language model provides the ability to interpret goals, break them into steps, and decide what to do next.
Tool use. The agent can call APIs, query databases, read documents, write files, send messages. Each tool extends the agent's capability beyond text generation.
Memory. The agent maintains context across multiple steps. What has it done? What did it learn? What should it do next? This is harder than it sounds - current context windows are finite and agents lose coherence over long task sequences.
Evaluation. The agent assesses whether its actions achieved the intended result. Did the search return relevant documents? Did the code change pass tests? Did the output match the goal?
The orchestration of these components is where the engineering challenge lives. Getting an LLM to generate text is straightforward. Getting it to reliably execute a ten-step workflow with tool calls, error handling, and quality assessment is significantly harder.

Enterprise Implications

What Changes

Automation scope expands. Current automation handles structured, rule-based tasks. Agents can handle semi-structured, reasoning-intensive tasks. The set of automatable work gets substantially larger.
The integration layer matters more. Agents are only as useful as the tools they can access. An agent that can read your knowledge base, query your CRM, and update your project management system is dramatically more useful than one that can only generate text. The investment in API infrastructure and system integration pays compounding dividends.
Monitoring becomes critical. An agent acting autonomously introduces a new category of risk. What happens when the agent takes an incorrect action? How do you audit what it did? How do you set boundaries on what it's allowed to do? These are governance questions that current frameworks don't fully address.

What Doesn't Change

Human oversight remains essential. Agents should operate within human-defined boundaries with human review of consequential actions. The "fully autonomous AI" narrative is premature and, for enterprise contexts, inappropriate.
Data quality still matters. Agents with access to bad data take bad actions. The data readiness challenge doesn't disappear - it becomes more consequential because the system is acting, not just informing.
Clear problem definition still matters. An agent can execute a well-defined goal. It cannot define the goal for you. The strategic thinking about where AI adds value is still a human responsibility.

What to Do Now

You don't need to build agents today. But you should:
  1. Watch the space. Agent capabilities are improving rapidly. What's fragile today will be robust in 12-18 months.
  2. Invest in your API layer. Every system that has an API is a potential agent tool. The enterprises with well-documented, accessible APIs will deploy agents faster.
  3. Think about boundaries. What should an AI agent be allowed to do in your organisation? What actions require human approval? Start thinking about this now.
  4. Identify high-value multi-step workflows. The tasks that involve multiple systems, multiple information sources, and multiple decision points. These are the workflows where agents will create the most value.
The future of enterprise AI isn't chatbots. It's agents that do real work. We're building toward that future - carefully, incrementally, and with appropriate governance.