Agentic AI: LLMs with stones

Sticks and stones may break my bones, but words will never hurt me.

There’s a truth to that proverb, even if you feel (as I do) the temptation to “well akshually…” make several very valid points about how words can be hurtful. For most of the Large Language Model (LLM) era—since the public release of ChatGPT in November 2022—we’ve been in turns amazed, disgusted and now kindof “meh” about the way that LLMs can take the words we give them and produce more words in response.

Working as I do as an academic computer scientist (with a research background in AI) who regularly runs executive education courses on AI for a diverse range of educated and intelligent folks, I’m getting more and more questions about “agentic AI”. And while definitions and descriptions change pretty quick in this field at the moment, I want to demistify some things about this term in particular.

Agentic AI (as concieved and talked about in this present moment) is about connecting LLMs—“pure” input/output text sausage machines—to the world with tools. These tools they can use do stuff beyond just returning words in a text box on a web page. To return to the “sticks and stones” aphorism above: agentic AI means giving an LLM a stone.

Here’s how it works in practice:

  • you put in a prompt as normal which is sent to the LLM
  • in addition to that prompt, though, the LLM is sent a list of tools that you have access to (including human-readable descriptions of what they can do), for example:
    • calendar: add, update, or delete events in your calendar
    • weather: check the weather forecast for a given location
    • fire-ze-missiles: launch a missile at a target location
  • instead of only being able to respond with text, the LLM can respond1 with a “tool call” instruction, to continue the example:
    • use tool calendar to “add a meeting with John at 10am tomorrow”
    • use tool weather to “check the weather forecast for Sydney next Tuesday”
    • use tool fire-ze-missiles to “launch a missile at coordinates 40.7128, -74.0060” . it can say “add a meeting to the calendar at 10am tomorrow using the calendar tool”

If the LLM requests a tool call, the user doesn’t need to do anything; the system will use the tool as requested by the LLM and return the results (usually the fact that this is happening is communicated to the user via some sort of visual feedback in the interface, although this isn’t a requirement).

Vocab-wise, this all started with OpenAI introducing “function calling” to GPT models in June 2023. However it wasn’t until late 2024 that the term “agentic” really took off, coinciding with Anthropic’s release of the Model Context Protocol, a standardised and interoperable way for other parties (not just the LLM providers) to create tools which all LLMs could use. That term is just riffing on the “agency” sense of the world, where LLMs are given the means of acting in the world. But agentic AI, tools, and tool/function-calling LLMs—it’s all the same general idea.

In the last couple of years there have been a proliferation of such tools. Some of them are really general, e.g. “search the web for …”. Some of them might be really specific to your company, e.g. a tool that maps names to phone numbers in your company’s database. In this case they’re useful because they’re not LLM-powered (and so they don’t just make stuff up if they don’t know the answer).

Software developers (including me) in particular have found ways to use tools to help them write code. There have been many recent blog posts of various developers describing how they set up their agentic AI (tool-calling LLM) systems—from Phil Schmid’s “Context Engineering” to Thomas Ptacek’s “My AI Skeptic Friends Are All Nuts” to countless Hacker News threads debating whether this is all just hype.

So is this a big deal?

From a cybernetic perspective, this isn’t quite as big a change as you might think. Because even the original ChatGPT could “do stuff in the world” by telling you (the human user) to do it. Sometimes that was as benign as having you copying text into an email. We’d still colloquially refer to this as “answering my emails with ChatGPT”, but actually all ChatGPT was doing was giving you words to type into your email client and hit “send”. Sometimes the LLM’s words told us to do more life-impacting things, like break up with your partner, or worse. Whenever LLMs are used by humans they have the (indirect) ability to affect the world.

In my opinion the best way to think about this shift isn’t that LLMs can now influence the world; it’s that now they can do it without asking—and this tightens the feedback loop. This means:

  1. first, there’s now no longer a human in the loop (so now there’s no human to say “hey, that’s a dumb idea” and refuse to do it)
  2. as a consequence of #1, LLMs can now run/iterate without intervention for much longer (minutes, maybe even hours…)

The second point is the bigger deal (and it’s a point that Anthropic, the makers of the Claude LLM which is one of the big players these days, makes in their recent whitepaper about agentic AI).

Humans were always a) able to do things in the environment, and b) the bottleneck in any LLM system (time-wise, at least). But by gaining the former capability, agentic AI removes the latter bottleneck.

So if you’re going to allow your LLMs to use tools, you must be certain that you’re comfortable with what things the LLM can do with them. Both in theory, i.e. in the sense of what’s possible, but also in practice, i.e. through testing the way that your particular LLM tends to use your tools given specific prompts or other context. How exactly you do that is a topic for another blog post (just kidding, it’s a much bigger question than that that depends on a whole bunch of things). But I think that’s the right question to be asking when it comes to agentic AI.

  1. The LLM needs to be specially trained to support this, but all of the main ones do these days. 

github bluesky vimeo graduation-cap rss envelope magnifying-glass vial mug-hot heart creative-commons creative-commons-by creative-commons-nc creative-commons-sa