Why your definition of “agent” matters CORE
You are about to spend a lot of hours directing AI systems to do real work. How well that goes depends almost entirely on one thing that nobody teaches: how you think about what you are directing.
If you think an AI agent is magic — a box that “just knows things” — you will hand it vague tasks and be disappointed when it invents facts. If you think it is a fancy search engine — a tool that looks things up — you will ask it to find sources it does not have access to and blame it for being wrong. Both of these ways of thinking produce the same result: a frustrated human and a mediocre output.
Good directors have a sharper picture. They see an agent as a specific kind of machine with specific parts, each of which can be broken or missing. When the output is bad, they can point at the part that failed. When the task is new, they can figure out what the agent needs to succeed. That sharper picture is what this lesson gives you.
The definition: model + loop + tools + state CORE
An AI agent is four things working together.
A model. A large language model — the AI you've used as a chatbot is one. The model is the “brain.” On its own, it can read, write, and reason, but it cannot do anything outside the conversation. (You don't need to know how a model is built to direct one. You only need to know what it does and where it stops.)
A loop. Code that runs the model more than once. After the model responds, the loop decides whether the task is done. If not, the loop feeds the model's output back in — along with new information — and runs it again. A chatbot runs the model once per user message. An agent runs the model many times, often without the user in between.
Tools. Specific actions the model is allowed to take. Reading a file. Searching the web. Sending an email. Running a shell command. Querying a database. (You'll sometimes hear these called “functions” — same idea: a named, defined thing the AI can do.) Without tools, the model can only produce text from what it already knows. With tools, it can reach out into the real world — your computer, the internet, your calendar — and act.
State. Information that persists between loop iterations. The conversation so far. A scratchpad of notes the agent writes to itself. Files it has read. Results from tools. State is the agent's memory for this task; without it, every step of the loop would start from nothing.
Chatbot vs. agent, in one concrete example CORE
Imagine you are writing a paper on the Dust Bowl.
You open a plain chatbot and type: “What were the top three causes of the Dust Bowl?” The chatbot responds in one turn. It produces three causes, drawn from what the model remembers from training. It cannot check a source. It cannot open a book. It cannot tell you how confident it is. The loop runs once. There are no tools. State is just your question and its answer.
Now imagine you open an agent and type: “Find the three most commonly cited causes of the Dust Bowl across at least five academic sources. For each cause, give me a one-paragraph explanation with a direct quote and a citation. Save it as dust-bowl-brief.md on my desktop.”
The agent does something very different. It plans. It calls a search tool. It reads the results. It notices that one source disagrees with three others and decides how to handle that. It writes the brief. It opens your filesystem and saves the file. Along the way, it has run the model maybe twenty times. It has called five or six tools. It has kept state — a running list of sources it has already checked, so it does not re-read the same page.
Same underlying model. Wildly different output. The difference is the loop, the tools, and the state — the parts that turn a model into an agent.
Two misconceptions that will make you a worse director CORE
Misconception 1: the “magic” frame. “It just knows things.” Students who hold this frame give vague commands like “make this better” or “fix my essay” and are disappointed when the output is generic. The model does not know what you have not told it. It does not know your assignment, your teacher's preferences, what you have already tried, or what you are actually optimizing for. Vague input, vague output — every time.
Misconception 2: the “search engine” frame. “It just looks stuff up.” Students who hold this frame assume every answer is grounded in a real source. It is not. A model without a search tool answers from its trained parameters — essentially, from compressed patterns of text it saw during training. Those patterns are often right and sometimes confidently wrong. This is where hallucinations come from: the model is not lying, it is pattern-completing. If you want grounded answers, you have to give the agent a tool that reaches real sources, and you have to direct it to use that tool.
Both misconceptions fail for the same reason: they ignore the parts. If you cannot name model, loop, tools, and state separately, you cannot tell which part is letting you down.
Why this changes everything you do next CORE
Everything you will learn in this course maps onto those four parts.
Module 2 sets up the models you will use — local models on your own machine, cloud models through an API. Module 3 teaches you to direct coding loops. Modules 4 through 6 teach you the tools that matter most — research, email, calendar, scheduled tasks. Module 7 teaches you to build your own tools when the defaults fall short. Module 8 teaches you to chain loops together, so one agent's output becomes another agent's input. Module 9 teaches you to protect state from leaking where it should not go.
Before you direct anything, ask four questions:
Do I have the right model for this task? Is my loop well-framed — does the agent know when to stop? What tools does it need, and does it have them? What state am I giving it up front, and what will it carry forward?
Those four questions are the whole course. This lesson gave you the words to ask them.
Try it CORE
Agentic or not?
Think through the questions and answer in your head — you don't need to write anything down.
Instructions. Below are eight interactions with an AI system. For each, decide whether it is chatbot-style (one-turn, no tools, no state beyond the conversation) or agent-style (multi-turn loop, tools, persistent state). Name the specific part — loop, tool, or state — that tips the balance. There is no trick: most are clearly one or the other, and one is deliberately ambiguous. Be ready to explain your reasoning.
- You ask an AI in a web chat, “Summarize the plot of The Great Gatsby.” It responds in one message.
- You ask an AI on your laptop to find the cheapest flight from Denver to Nashville next weekend; it searches three airline sites, compares prices, and reports back.
- You paste an essay into a chat window and say, “Make this better.” The AI returns an edited version.
- You tell an AI, “Every morning at 7am, check my calendar, check the weather, and draft a short daily plan in a note titled ‘Today.’” It does this every day.
- You ask an AI to translate a Spanish paragraph into English. It returns the translation in one response.
- You ask an AI to triage your inbox: star urgent emails, archive newsletters, and draft replies to anything that mentions your part-time job.
- You ask an AI, “Plan a weekend road trip from my house to a national park within four hours, with a budget of $300. Find lodging options, build a packing list, and save it all to a note called Road Trip.” It does so over the next few minutes, checking maps, hotel sites, and weather.
- You ask an AI, “What's the capital of Mongolia?” It answers in one sentence.
Do this for real, today. Pick one chatbot-style item from the list above — whichever you would actually use — and try it in any AI you can open (ChatGPT, Claude, Gemini, the chat box on your phone). One turn, one answer. Pay attention to what it does well and where it stops. You don't need to write anything down. The point is to feel a one-turn loop before we draw one on the page in Lesson 1.3.
Done with the hands-on?
When the recipe steps and any activity above are complete, mark this stage to unlock the assessment, reflection, and project checkpoint.
Key concepts CORE
- An AI agent is a model + a loop + tools + state. If any of the four is missing or broken, the agent will fail in a predictable way.
- A chatbot is a model with a one-turn loop and no tools. An agent runs the model many times, calls tools, and keeps state.
- The magic frame (“it just knows things”) leads to under-specified tasks and generic output.
- The search engine frame (“it just looks stuff up”) leads to wrong expectations about grounding and invisible hallucinations.
- Good directors can name which part of the agent failed when the output is bad.
Quick check CORE
Four questions. Pick the best answer, then reveal the explanation — the why matters more than the letter.
Show explanation
Answer: B. The four parts are what make an agent an agent. A confuses length of conversation with agency; a long chatbot conversation is still one-turn-at-a-time with no tools. C is too narrow — an agent does not require internet access; a local agent calling local tools is still an agent. D confuses training size with architecture; a huge model with no loop and no tools is still a chatbot.
Show explanation
Answer: B. This is the “search engine frame” failure mode. Without a grounding tool, the model answers from trained parameters — compressed patterns of text. Confidence is not the same as correctness. A anthropomorphizes — the model is not intending anything. C and D misapply the vocabulary; a chatbot has a one-turn loop and minimal state, so neither was the failure.
Show explanation
Answer: B. The magic frame — “it just knows things” — leads to under-specified tasks. The model did not know your rubric, your argument, or what “better” meant to you, because you did not tell it. Naming this failure mode is the first step toward fixing it: next time, give the agent your rubric, your draft, and a clear target for “better.”
Show explanation
Answer: D. State is the agent's working memory for the task — the conversation so far, scratchpad notes, tool results. The model itself does not “remember” between calls; each call is stateless. The loop decides when to run the model again, but it does not store the information — it passes state in and out. Tools produce information that goes into state. Lesson 1.4 spends an entire session on this, because state is where most real-world agent failures begin.
Reflection prompt
Name the frame you've been using.
In 4–6 sentences, answer: Which of the two misconceptions — the magic frame or the search engine frame — have you been more guilty of in your own use of AI so far? Give a specific example from the past month. Then describe, using the four parts, what you would change next time.
The goal is not self-criticism. The goal is to notice the pattern while it is fresh, so you can break it as we move forward.
Project checkpoint
One sentence. One note.
You will not start your capstone design until Lesson 1.5. For now, do this small thing and keep the result.
Open a new note in any text editor or notes app — TextEdit on Mac, Notepad on Windows, Apple Notes, Google Docs, anything you'll find again. Title it “My first loop” and write one sentence: “A task I spend real time on every week that I suspect an agent could help with is ____.”
You do not need to commit to it. Do not research tools. Do not plan a build. Just name one candidate task, in one sentence, and save the note somewhere you'll see it again. You'll add to it across the next four lessons; by Lesson 1.5 it becomes the seed for your first loop design — and by Module 10, the seed of your capstone.
Want to keep going?
The full course picks up where this leaves off — local models, coding agents, research and email agents, multi-agent orchestration, and a capstone system the student ships and operates.
30-day no-questions refund. First year of Recipe Book updates included. Credit documentation for parents included.
Next free lesson
Lesson 1.2 — How a language model reasons.
Tokens, next-token prediction, why confidence is not correctness, and the mechanism behind “think step by step.”