Skip to main content
Back to Resources
AI in Education

The Centaur Assignment: 3 Examples That Flip AI From Crutch to Gym

By Nathan Critchett · January 7, 2026

In 1997, Garry Kasparov, the greatest chess player who ever lived, lost to a machine. Deep Blue beat him. It was supposed to be the end of human chess.

It wasn't. What happened next was more interesting.

Kasparov noticed something. In freestyle chess tournaments, where humans could team up with computers, the winners weren't the best human players. They weren't the best computers, either. The winners were average players who were excellent at collaborating with machines. They knew when to trust the computer's calculation and when to override it with human intuition. Kasparov called them Centaurs: half human, half machine, stronger than either alone (Kasparov, 2017).

The most powerful chess player on Earth was neither human nor computer. It was a human-computer team where the human provided strategic judgment and the computer provided computational depth.

That was chess. This is education. And the principle is identical.

The Problem With AI On and AI Off

Right now, most schools have two modes: AI banned, or AI allowed. Neither works.

AI banned means students develop cognitive skills in an artificial vacuum, then graduate into a world where AI is everywhere. It's like banning calculators in a math class and then sending kids to work at NASA.

AI allowed, without redesigning the assignment, means students outsource the thinking. They type a prompt, get a polished output, change the font, and submit. The AI does the cognitive work. The student does the formatting.

Both approaches fail because they're asking the wrong question. As we explore in our whitepaper Cognitive Offloading: How AI Is Simultaneously Enhancing and Eroding Student Thinking, the real risk of "AI allowed" without structure is that students outsource the productive struggle that builds cognitive capacity. The question isn't whether to allow AI. It's what cognitive demand you're placing on the student when AI is in the room.

The Cognitive Demand Matrix

Think of it as a 2x2:

AI OffAI On
Low Cognitive DemandWorksheet busyworkAI does the busywork (fine, but pointless)
High Cognitive DemandTraditional deep thinkingThe Centaur Quadrant

Most "AI-integrated" assignments live in the bottom-right: AI On, Low Demand. The student uses AI to complete a task that didn't require deep thinking in the first place. The AI makes it faster. Nobody gets smarter.

The magic quadrant is the top-right: AI On, High Demand. The student uses AI AND the cognitive load goes up, not down. The AI becomes a tool that amplifies the difficulty of the thinking, not a shortcut that eliminates it.

This is the Centaur Quadrant. Here's what it looks like in three subjects.

Example 1: English, The Thesis Architect

The old assignment: "Write a five-paragraph essay arguing whether the American Dream is alive or dead in The Great Gatsby."

Students know the game. Open ChatGPT. Type the prompt. Receive 500 words of competent literary analysis. Change a few phrases. Submit. A-minus.

The AI did the thinking. The student did the typing.

The Centaur version:

Step 1: Generate. Use AI to produce three different thesis statements about the American Dream in Gatsby. Not one. Three. Each must take a fundamentally different position.

Step 2: Evaluate. For each thesis, write a paragraph explaining its strengths and weaknesses. Which one has the most interesting tension? Which one would be hardest to argue, and why might that make it the best choice? Which one does the AI seem most "comfortable" with, and what does that tell you about its limitations?

Step 3: Build. Choose the strongest thesis. Use AI to draft the essay. Then identify and improve at least three places where the reasoning is shallow, the evidence is weak, or the AI is pattern-matching instead of actually analyzing the text.

Step 4: Defend. Write a one-page reflection: Why this thesis? What did the AI get wrong? What did you change and why? Where is YOUR thinking in this essay?

Notice what happened. The student's job shifted from producing the essay to architecting the thinking. They're evaluating, choosing, critiquing, and defending, all cognitive work the AI cannot do for them. The AI generates raw material. The student is the architect.

The essay itself is almost beside the point. The thinking IS the assignment.

Example 2: Science, The Lab Report Auditor

The old assignment: "Write a lab report for our pendulum experiment, including hypothesis, methodology, data analysis, and conclusion."

Lab reports are particularly vulnerable to AI because they follow rigid templates. Any LLM can produce a flawless lab report from a data set. Students know this. Teachers know students know this. The mutual pretending is exhausting.

The Centaur version:

Step 1: Generate. Feed your raw experimental data to AI. Have it produce a complete lab report: hypothesis, methodology, analysis, conclusion, the works.

Step 2: Audit. Now tear it apart. Answer these questions in writing:

  • Did the AI correctly interpret statistical significance, or did it just say the results "suggest" something without doing the math?
  • Did it identify the actual sources of error in YOUR experiment, or did it generate generic error sources that could apply to any pendulum lab? (Hint: AI almost always does the second one.)
  • Does the conclusion follow from the data, or did the AI write a conclusion that sounds good but isn't actually supported by what you measured?
  • Where did the AI confuse correlation with causation, or assume linearity where your data shows something else?

Step 3: Rewrite. Rewrite the analysis and conclusion sections using your understanding of what actually happened in the lab. Not what a language model thinks a pendulum experiment should show, but what YOUR pendulum actually did, including the weird parts.

Step 4: Compare. Submit both versions (the AI's and yours) with annotations explaining every place they differ and why your version is more scientifically accurate.

The student who does this well has demonstrated something no AI detector could ever measure: they understand the science. They know what happened in the lab. They can distinguish between real analysis and plausible-sounding nonsense. They have scientific judgment.

The student who can't do this, who can't find the errors, who can't tell the difference between their data and a generic template, has revealed exactly where their understanding breaks down. That's diagnostic gold.

Example 3: Math, The Trap Designer

The old assignment: "Solve these 20 equations."

Twenty repetitions of a mechanical procedure. AI solves all twenty in seconds, showing work that looks identical to a textbook solution. This assignment was already low-value before AI. Now it's zero-value.

The Centaur version:

Step 1: Solve with AI. Give AI five equations. Have it solve each one and show its method.

Step 2: Verify. Check its work. For each equation, answer: Did the AI choose the most efficient method? Is there a faster or more elegant approach it missed? Did it make any errors? (AI makes math errors more often than students expect, and this is genuinely surprising to most kids.)

Step 3: Analyze. For each equation, explain WHY the AI chose the method it did. What pattern in the equation triggered that approach? When would that approach fail?

Step 4: Design a trap. This is the hard part. Create an equation that would trick the AI into using the wrong approach. Explain why the trap works: what about the equation's structure would mislead a pattern-matching system? Then solve it correctly yourself, using the method the AI should have used.

A student who can design a mathematical trap for an AI understands mathematics at a level that no amount of problem-set grinding would produce. They're not just following procedures. They're reasoning about procedures, understanding when methods work, when they break, and why. That's mathematical thinking.

The trap design also reveals something profound: it forces students to understand AI's failure modes. Students who do this exercise walk away knowing, at a gut level, that AI is not infallible. That's a lesson worth more than a hundred solved equations.

The Spotter Methodology

All three examples follow the same principle, and it comes from weightlifting, not education.

A good spotter stands behind the lifter at the bench press. Not grabbing the bar. Not doing the work. But close enough to prevent catastrophe. If you grab the bar every rep, you guarantee the lifter stays weak. They never bear the full weight. If you walk away entirely, you risk a crushed sternum. The art is in the distance: close enough to keep them safe, far enough to make them feel every pound.

That's the teacher's stance in a Centaur classroom. Not banning AI (walking away). Not doing the thinking for students (grabbing the bar). Standing close enough to structure the cognitive challenge, far enough for students to feel the full weight of the thinking.

This changes the teacher's default responses:

When a student asks, "What should I do?", the old answer was an instruction. The Centaur answer: "What are three approaches you see? Which would you try first, and why?"

When a student says, "The AI gave me this. Is it right?", the old answer was verification. The Centaur answer: "What would have to be true for that to be wrong? How would you check?"

When a student says, "I'm stuck", the old answer was a hint toward the solution. The Centaur answer: "Where exactly did you get stuck? What did you try? What happened?"

Every response redirects the cognitive load back to the student. Not cruelly. Not unhelpfully. But deliberately. The student must bear the weight.

The Pattern Across All Three

For the full Centaur framework, including the Cognitive Demand Matrix, the Spotter Methodology, and rubric design, see our whitepaper The Centaur Classroom: Designing Human-AI Learning That Builds Thinkers, Not Dependents.

Look at the three assignments again. In every case, the student's job shifted from producing the answer to architecting the thinking.

In English: the student evaluates, selects, critiques, and defends. In Science: the student audits, compares, and corrects. In Math: the student verifies, analyzes, and designs.

The AI handles the computational, generative, template-following work, the stuff it's good at. The student handles the judgment, evaluation, and strategic decision-making, the stuff humans are good at and AI isn't.

That's not using AI as a crutch. It's using AI as a gym. The machine provides the resistance. The student does the lifting.

Start This Week

You don't need a curriculum overhaul. You don't need a PD series. You need one assignment.

Pick a unit you're teaching in the next two weeks. Find ONE assignment that AI could easily complete. Then redesign it using this framework:

  1. Let the AI generate. Don't fight it. Use it.
  2. Make the student evaluate the output. What's wrong? What's shallow? What's missing?
  3. Make the student improve it. Using their own knowledge, not another prompt.
  4. Make the student defend their choices. In writing. Out loud. Both.

Try it once. See what happens. You'll learn more about what your students actually understand in one Centaur assignment than in a semester of traditional output grading.

The Centaur doesn't ban the machine or surrender to it. The Centaur rides it.

Teach your students to ride.

Related Reading

Want to see this in action?

We'll walk you through a real report and recommend the right starting point for your team.