Irreducibly Human Series · Northeastern University · College of Engineering

AImagineering: The Full Design Pipeline

Irreducibly Human: What AI Can't Do
Research Report | March 2026 | Ingested via Dev the Dev

Design thinking courses have always had a dirty secret: most of the semester is spent making the artifact.

Sketching, prototyping, iterating — the physical production of a thing to test with a user. The cognitive work of design — understanding who has the problem, deciding which problem is worth solving, choosing among fifty plausible directions, committing to one with your name on the consequences — gets compressed into the margins. Not because it is less important. Because making was hard, and it consumed the time.

AI has eliminated that excuse.

A brief that would have produced a working prototype in a week now produces one in an afternoon. Thirty concept sketches that once required a sprint now require a prompt. The constraint that forced design courses to spend eighty percent of their time on production has collapsed — and it has exposed what was always the harder work underneath it. That work is not making. It is thinking.

This course is what design thinking looks like when making is no longer the bottleneck.

The shift sounds like a gift. It is not straightforward. When ideation is a tool operation rather than a cognitive achievement, the discipline required to do it well changes entirely. The engineer who can generate fifty concepts in an afternoon now faces a problem the design sprint methodology was not built to address: how do you choose? Not which concept is most polished — the tool makes them all polished. Which concept is worth building. Which problem was worth solving in the first place. Whether the brief you were handed was even asking the right question. And at the end, when you have built and tested something: what do you actually commit to, and what are you accountable for if it fails?

Design Thinking as it is typically taught ends at Test. Prototype, test, iterate — and then, implicitly, hand the results to someone else. That omission was tolerable when the pipeline took a semester. It is not tolerable when AI compresses the pipeline to three weeks and leaves the human with nothing to do but make decisions.

The missing stage is the Commit — a specific course of action, with its evidential basis stated, its uncertainty acknowledged, and its accountability owned. No AI tool makes a Commit. That is the point of this course.

Course information

Course title Irreducibly Human: What AI Can't Do — AImagineering: The Full Design Pipeline
Credit hours 4
Delivery In-person | Lecture/Seminar (weekly) + TA-led Studio Lab (weekly)
Level Graduate
Prerequisite Botspeak or equivalent AI fluency foundation
Instructor Nik Bear Brown · ni.brown@neu.edu
Series Part of the Irreducibly Human series at Northeastern University — College of Engineering. Companion courses: Conducting AI, Causal Reasoning.

Who this course is for

This course is for engineers and technical practitioners who use AI tools in design work — and who have noticed that generating outputs is no longer the hard part.

What this course assumes

Botspeak proficiency or equivalent — you understand the difference between pattern completion and knowledge retrieval, you have used AI tools at specification-and-delegation level, and you have been asked to evaluate AI-generated design outputs rather than just produce them. No prior design thinking training required.

What this course does not assume

Prior engineering design coursework. Human-centered design, UX, or product development background. Any particular domain — the pipeline applies across engineering disciplines and the Studio Lab is structured to develop it in yours.

A note for students with prior design thinking training: Students who arrive with design methodology background sometimes find the early weeks disorienting — specifically, the course's insistence that conventional Design Thinking ends at Test and omits the most consequential stage. That disorientation is the course working as intended. The Commit is not a harder version of the prototype review. It is a different professional operation entirely. Students who treat this course as advanced prompting will produce technically correct exercises and miss it. Students who approach the Commit as genuinely new terrain will get the most from it.

What you will leave with

The complete AImagineering pipeline — Empathize, Define, Ideate, Prototype, Test, Commit — applied to a real design problem in your own domain, with explicit documentation of every human judgment call that an AI tool could not have made on your behalf.

A Commit document: a specific course of action you are willing to stand behind, with the evidence stated, the uncertainty acknowledged, and the accountability owned — not a recommendation, not a direction, not a set of options. A decision.

I can stand in front of a design problem, resist the pull to generate immediately, ask the right human questions first, direct AI tools with authority rather than dependency, and commit to a course of action I can defend on its merits — not just its polish.

What this course builds

By the end of this course, students can:

How the course is assessed

Grading is structured around a premise the course does not hide: AI tools handle the generation. What is being assessed is everything before and after.

Every assignment requires an AI Use Disclosure — not as compliance, but as the primary assessment instrument. Students document what they used, how they used it, what they changed, and — this field is not optional — what the AI could not do. Specifically: at least one judgment call that required their values, domain knowledge, or professional accountability. The disclosure that cannot name one such judgment call has not demonstrated that the student performed the irreducibly human layer of the design work.

The Irreducibly Human section of the final pipeline presentation carries 50% of the final project grade. Not because the design work is less important — because the honest, specific account of what required human judgment is exactly what the course is building.

Relative grading applies at the top of the scale, comparing students on depth of design judgment and specificity of the human judgment call identification. Absolute grading applies below the threshold.

How the course is structured

The course runs in three acts, each one targeting a different part of the pipeline that AI cannot perform.

Act One — The Material and the Human · Weeks 1–7

Act One is everything before a concept is generated. It opens with an experiment: thirty minutes, a brief, AI tools open. The output is good. That is not the point. The act then builds the capacities that need to happen before the tool runs: understanding the grain of the tool you're using and what it resists; conducting an empathy investigation that produces something a plausible AI persona would not; and reframing the brief itself, because the brief as given is a hypothesis, and it is often the wrong one. Act One closes with the midterm: a novel design situation, no scaffolding, demonstrate all four pre-ideation capacities as practice. The Act One gate is not whether the reframe is clever. It is whether the student can defend it against a skeptic.

Week 1

The Thirty-Minute Designer

The course opens with an experiment before any theory arrives. A brief is on the table. Thirty minutes. AI tools open. The output is good. Session B names what was assumed: a taxonomy of the human judgment the session skipped — about the user, the problem, the context, the values at stake, and what would happen if the output failed. The course argument is stated directly: AI has made ideation easy, which means everything before and after ideation is now the work of design.

Reading Response #1 — 30 pts
Weeks 2–3

Finding the Grain

A community health intervention designed entirely with a language model produces communication that is coherent, complete, and addressed to the community rather than from it. The grain as the explanation — every AI tool has affordances and resistances that shape what it produces, and working against the grain without knowing it produces the wrong kind of output. Students map the grain of their primary tool against their capstone domain brief: what it does naturally, what it resists, one specific design decision that changes because of this knowledge.

Studio Exercise #1 — 25 pts  Reading Response #2 — 30 pts
Weeks 4–5

What Simulation Cannot Feel

An AI-generated user persona next to a field note from an actual conversation. They are not the same document. AI personas are trained on population patterns — they produce plausible users, not specific ones. The empathy investigation protocol goes where simulation cannot: the unexpected, the contradictory, the embodied. Students conduct a genuine empathy investigation with minimum two human contacts and identify at least one finding that would not appear in any AI-generated persona for their user group — and name one design decision that would have been wrong without it.

Studio Exercise #2 — 25 pts  Reading Response #3 — 30 pts
Weeks 6–7

The Brief Is a Hypothesis

The brief as given is a hypothesis, and it is often the wrong one. "Design a faster horse" generates a very different AI output than "help people move between places efficiently" — the choice of brief is a human judgment made before any tool runs. Students apply the reframe protocol to their capstone brief, produce three alternative problem definitions, evaluate each against explicit criteria, and defend a selection in a live three-minute presentation that fields peer objection. The reframe is not a design decision. It is a values decision. The Act One midterm follows: novel brief, no scaffolding, all four pre-ideation capacities demonstrated as practice.

Studio Exercise #3 — 25 pts  Midterm — 100 pts
Act Two — The Human Capacities · Weeks 8–13

Act Two moves through the pipeline stages where AI assists but does not decide. The Dreamer's week — the only week AI owns the session — generates fifty concepts and immediately faces the harder problem: how do you choose among fifty plausible outputs? Curation requires criteria, and criteria require judgment. The Realist builds, but every build begins with identification decisions that are not in the brief. The Critic tests, and the test results answer pragmatic questions while leaving moral and cognitive legitimacy entirely to the human interpreter. Act Two closes with the Pipeline Protocol Checkpoint — the full design process documented to the point of test, with every human judgment call named.

Week 8

One Week for the Dreamer

This is the Dreamer's week — the only week AI owns the session. The Dreamer's job is not to evaluate. It is to generate without premature constraint. Students direct an AI ideation session producing minimum thirty concepts for their capstone reframe, then face the harder problem: curation. Choosing among fifty plausible outputs requires criteria, and criteria require judgment. Why do the surprising concepts survive when the plausible ones shouldn't? Students reduce thirty concepts to three with explicit criteria and a 300-word written defense of each selection.

Studio Exercise #4 — 25 pts  Reading Response #4 — 30 pts
Weeks 9–11

The Realist Builds

AI builds the interface in an afternoon. The identification decisions the human must supply are not in the build brief until the human puts them there: which metric is most motivating, where to surface it in the user journey, what happens when a user sees demoralizing data. These are judgment calls that shape what gets built — and none of them are generated by the tool. Students write a complete build specification that makes every identification decision explicit, then build a low-fidelity prototype using AI acceleration with all identification decisions documented. The prototype answers a specific question. It is not a finished product.

Studio Exercise #5 — 25 pts  Studio Exercise #6 — 25 pts
Weeks 12–13

The Critic Tests

The test results are in. They show the prototype works. The question the data cannot answer: should it? The three legitimacy types — pragmatic (does it work), moral (should it exist), cognitive (can it be trusted) — structure the interpretation. AI achieves pragmatic legitimacy readily. It does not adjudicate the other two. Students conduct testing with minimum five users and produce an interpretive judgment document that addresses all three legitimacy types and names what the data cannot resolve.

Reading Response #5 — 30 pts  Pipeline Protocol Checkpoint — 100 pts
Act Three — The Commit · Weeks 14–15

Act Three is the stage Design Thinking omits — and it was invisible when prototyping was hard. AI has made the omission catastrophic. One week of instruction, one week of presentation.

The five required elements of the Commit document

  1. The decision — a specific course of action, not a recommendation
  2. The evidence — direct citations from the test phase
  3. The uncertainty — what is not known before deployment
  4. The accountability — what the designer is responsible for if it fails
  5. The revision condition — what new information would change the decision
Week 14

The Commit

Students produce a Commit document meeting all five required elements. The peer critique round is mandatory and supervised — it cannot be completed by reading alone, because it requires reviewing a real Commit document with real stakes and delivering written feedback that names specific failures of specificity, evidence, honesty about uncertainty, and accountability.

Studio Exercise #7 (supervised peer critique) — 25 pts  Peer Review Checkpoint — 100 pts
Week 15

Full Pipeline Presentations

The terminal deliverable: the complete AImagineering pipeline presented as a narrative of judgment — not the output, but every human decision from brief to Commit, in sequence, with the metacognitive switches named. The Irreducibly Human section carries half the grade: three specific judgment calls that required values, domain knowledge, or accountability; one judgment call that was tried-as-delegation and then reclaimed; and an honest assessment of the collaboration — where AI was genuinely useful, where it produced confident-sounding noise, and what the student would do differently. The pipeline graduation statement closes it — earned, not recited.

Full Pipeline Presentation — 250 pts

Studio Lab participation (100 pts) is assessed continuously across all 15 weeks. The lowest-scoring Studio Exercise is dropped — 8 of 9 count toward the final grade. Week 7's reframe defense and Week 14's Commit peer critique cannot be made up by reading alone — both sessions depend on the presence of peers whose work you are responding to and who are responding to yours.

Irreducibly Human: What AI Can't Do — AImagineering: The Full Design Pipeline · Graduate seminar · College of Engineering · Northeastern University · 4 credit hours · Prerequisite: Botspeak or equivalent · Instructor: Nik Bear Brown · ni.brown@neu.edu