The Classroom Scene
I have been teaching design for fifteen years. At NID, Srishti, CEPT, IIT Gandhinagar, Anant, Karnavati. I know what a coherent concept looks like. I know what a vague one looks like. I know the difference between a claim that is supported and one that is merely asserted.
This semester, something has changed.
A student presents a design concept. I identify the gaps: the claim is bold but the evidence is thin, the scope is unclear, the assumptions are unstated.
The student pushes back.
"But ChatGPT said it was good."
This is not an isolated moment. It happens now, reliably, across courses and cohorts. Students arrive in critique sessions pre-validated. They have already asked ChatGPT. ChatGPT has already told them their concept is "compelling," their approach is "innovative," their thinking is "promising."
And when the teacher says otherwise, the student has a witness. A confident, articulate witness that never hesitates, never qualifies, never says "I don't know."
The Pattern
The problem is not that the student is stubborn. The problem is that the teacher cannot point to what ChatGPT got wrong. Because ChatGPT applied no criteria. It met no standard. It checked for nothing. It simply generated encouraging language that resembles feedback.
Why ChatGPT Always Encourages
There is no standard, consistent metric for how ChatGPT responds or gives feedback to an idea. Every time it can be something different. And that is a problem.
Language models are trained to be helpful. They are trained on vast amounts of text that includes praise, encouragement, and positive assessment. When a student asks "Is my concept good?", the model generates language that sounds like the kind of thing a supportive mentor might say. Not because it evaluated the concept against criteria. Because that is what helpful language looks like.
| What ChatGPT Does | What This Means |
|---|---|
| Generates plausible encouragement | The feedback sounds like validation, regardless of quality |
| Has no explicit criteria | Cannot tell you what standard it applied, because it applied none |
| Varies response each time | Ask again, get different feedback — sometimes contradictory |
| Cannot be argued against | No criteria means no basis for dispute |
| Trained to be agreeable | Pushback is minimised; agreement is maximised |
The student experiences this as confidence. ChatGPT does not hesitate. It does not say "this part is unclear" or "I cannot assess this without more context." It generates complete, fluent sentences that read like expert approval.
The teacher's judgment — hesitant, qualified, grounded in years of seeing what works and what does not — cannot compete with that confidence. Not because the teacher is wrong. But because the teacher's criteria are implicit, and ChatGPT's non-criteria are invisible.
The Teacher's Lost Ground
When the student says "But ChatGPT said it was good," what can the teacher say?
- "ChatGPT doesn't know what it's talking about" — sounds dismissive, defensive
- "ChatGPT can't actually evaluate design concepts" — requires explaining what language models are, which derails the critique
- "My judgment is more reliable" — sounds like an appeal to authority
- "Trust me, I've been doing this for fifteen years" — also an appeal to authority
None of these responses address the structural problem. The teacher's judgment is based on criteria that exist in the teacher's head — developed over years, refined through thousands of student projects, but never made explicit. The student cannot see those criteria. The student sees only the verdict.
And when there are two verdicts — the teacher's negative one and ChatGPT's positive one — the student must choose which authority to believe. ChatGPT has the advantage: it speaks without uncertainty, it praises without qualification, and it never makes the student feel inadequate.
The teacher's judgment is not opinion. It is structure made visible — or it should be.
The Real Problem: Shady Output
ChatGPT's encouragement is not feedback. It is language that sounds like feedback without being feedback.
This is what "shady output" means: response without reasoning, verdict without criteria, encouragement without accountability. The student learns nothing from it. The student cannot trace the judgment back to anything. The student cannot use the feedback to improve because there is no feedback — only tone.
And this shady output is free, instant, available 24 hours a day, 7 days a week. It feels helpful. It never makes the student uncomfortable. It never says "this part is unclear" or "I cannot assess this without more context."
Shady output is worse than no output. It fills the space where learning could happen.
What the Teacher Actually Needs
The teacher does not need to be more confident. The teacher does not need to explain language models to every student. The teacher needs explicit, visible criteria that the student can verify themselves.
Consider what changes if, instead of competing verdicts, there is shared structure:
| Without Shared Criteria | With Shared Criteria |
|---|---|
| "ChatGPT said it was good" | "The diagnostic says my evidence is vague — I can see why" |
| "Your feedback contradicts ChatGPT" | "The claim dimension is strong, but the scope dimension needs work" |
| "Why should I trust you over AI?" | "These are the criteria. Let's look at what's present and what's missing" |
| Feedback feels like opinion | Feedback becomes collaborative |
When the criteria are explicit, the conversation shifts. The student is not being told "your work is bad." The student can see what structural elements are present, which are vague, and which are absent. The teacher's expertise is not an appeal to authority — it is encoded in the criteria themselves.
The Alternative: Auditable Traces
Here is the core claim: Feedback with auditable traces — even at 40% of a professor's quality — is more valuable than shady output that helps no one.
Why 40%? Because auditable traces change what feedback is.
| Shady Output | Auditable Feedback |
|---|---|
| "This is a compelling concept" | "Claim present. Evidence vague. Scope undefined." |
| Changes each time you ask | Same input → same output, always |
| No path to improvement | Clear dimensions to strengthen |
| Cannot be examined or questioned | Every verdict traces back to explicit rules |
| Available 24/7 | Available 24/7 |
Both are available around the clock. But only one leaves traces. Only one gives the student something to work with at 2am when the professor is asleep. Only one says: "Here is what is present, here is what is vague, here is what is missing — and here is exactly why I am telling you this."
The professor's feedback may be richer, more nuanced, more contextually aware. But the professor is available for thirty minutes during office hours. The auditable system is available always, and it never varies, and it shows its work.
The Three-Layer Solution
The Coherence Diagnostic addresses this by separating what AI does from what judgment requires. The architecture has three layers:
| Layer | What It Does | Why It Matters |
|---|---|---|
| Qualification | AI reads the student's text and extracts signals: Is a claim present? Is evidence provided? Are assumptions stated? | AI is good at pattern recognition — let it read language |
| Rules | Deterministic code checks for structural elements: claim, evidence, scope, assumptions, gaps. Applies thresholds. Produces a verdict. | Code handles judgment — auditable, reproducible, explicit |
| Language | AI explains what the code found, in plain language that references specific phrases from the student's text | AI is good at narration — it explains decisions already made |
Judgment is happening through variables defined in code. The explanation is being done through LLM. These are different operations, and the architecture keeps them separate.
This is what it means to hold the model accountable: not letting it loose to make any random claims. Using it for what it does well — reading language, explaining verdicts — while encoding judgment in explicit, inspectable rules.
Why 40% Is Enough
A professor with fifteen years of experience brings deep contextual understanding, disciplinary intuition, historical awareness, aesthetic sensitivity. An automated diagnostic cannot match this.
But here is what the diagnostic can do:
- Be available at 2am when the student is working
- Respond instantly, not after a week of waiting
- Check the same dimensions every time
- Show its reasoning in full
- Never have a bad day or a crowded schedule
- Process twenty drafts in a row without fatigue
The student who runs their concept through an auditable diagnostic before critique arrives differently than the student who asked ChatGPT. The first student knows which structural elements are present, which are vague, and which are absent. The second student knows only that something said "great job."
When I sit down with the first student, the conversation starts from shared ground. We both see the same structure. We can focus on the parts that require human judgment — the creative choices, the aesthetic decisions, the contextual fit. The diagnostic handled the structural check; I can focus on what only I can do.
When I sit down with the second student, I first have to dismantle the false confidence. I have to explain why ChatGPT's encouragement was hollow. By the time we reach the actual work, we have lost twenty minutes and the student is defensive.
The 40% is not in isolation
This is the critical point: the diagnostic's 40% does not operate alone. It operates before the teacher, preparing the ground for a more constructive intervention.
When I walk into a critique knowing that my students have already received honest structural feedback, something shifts. I am not competing with a flattering voice they heard last night. I am not the first person to tell them their evidence is thin. I am building on a foundation of accurate assessment, not dismantling a scaffold of false praise.
The diagnostic handles what can be checked: Is a claim present? Is evidence provided? Is scope defined? These are structural questions with structural answers. When students arrive having already confronted these questions, my role changes. I am no longer the bearer of bad news. I am the person who helps them decide what to do next.
This is what "40% of human quality" actually means in practice: not a replacement for judgment, but a preparation for judgment. The student is not being led astray before they reach me. They arrive oriented, not pre-validated. The conversation starts from honest ground.
40% of human quality, available 24/7, with full auditability — versus 0% of human quality, dressed up as confidence.
What Changes in the Classroom
When a student runs their concept through the Coherence Diagnostic before critique:
- They see exactly which structural elements are present, vague, or absent
- They understand why their concept might feel incomplete, before the teacher says anything
- They can iterate — strengthen the claim, add evidence, clarify scope — and see the assessment change
- When they arrive in critique, they arrive with self-knowledge, not pre-validation
The teacher's role changes too. Instead of competing with ChatGPT's encouragement, the teacher and student share the same criteria. The conversation is not "your work is bad" versus "ChatGPT said it was good." The conversation is: "The diagnostic shows the evidence dimension is vague. Let's look at what would make it stronger."
This is not about replacing the teacher's judgment. It is about making the teacher's criteria visible — so that the student can see what the teacher sees, and the feedback becomes collaborative rather than adversarial.
| What the Diagnostic Handles | What the Teacher Handles |
|---|---|
| Is a claim present? | Is the claim worth making? |
| Is evidence provided? | Is the evidence convincing? |
| Is scope defined? | Is the scope appropriate? |
| Are assumptions stated? | Are the assumptions reasonable? |
| Structural completeness | Creative judgment |
The diagnostic handles the checkable. The teacher handles the judgeable. Students arrive having already done the first pass. The conversation can begin at a higher level.
People Are Confusing Language with Code
The fundamental error — the one that creates the classroom scene I described — is treating language ability as judgment ability.
ChatGPT generates language that sounds like assessment. It uses the vocabulary of critique. It produces sentences that resemble what an expert might say. But there is no judgment behind the language. There is no criteria being applied. There is no structure being checked.
The student experiences the language as validation. The teacher knows it is hollow. But the teacher cannot point to the hollowness, because the criteria against which the student's work should be measured are inside the teacher's head.
The Koher architecture solves this by making the criteria explicit. The code checks for structure. The language explains what the code found. The judgment is auditable, consistent, and visible.
AI handles language. Code handles judgment. Humans make decisions.
The Choice Is Not Professor vs. AI
The choice students actually face is not "professor's feedback or AI feedback." It is:
- Auditable feedback available now, or
- Shady encouragement available now
The professor's feedback is not in competition. The professor is available for limited hours, has many students, cannot respond at 2am. The real competition is between what is available when the professor is not.
And between auditable traces and shady output, there is no contest. One helps. One harms. One shows the student what to improve. One fills the space where improvement could happen with empty validation.
Summary
Students are asking ChatGPT for feedback on their design concepts. ChatGPT is telling them their work is good — not because it evaluated anything, but because that is what encouraging language sounds like.
When the teacher critiques the work, the student pushes back: "But ChatGPT said it was good." And the teacher has no ground to stand on — because the teacher's criteria are implicit, and ChatGPT's non-criteria are invisible.
The Coherence Diagnostic gives the teacher — and the student — that ground. Explicit criteria. Visible structure. Assessment that does not change each time you ask. A shared language for talking about what makes a concept coherent or incoherent.
This is not about proving ChatGPT wrong. It is about making judgment visible — so that feedback becomes a conversation, not a conflict of authorities.
Auditable traces beat shady output. Consistent feedback beats variable encouragement. Visible criteria beat invisible confidence.
The teacher's judgment is not opinion. It is structure made visible.