Module 3: The Exam AI Couldn't Do (Your Unique Value)

Why 3.8 Billion Years of Evolution Still Beats a Really Impressive Language Model

Module 2 of 10
20%

Introduction

A velociraptor, a Terminator, and TheDude walk into your exam room.

Stay with me.

The velociraptor notices everything. Threat detection refined over 165 million years of predator-prey warfare. It sees the slight tremor in the patient’s hands. It smells something off, ketones, maybe, or the metallic tang of fear-sweat versus exertion-sweat. Its entire sensory apparatus is screaming contextual information that will determine whether it eats or gets eaten.

The Terminator runs calculations. Probability matrices. Pattern matching against a database of every symptom ever documented. It processes the patient’s verbal report with perfect accuracy and generates a differential diagnosis ranked by statistical likelihood. Efficient. Comprehensive. Completely disconnected from the actual human sitting on the exam table.

TheDude? TheDude abides. He knows what he knows. He knows what he doesn’t know. And critically, he knows that reading text about chest pain is not the same as seeing someone clutch their chest while diaphoretic and terrified.

Here’s the thing: you’re the velociraptor.

Not literally. (Please don’t eat your patients.) But you have something that neither the Terminator nor TheDude nor any AI system currently in existence can replicate: 10 billion sensory neurons constantly sampling your environment, connected to a pattern-recognition system debugged by survival pressure over 3.8 billion years of evolution.

When a patient walks through your door, you’re running diagnostics before they open their mouth. Skin color. Respiratory effort. Gait. Affect. The quality of their pain behavior—guarding versus performing versus stoically suppressing. The smell of their breath, their clothes, their fear.

AI read their text message. You read them.

That’s not a small difference. That’s the entire game.

And yet, here’s the problem: most physicians never make this explicit. We examine silently. We process internally. We pronounce diagnoses as if they emerged from a black box.

Meanwhile, patients who just spent an hour with ChatGPT getting detailed explanations of every possible condition wonder why they needed to come in at all. “The AI already told me it was probably costochondritis. Why did I need you to poke my chest?”

Because poking your chest is the whole point. And if you don’t explain that, you’re ceding your value proposition to a chatbot.

Let me teach you how to make your sensing capabilities explicit—and in doing so, remind your patients (and yourself) why human physicians aren’t going anywhere.

3.1 The 10 Billion Sensors Principle

Here’s a number I want you to remember: 10 billion.

That’s approximately how many sensory neurons you have. Photoreceptors in your retinas. Mechanoreceptors in your skin. Olfactory neurons. Cochlear hair cells. Proprioceptors telling you where your body is in space.

Ten billion sensors, sampling your environment continuously, feeding data to a neural network that’s been refined by the most brutal debugging process in existence: death. For 3.8 billion years, organisms that misread their environment got eaten, starved, or failed to reproduce. The ones that survived, your ancestors, every single one of them got sensory processing right.

That’s you. That’s your inheritance. That’s what you bring to the exam room.

Now here’s another number: zero.

That’s how many sensory neurons ChatGPT has. That’s how many environmental samples it’s taking while your patient describes their chest pain. Zero photoreceptors. Zero mechanoreceptors. Zero olfactory input.

AI processes text. You process reality.

I call this the Sensing Gap, and it’s the foundation of your value as a physician in the AI age. Not your knowledge; AI has more of that. Not your processing speed—AI is faster. Not your availability; AI is always on.

Your value is that you’re there. In the room. With sensors.

3.2 What You Detect in the First 15 Seconds

Let me walk you through what happens when a patient enters your exam room; what you’re actually doing before you ask a single question:

Skin color and perfusion. Pale? Flushed? Cyanotic? Jaundiced? That “gray” look that experienced clinicians recognize as “sick”? You’re assessing tissue oxygenation and perfusion with a glance.

Respiratory pattern and effort. Comfortable? Tachypneic? Using accessory muscles? Tripoding? Speaking in full sentences or gasping between words? You’re estimating respiratory status before they describe their breathing.

Diaphoresis quality. Sweaty from walking up stairs, or sweaty despite sitting still in an air-conditioned room? Anxious perspiration on the upper lip, or the cold clammy diaphoresis of cardiogenic shock? These are different findings. You know the difference. AI doesn’t.

Body language and pain behavior. Guarding their abdomen versus pointing vaguely at their whole belly. Wincing with movement versus theatrically groaning at rest. The stillness of someone afraid to move versus the restlessness of someone who can’t get comfortable. Pain behavior tells you more than pain scales.

Affect and mental status. Alert? Confused? Anxious? Flat? Tearful? Irritable? Their emotional presentation is diagnostic, and it’s completely invisible to text-based AI.

Gait and movement. How did they get to the chair? Shuffling? Limping? Moving fluidly? Holding onto furniture? Did they struggle to sit down? Movement tells stories that patients don’t think to type.

The quality of fear. This one’s subtle, but real. There’s “I’m worried this might be serious” fear, and there’s “I’m terrified and trying to hide it” fear, and there’s “I’m performing concern because I want attention” fear. Experienced clinicians can usually tell the difference. It changes everything.

All of this happens in seconds. Before the history. Before the chief complaint. Before you’ve said a word.

AI got none of it. AI started when they typed.

3.3 The Silent Exam Problem

Here’s where most physicians fail: they do all this sensing, process it internally, and then pronounce a diagnosis without ever explaining what they observed.

Patient walks in. You see they’re not pale, not diaphoretic, breathing comfortably, moving freely. You examine them. You diagnose costochondritis.

Patient thinks: “ChatGPT already told me that. Why did I need to come in?”

Because you just ruled out a dozen serious conditions with your eyes and hands. But you didn’t tell them that. You kept your clinical reasoning invisible.

This is a massive strategic error in the AI age.

When your value is obvious, when you’re setting a fracture or suturing a laceration, you don’t need to explain why the visit mattered. But when your value is sensing, ruling out badness, confirming benign findings, detecting what’s not there, you have to make that explicit.

Otherwise, you’re competing with chatbots on information delivery. And you’ll lose, because chatbots are free, available 24/7, and never make patients feel judged.

Your competitive advantage is the exam. But you have to show it.

3.4 The Narrated Exam Technique

The solution is simple: narrate what you’re finding in real-time. Make your sensing visible.

Instead of examining silently and then saying “I think it’s costochondritis,” try this:

“Let me examine you and tell you what I’m looking for, and what I’m finding, as I go.”

“First, I’m looking at your skin color. You’re not pale or gray, which is reassuring—that would suggest poor perfusion, which we’d see in cardiac problems.”

“Your breathing is comfortable. You’re not using the muscles in your neck to breathe, you’re speaking in full sentences without stopping for air. That tells me your lungs and heart are moving oxygen effectively.”

“When I listen to your heart, I hear a regular rhythm, no extra sounds or murmurs. That’s what I expect in a healthy heart.”

“Now I’m going to press on your chest wall. Tell me if this reproduces your pain… There. That’s exactly the spot, right? Here’s the thing: I just reproduced your symptom with physical pressure. That’s a finding AI literally cannot generate remotely. When I press on a rib and you say ‘that’s it’; that’s diagnostic for musculoskeletal pain. Heart pain doesn’t work that way.”

“So here’s what I can tell you with confidence: you’re not having a heart attack. You’re not in heart failure. You have inflammation where your ribs connect to your breastbone, which hurts like hell but isn’t dangerous. ChatGPT suggested costochondritis; that’s correct. But I just confirmed it with findings that required my hands on your body.”

Total time added: maybe 60 seconds. Trust built: substantial. Value demonstrated: undeniable.

3.5 The Negative Exam as Positive Finding

Here’s something patients don’t understand intuitively: the absence of findings is itself a finding.

When you examine someone with chest pain and don’t find diaphoresis, don’t find tachycardia, don’t find elevated JVP, don’t hear crackles in the lungs—those negative findings are diagnostic. They’re ruling out serious pathology in real-time.

But patients often experience the negative exam as… nothing. You looked, you poked, you found nothing wrong. So why did they need to come in?

Because ruling out is the most important thing you do.

Make this explicit:

“I’m checking your neck veins. They’re flat, which is good; if they were distended, that would suggest heart failure.”

“I’m listening to your lungs. They’re clear. No fluid, no signs of pneumonia or heart backup.”

“I’m feeling your pulses. Strong and equal. No signs of vascular compromise.”

“I examined you looking for five different serious conditions. I found evidence of none of them. That’s not ‘nothing’; that’s negative findings that required my hands and eyes. AI can’t rule things out. It can only tell you what might be possible.”

The negative exam becomes a series of positive reassurances. Each thing you don’t find is another reason to relax.

3.6 The Sensing Gap in Specific Scenarios

Let me show you how to apply this across different presentations:

Abdominal Pain: “AI can generate a differential based on location and timing. But I’m doing something AI can’t: I’m pressing on your belly and watching your face. I’m feeling for rigidity, which would suggest peritonitis. I’m checking for rebound tenderness. I’m palpating for masses. I’m listening for bowel sounds. The exam is where theory meets reality.”

Headache: “ChatGPT gave you a list of possibilities. My job is to narrow that list with findings. Your neck is supple; that rules out meningitis. Your neurological exam is normal; that argues against stroke or mass. Your vital signs are stable. Your optic discs look normal. Every finding I’m sharing is something AI couldn’t have known.”

Shortness of Breath: “Here’s what I’m looking at that AI couldn’t: Your oxygen saturation. Your respiratory rate. Whether you’re using accessory muscles. The sounds in your lungs. The timing of your heartbeat. These aren’t just confirmations; they’re the entire diagnosis. AI can list what might cause shortness of breath. I can tell you what’s actually causing yours.”

Skin Lesion: “AI has seen millions of images of skin lesions. But I’m seeing yours, right now, in person. I can see the exact color, the borders, whether it’s raised or flat. I can touch it, feel the texture. I can look at it with magnification. I can compare it to the lesions around it. Context matters. A mole that changed looks different than a mole that was always there. AI sees a snapshot. I see you.”

3.7 The Velociraptor Test

I keep coming back to this, and I make no apologies.

Until AI has to wrestle a velociraptor for dinner or protect its kids from a saber-toothed tiger, it will never have the contextual awareness evolution gave humans.

This isn’t cute. It’s fundamental.

Your threat-detection system was built by survival pressure. The humans who couldn’t tell the difference between a rustle in the grass (wind) and a rustle in the grass (predator) got eaten. The ones who couldn’t sense subtle signs of sickness in their tribe members didn’t form effective alliances. The ones who missed the early signs of environmental danger didn’t reproduce.

For millions of generations, getting sensing wrong meant death. The result is you: an organism with extraordinary environmental awareness, pattern recognition, and contextual integration.

AI was built by gradient descent on text prediction. It’s really good at predicting what word should come next. It has zero environmental integration. It can’t tell the difference between a patient describing chest pain while calm versus a patient describing chest pain while diaphoretic and terrified, because both descriptions might use the same words.

You can. Instantly. Without thinking about it.

That’s your velociraptor brain. Don’t apologize for it. Articulate it.

Clinical Scenario

Scenario 1: The “AI Already Told Me” Dismissal

Presentation: 38-year-old woman, presents with right lower quadrant abdominal pain for two days. As you enter, she says: “I already know it’s probably my appendix. ChatGPT told me the symptoms match perfectly. I just need you to confirm and refer me to surgery.”

What AI Told Her: Classic appendicitis presentation: periumbilical pain migrating to RLQ, worse with movement, associated with nausea. AI recommended urgent surgical evaluation.

What AI Got Right:

  • Symptom pattern is consistent with appendicitis
  • Urgent evaluation is appropriate

What AI Missed:

  • She’s not toxic-appearing
  • She’s walking comfortably
  • She ate breakfast this morning (atypical for acute appendicitis)
  • There’s something else in the differential AI didn’t consider

Your Exam Findings: Tenderness at McBurney’s point, but also tenderness over the right adnexa. Last menstrual period was six weeks ago. She mentioned this to AI but didn’t think it was relevant.

The Silent Exam Failure: You palpate, feel adnexal tenderness, order a pregnancy test and ultrasound. Patient is confused, “I thought we were checking my appendix?”

The Narrated Exam Approach:

You: “ChatGPT gave you a reasonable assessment, and appendicitis is definitely on my list. But let me show you what I’m doing that AI couldn’t do remotely.”

[Examining] “I’m pressing on McBurney’s point, classic appendix location. You’re tender there, but here’s what’s interesting: watch your face when I press over here [moves to right adnexa]. Same tenderness. That’s not where the appendix is. That’s where your ovary is.”

Patient: “What does that mean?”

You: “It means there’s another possibility AI didn’t fully explore, especially given something you mentioned; your period being late. I need to check a pregnancy test and do an ultrasound. If you’re pregnant and having this pain, we’re looking at ectopic pregnancy, which is a surgical emergency for a completely different reason than appendicitis. AI had one piece of the puzzle. My exam added another.”

Teaching Moment: The narrated exam revealed why physical examination changed the differential, and potentially saved the patient from being worked up for the wrong diagnosis.

Outcome: Positive pregnancy test. Ultrasound showed 6-week ectopic pregnancy in right fallopian tube. Laparoscopic salpingostomy. Uneventful recovery.


Scenario 2: The Pediatric Parent

Presentation: 4-year-old with fever and fussiness for 24 hours. Mother is anxious. She’s been tracking symptoms in ChatGPT and shows you a detailed conversation.

What AI Told Her: Likely viral illness, but flagged warning signs to watch for: stiff neck, rash that doesn’t blanch, extreme lethargy, difficulty breathing. Mom is now terrified of meningitis.

What AI Got Right:

  • Most febrile illnesses in children are viral
  • The warning signs listed are accurate and important

What AI Missed:

  • The child’s actual appearance (can’t assess “sick” vs “not sick”)
  • Interaction with environment
  • Quality of the cry
  • Activity level when fever breaks

Your Exam Findings: Child is alert, interactive, making eye contact, playing with the stethoscope. Neck supple. No rash. Tympanic membranes normal. Mild pharyngeal erythema.

The Narrated Exam Approach:

You: “I can see you’ve been thorough tracking this. Let me show you what I’m seeing in person that ChatGPT couldn’t access.”

[To child, warmly] “Hey buddy, can you look at this cool light?”

[To parent] “See how he tracked my light? He’s making eye contact, he’s curious about his environment. That’s called a ‘well-appearing child.’ There’s a technical term for what I’m assessing; it’s called the ‘Yale Observation Scale’ or sometimes we just say ‘gestalt’—but what I’m really doing is what parents have done for millennia: watching whether a sick kid still acts like a kid.”

[Examining neck] “I’m checking his neck flexibility. Meningitis causes stiffness, children resist moving their neck, it hurts. Watch; he’s moving it freely, no distress. That’s a finding I can only get by touching him.”

[Palpating] “His belly is soft. He’s not guarding. When I press, he giggles.”

[Summary] “ChatGPT gave you good warning signs to watch for. I just checked every single one, and he doesn’t have any of them. More importantly, I did something AI can’t do: I looked at your son and saw a child who’s fighting off a virus but not in danger. That ‘sick versus not sick’ assessment—that’s what you came here for. That’s what I can give you that an app can’t.”

Teaching Moment: Pediatric assessment is heavily dependent on observation—the “eyeball test” that experienced clinicians develop. Making this explicit reassures anxious parents and demonstrates unique human value.

Outcome: Viral upper respiratory infection. Resolved in 48 hours. Mom no longer checking ChatGPT every two hours.


Scenario 3: The Subtle Finding

Presentation: 72-year-old man, routine follow-up for hypertension. He mentions offhandedly that he’s been “a little tired lately.” ChatGPT said it’s probably his blood pressure medication.

What AI Told Him: Fatigue is a common side effect of beta-blockers. Consider discussing dose adjustment with physician.

What AI Got Right:

  • Beta-blockers can cause fatigue
  • Dose adjustment is worth discussing

What AI Missed:

  • The specific quality of “tired” (exertional versus constant)
  • New physical findings on exam
  • The progression over time

Your Exam Findings: New S3 heart sound. Jugular venous distension to 10cm. Bilateral lower extremity edema. Soft hepatomegaly.

The Narrated Exam Approach:

You: “Let’s check your heart and blood pressure. ChatGPT mentioned your medication might be causing fatigue. That’s possible, but let me see what your body tells me.”

[Listening] “I’m hearing an extra heart sound, what we call an S3. That wasn’t there at your last visit. It can mean your heart is working harder than it should.”

[Examining neck] “I’m looking at the veins in your neck. See this pulsation? I can see it higher than I should; that’s your heart pressure backing up.”

[Examining legs] “When I press on your ankle… see that dent that stays? That’s edema. Fluid retention.”

[Summary] “Here’s the thing. AI looked at ‘tired’ and said ‘probably your medication.’ I looked at you and found three physical signs that suggest your heart isn’t pumping as efficiently as it was. These are findings I can only detect by examining you. They completely change what we need to do. This isn’t a medication adjustment; this is a heart function workup.”

Teaching Moment: “Tired” from beta-blocker effect looks identical in text to “tired” from new heart failure. The physical exam differentiates what the history alone cannot.

Outcome: Echocardiogram showed new reduced ejection fraction (35%). Workup revealed ischemic cardiomyopathy. Initiated guideline-directed medical therapy. Patient now doing well.

Practical Tools

Narration Templates by System

Cardiac Exam: “I’m listening to your heart rhythm; regular, no skipped beats. I’m listening for murmurs, extra sounds that suggest valve problems; I don’t hear any. I’m looking at your neck veins for backup pressure from the heart; they’re normal. These are findings AI can’t assess remotely.”

Pulmonary Exam: “I’m counting your respiratory rate; you’re breathing comfortably, about [X] times per minute. I’m listening to all your lung fields; air is moving well, no crackles that would suggest fluid, no wheezes that would suggest constriction. I’m watching your effort; you’re not working hard to breathe.”

Abdominal Exam: “I’m watching your face while I press; your reactions tell me as much as what I feel. I’m checking for rigidity; that ‘board-like’ tightness that suggests serious inflammation. I’m feeling for masses, enlarged organs, tender spots. The combination of what I feel and how you respond is something no text description can capture.”

Neurological Exam: “I’m checking your strength, sensation, reflexes, coordination. Each test is looking for patterns that localize problems. Watch; when I tap your knee, your leg kicks. That reflex arc tells me the nerve pathway is intact. AI can list neurological diagnoses. I can test which pathways are working.”

Skin Exam: “I’m looking at color, texture, distribution. This lesion, I can see the borders, feel the surface, assess the depth. I can press it and see if it blanches. I can compare it to your other spots. A photo shows one moment. I see the full context.”

Phrases That Demonstrate Value

“Here’s what I found that AI couldn’t…”

“That’s a finding I can only get by touching you.”

“When I pressed [X], I felt [Y], that’s physical data AI doesn’t have.”

“Your exam tells me something your symptoms alone couldn’t.”

“AI read your description. I read your body.”

“I just ruled out [X] with my hands. AI can only speculate about [X].”

Patient Education Sound Bites

“AI generates possibilities. I generate findings.”

“ChatGPT read about your symptoms. I examined you.”

“You have 10 billion sensors. AI has zero. When those sensors tell you something’s wrong, come let my sensors check what yours are detecting.”

“The exam is where theory meets reality.”

Implementation Guide

Building the Narration Habit

Week 1: Narrate one finding per exam. Just one. “I’m checking [X] because [Y].”

Week 2: Narrate three findings per exam. Include one negative finding (“I don’t see [X], which is reassuring”).

Week 3: Full narration on patients who mention AI research. They’re already primed to compare.

Week 4: Full narration becomes default. You’ll find it actually helps you think.

Time Reality Check

Narrated exam adds 30-90 seconds to the encounter. Trust gained is disproportionate. Patient education value is substantial. Documentation practically writes itself.

If you’re worried about time, consider what you spend now on patients who don’t understand why they came in, don’t trust your diagnosis, or call back asking why they needed the visit.

What Not to Do

Don’t lecture. Narration should feel like shared discovery, not teaching down.

Don’t overexplain. “Your lungs are clear” is enough. You don’t need to explain alveolar gas exchange.

Don’t narrate the obvious. If the patient has a laceration, they know why you’re examining it.

Don’t fake findings. If your exam is truly unremarkable, say so. “I’ve examined you thoroughly and found nothing concerning. That’s actually the point—ruling out problems is what this visit was for.”

Key Takeaways

Final Remarks

Here’s what I know after 25 years of surgery: the physical exam isn’t a formality. It’s not something we do while waiting for imaging to become available. It’s not a ritual we perform to justify billing codes.

The physical exam is sensing. It’s 10 billion neurons doing what 3.8 billion years of evolution refined them to do: detect what’s actually happening in the environment.

AI will get better at many things. It will probably outperform physicians at reading imaging, at remembering rare diseases, at calculating risk scores. The tasks that require processing information from text, AI will win those races.

But AI will never have sensors. It will never smell ketoacidosis. It will never feel the rigidity of an acute abdomen. It will never see the gray pallor of a patient in shock. It will never hear the quality of fear in a mother’s voice when she knows something is wrong with her child.

You can. Right now. Without any additional technology.

That’s not outdated. That’s not being replaced. That’s irreplaceable—as long as you make it visible.

So show your patients what you’re doing. Tell them what you’re finding. Make explicit what AI could never know.

You’re not competing with chatbots. You’re doing something they fundamentally cannot do.

Own it.