• My Feed
  • Home
  • What's Important
  • Media & Entertainment
Search

Stay Curious. Stay Wanture.

© 2026 Wanture. All rights reserved.

  • Terms of Use
  • Privacy Policy
Science/Mind
AI can't read the room — and that's a problem

New research reveals why even advanced AI fails at understanding human social dynamics

7 November 2025

—

Explainer *

Adrian Vega
banner

A groundbreaking 2025 study from Johns Hopkins University tested over 350 AI models against human perception of social interaction. The result: no AI could match how people instantly interpret collaboration, competition, or social cues. This limitation affects autonomous vehicles, delivery robots, and any technology navigating human spaces — revealing a fundamental gap between seeing and understanding.

telegram-cloud-photo-size-2-5211058007943351500-y

Summary:

  • Johns Hopkins study reveals AI struggles to interpret human social interactions in 3-second video tests
  • Current AI models cannot match human ability to read subtle social dynamics and collaborative behaviors
  • Research highlights critical limitations for autonomous technologies like self-driving cars and service robots

Two people glance at each other across a crowded room. In milliseconds, you know they're collaborating—not competing, not strangers, not waiting. An AI watching the same scene? It's still guessing.

That gap—between human intuition and machine interpretation—is wider than we thought. A study published at the International Conference on Learning Representations (ICLR) in April 2025 reveals that even the most advanced AI models struggle to interpret the social dynamics humans read effortlessly.

The research, led by scientists at Johns Hopkins University, tested over 350 large language models and generative AI systems against human perception. The result: no AI model could adequately match how people understand and respond to social behavior in real time.

This isn't just an academic curiosity. It's a fundamental limitation with real-world stakes—for autonomous vehicles navigating pedestrian crossings, delivery robots interpreting when someone holds a door open, and any technology that must move safely through human spaces.

What Social Interaction Actually Involves

Before we understand what AI can't do, we need to clarify what humans do without thinking.

Social interaction isn't just seeing people move. It's reading body language, interpreting context, predicting intentions, and sensing collaboration or conflict in a glance. When two people assemble furniture together, you instantly recognize coordination. When they work on separate tasks in the same room, you know they're coexisting, not cooperating.

These judgments happen in fractions of a second. They rely on pattern recognition, contextual memory, and emotional inference—cognitive processes woven so deeply into perception that we barely notice them.

AI, by contrast, sees pixels and patterns. It lacks the lived experience that teaches humans what collaboration looks like versus competition, what hesitation means versus confidence.

How Scientists Tested AI Against Human Perception

The Johns Hopkins team designed an experiment to measure this gap precisely.

Researchers Kathy Garcia, Emalie McMahon, Colin Conwell, Michael F. Bonner, and Leyla Isik led the study.

The Three-Second Video Experiment

Participants watched 250 short video clips—each just three seconds long—drawn from the Moments in Time dataset. In these clips, people performed tasks together or independently, demonstrating different aspects of social interaction.

After watching, participants rated characteristics important for understanding social dynamics on a scale from 1 to 5. Questions included: Are these people working together? Is this interaction cooperative or independent? What is the social relationship here?

What AI Models Were Asked to Do

Researchers fed the same videos to over 350 AI systems—including large language models (AI systems trained on vast text to predict and generate human-like responses) and generative AI models (systems that create new content based on patterns).

The models were asked to predict how humans would rate the videos. Additionally, language models evaluated short captions written by humans describing the social interactions.

To deepen the comparison, the team also collected fMRI brain response data from four participants, measuring neural activity in regions associated with social cognition—specifically, lateral-stream brain responses, which process social information.

Why AI Struggles With Social Dynamics

The results were clear: AI models could not reliably predict human judgments about social behavior.

Language models performed relatively well at predicting human ratings when given text captions. Video models showed some ability to predict brain responses in certain regions. But no single model excelled at both behavioral judgments and social brain activity.

Think of it like reading sheet music versus feeling rhythm. AI sees the notes but misses the beat that makes humans move together.

The researchers concluded that current AI architecture lacks a fundamental aspect that allows the human brain to interpret dynamic social interaction quickly and accurately. That missing piece isn't just more data or better algorithms—it's something closer to lived understanding, the kind that comes from being a social creature navigating a social world.

What This Means for Autonomous Technology

This limitation isn't abstract. It has immediate implications for technologies already entering public spaces.

Self-Driving Cars and Social Navigation

Autonomous vehicles rely on AI to interpret pedestrian behavior. A person making eye contact at a crosswalk signals intent to cross. A group hesitating on the curb suggests uncertainty. These cues—invisible to current AI—are critical for safe navigation.

If an AI can't distinguish collaboration from coexistence in a three-second video, how reliably can it interpret the social choreography of a busy intersection?

Assistant Robots in Human Spaces

Delivery robots, warehouse assistants, and service machines must navigate environments filled with people. They need to recognize when someone is blocking a path intentionally versus accidentally, when a gesture means "go ahead" versus "wait."

Without the ability to read social dynamics, these systems risk awkward interactions at best—and safety failures at worst.

The Missing Piece in AI Architecture

What exactly do humans possess that AI lacks?

The Johns Hopkins researchers point to something deeper than pattern recognition. Humans don't just process visual information—they interpret it through layers of social experience, emotional context, and predictive modeling built over a lifetime of interaction.

AI models, even those trained on billions of images and videos, lack this embodied knowledge. They can identify objects, track motion, and classify actions. But they can't feel the difference between a tense silence and a comfortable one, between cooperation and competition, between invitation and dismissal.

That gap—between seeing and understanding—is where current AI architecture falls short.

What Comes Next for AI Development

The research team made their findings publicly available, inviting other researchers to build on their work.

They shared code, captions, behavioral data, and fMRI data through the Open Science Framework.

In a follow-up study posted in October 2025, Garcia and Isik introduced a human-similarity benchmark with approximately 49,000 odd-one-out judgments. They also developed a method to fine-tune video models to better align with human social judgments.

These steps suggest a path forward: not just training AI on more data, but training it to recognize the patterns that matter most to human social cognition.

The question isn't whether AI will learn to read social cues—it's how researchers will teach machines something the human brain does without thinking. Until then, the room remains unreadable to the algorithm watching from the corner.

What is this about?

  • Explainer */
  • Adrian Vega/
  • Science/
  • Mind

Feed

    Google adds Gmail mobile encryption for Enterprise Plus

    Google adds Gmail mobile encryption for Enterprise Plus

    Mobile Gmail now provides end-to-end encryption, dropping third-party tools

    about 5 hours ago
    Microsoft removes Copilot disclaimer on April 10, 2026

    Microsoft removes Copilot disclaimer on April 10, 2026

    2025 Nadella interview frames the removal as a push to make Copilot a tool

    about 5 hours ago
    Artemis-2 Returns: Orion Splashdown at 3:00 a.m. PT

    Artemis-2 Returns: Orion Splashdown at 3:00 a.m. PT

    Four astronauts end a nine‑day, 406,765 km lunar arc—Moon flight since Apollo 17

    about 5 hours ago
    Button AI Assistant Debuts, Offering Screen‑Free Voice Help

    Button AI Assistant Debuts, Offering Screen‑Free Voice Help

    Nostalgic iPod Shuffle design meets privacy‑first press‑to‑talk AI

    1 day ago
    Razer Hammerhead V3 HyperSpeed Debuts with Dual‑Mode Case

    Razer Hammerhead V3 HyperSpeed Debuts with Dual‑Mode Case

    The USB‑C case also serves as a 2.4 GHz receiver, cutting dongles for PS5 and phones

    1 day ago
    Apple ships 6.2 million Macs Q1 2026, M5‑MacBook Pro leads

    Apple ships 6.2 million Macs Q1 2026, M5‑MacBook Pro leads

    Apple’s share rises to 9.5%, moving it into fourth place among global PC makers

    1 day ago
    Galaxy S22 Ultra can be bricked after factory reset

    Galaxy S22 Ultra can be bricked after factory reset

    US owners report IMEI‑level lock that hands control to unknown administrator Numero LLC

    1 day ago
    Mouse: P.I. for Hire arrives April 16 on PC, PS5, and Xbox

    Mouse: P.I. for Hire arrives April 16 on PC, PS5, and Xbox

    Modes: 4K 60 fps quality or 120 fps performance on PS5 and Xbox Series X

    1 day ago
    YouTube Rolls Out Auto Speed for Premium Users

    YouTube Rolls Out Auto Speed for Premium Users

    The AI‑driven playback boost aims to cut dead air on long videos

    2 days ago
    Blackwell Set to Capture Majority of the 2026 GPU Market

    Blackwell Set to Capture Majority of the 2026 GPU Market

    GB300/B300 GPUs Push Blackwell to 71% of Shipments; Rubin Falls to 22%

    2 days ago
    Google launches AI avatar tool for Shorts on April 9, 2026

    Google launches AI avatar tool for Shorts on April 9, 2026

    Ages 18+ can create digital replicas, with Synth ID tags and a 3‑year auto‑delete

    2 days ago
    Mac OS X 10.0 Cheetah runs on Wii

    Mac OS X 10.0 Cheetah runs on Wii

    Ports Mac OS X 10.0 Cheetah to the Wii, showing the PowerPC 750CL can run an OS

    2 days ago
    DuoBell Beats ANC: Safer Cycling with Apple AirPods Max

    DuoBell Beats ANC: Safer Cycling with Apple AirPods Max

    A 750 Hz blind‑spot lets DuoBell cut through ANC on popular headphones

    2 days ago
    Škoda DuoBell prototype unveiled on April 5, 2026

    Škoda DuoBell prototype unveiled on April 5, 2026

    750 Hz pulse and 2,000 Hz chime cut through ANC, alerting riders faster at 15 mph

    2 days ago
    SteamGPT Leak Reveals Dual‑Role AI on Steam

    SteamGPT Leak Reveals Dual‑Role AI on Steam

    Leak shows AI handling support and cheat‑detection for millions on the platform

    2 days ago
    Oppo Pad mini challenges Apple with Snapdragon 8 Gen 5

    Oppo Pad mini challenges Apple with Snapdragon 8 Gen 5

    April 21: Oppo Pad mini 8.8‑inch, Snapdragon 8 Gen 5, 5.39 mm, 279 g, 144 Hz OLED

    2 days ago
    Apple to ship 3 million foldable iPhones by end‑2026

    Apple to ship 3 million foldable iPhones by end‑2026

    Limited rollout equals 12 % of iPhone volume and rivals Samsung’s 2.4 million Galaxy Z Fold 7 sales

    2 days ago
    Apple unveils iPhone 18 Pro, iPhone 18 Pro Max, and iPhone Ultra

    Apple unveils iPhone 18 Pro, iPhone 18 Pro Max, and iPhone Ultra

    Mockups match leaked renders; 20 million Samsung panels for iPhone Ultra

    3 days ago
    Sony launches Playerbase program for Gran Turismo 7

    Sony launches Playerbase program for Gran Turismo 7

    PlayStation gamers can win a flight, facial scan, and an avatar in Gran Turismo 7

    3 days ago
    Claude Mythos Preview Beats Opus 4.6 in Cybersecurity!

    Claude Mythos Preview Beats Opus 4.6 in Cybersecurity!

    Claude Mythos Preview for five partners—pricing after a 100 million token credit

    3 days ago
    Loading...
Science/Mind

AI can't read the room — and that's a problem

New research reveals why even advanced AI fails at understanding human social dynamics

November 7, 2025, 6:14 pm

A groundbreaking 2025 study from Johns Hopkins University tested over 350 AI models against human perception of social interaction. The result: no AI could match how people instantly interpret collaboration, competition, or social cues. This limitation affects autonomous vehicles, delivery robots, and any technology navigating human spaces — revealing a fundamental gap between seeing and understanding.

telegram-cloud-photo-size-2-5211058007943351500-y

Summary

  • Johns Hopkins study reveals AI struggles to interpret human social interactions in 3-second video tests
  • Current AI models cannot match human ability to read subtle social dynamics and collaborative behaviors
  • Research highlights critical limitations for autonomous technologies like self-driving cars and service robots

Two people glance at each other across a crowded room. In milliseconds, you know they're collaborating—not competing, not strangers, not waiting. An AI watching the same scene? It's still guessing.

That gap—between human intuition and machine interpretation—is wider than we thought. A study published at the International Conference on Learning Representations (ICLR) in April 2025 reveals that even the most advanced AI models struggle to interpret the social dynamics humans read effortlessly.

The research, led by scientists at Johns Hopkins University, tested over 350 large language models and generative AI systems against human perception. The result: no AI model could adequately match how people understand and respond to social behavior in real time.

This isn't just an academic curiosity. It's a fundamental limitation with real-world stakes—for autonomous vehicles navigating pedestrian crossings, delivery robots interpreting when someone holds a door open, and any technology that must move safely through human spaces.

What Social Interaction Actually Involves

Before we understand what AI can't do, we need to clarify what humans do without thinking.

Social interaction isn't just seeing people move. It's reading body language, interpreting context, predicting intentions, and sensing collaboration or conflict in a glance. When two people assemble furniture together, you instantly recognize coordination. When they work on separate tasks in the same room, you know they're coexisting, not cooperating.

These judgments happen in fractions of a second. They rely on pattern recognition, contextual memory, and emotional inference—cognitive processes woven so deeply into perception that we barely notice them.

AI, by contrast, sees pixels and patterns. It lacks the lived experience that teaches humans what collaboration looks like versus competition, what hesitation means versus confidence.

How Scientists Tested AI Against Human Perception

The Johns Hopkins team designed an experiment to measure this gap precisely.

Researchers Kathy Garcia, Emalie McMahon, Colin Conwell, Michael F. Bonner, and Leyla Isik led the study.

The Three-Second Video Experiment

Participants watched 250 short video clips—each just three seconds long—drawn from the Moments in Time dataset. In these clips, people performed tasks together or independently, demonstrating different aspects of social interaction.

After watching, participants rated characteristics important for understanding social dynamics on a scale from 1 to 5. Questions included: Are these people working together? Is this interaction cooperative or independent? What is the social relationship here?

What AI Models Were Asked to Do

Researchers fed the same videos to over 350 AI systems—including large language models (AI systems trained on vast text to predict and generate human-like responses) and generative AI models (systems that create new content based on patterns).

The models were asked to predict how humans would rate the videos. Additionally, language models evaluated short captions written by humans describing the social interactions.

To deepen the comparison, the team also collected fMRI brain response data from four participants, measuring neural activity in regions associated with social cognition—specifically, lateral-stream brain responses, which process social information.

Why AI Struggles With Social Dynamics

The results were clear: AI models could not reliably predict human judgments about social behavior.

Language models performed relatively well at predicting human ratings when given text captions. Video models showed some ability to predict brain responses in certain regions. But no single model excelled at both behavioral judgments and social brain activity.

Think of it like reading sheet music versus feeling rhythm. AI sees the notes but misses the beat that makes humans move together.

The researchers concluded that current AI architecture lacks a fundamental aspect that allows the human brain to interpret dynamic social interaction quickly and accurately. That missing piece isn't just more data or better algorithms—it's something closer to lived understanding, the kind that comes from being a social creature navigating a social world.

What This Means for Autonomous Technology

This limitation isn't abstract. It has immediate implications for technologies already entering public spaces.

Self-Driving Cars and Social Navigation

Autonomous vehicles rely on AI to interpret pedestrian behavior. A person making eye contact at a crosswalk signals intent to cross. A group hesitating on the curb suggests uncertainty. These cues—invisible to current AI—are critical for safe navigation.

If an AI can't distinguish collaboration from coexistence in a three-second video, how reliably can it interpret the social choreography of a busy intersection?

Assistant Robots in Human Spaces

Delivery robots, warehouse assistants, and service machines must navigate environments filled with people. They need to recognize when someone is blocking a path intentionally versus accidentally, when a gesture means "go ahead" versus "wait."

Without the ability to read social dynamics, these systems risk awkward interactions at best—and safety failures at worst.

The Missing Piece in AI Architecture

What exactly do humans possess that AI lacks?

The Johns Hopkins researchers point to something deeper than pattern recognition. Humans don't just process visual information—they interpret it through layers of social experience, emotional context, and predictive modeling built over a lifetime of interaction.

AI models, even those trained on billions of images and videos, lack this embodied knowledge. They can identify objects, track motion, and classify actions. But they can't feel the difference between a tense silence and a comfortable one, between cooperation and competition, between invitation and dismissal.

That gap—between seeing and understanding—is where current AI architecture falls short.

What Comes Next for AI Development

The research team made their findings publicly available, inviting other researchers to build on their work.

They shared code, captions, behavioral data, and fMRI data through the Open Science Framework.

In a follow-up study posted in October 2025, Garcia and Isik introduced a human-similarity benchmark with approximately 49,000 odd-one-out judgments. They also developed a method to fine-tune video models to better align with human social judgments.

These steps suggest a path forward: not just training AI on more data, but training it to recognize the patterns that matter most to human social cognition.

The question isn't whether AI will learn to read social cues—it's how researchers will teach machines something the human brain does without thinking. Until then, the room remains unreadable to the algorithm watching from the corner.

What is this about?

  • Explainer */
  • Adrian Vega/
  • Science/
  • Mind

Feed

    Google adds Gmail mobile encryption for Enterprise Plus

    Google adds Gmail mobile encryption for Enterprise Plus

    Mobile Gmail now provides end-to-end encryption, dropping third-party tools

    about 5 hours ago
    Microsoft removes Copilot disclaimer on April 10, 2026

    Microsoft removes Copilot disclaimer on April 10, 2026

    2025 Nadella interview frames the removal as a push to make Copilot a tool

    about 5 hours ago
    Artemis-2 Returns: Orion Splashdown at 3:00 a.m. PT

    Artemis-2 Returns: Orion Splashdown at 3:00 a.m. PT

    Four astronauts end a nine‑day, 406,765 km lunar arc—Moon flight since Apollo 17

    about 5 hours ago
    Button AI Assistant Debuts, Offering Screen‑Free Voice Help

    Button AI Assistant Debuts, Offering Screen‑Free Voice Help

    Nostalgic iPod Shuffle design meets privacy‑first press‑to‑talk AI

    1 day ago
    Razer Hammerhead V3 HyperSpeed Debuts with Dual‑Mode Case

    Razer Hammerhead V3 HyperSpeed Debuts with Dual‑Mode Case

    The USB‑C case also serves as a 2.4 GHz receiver, cutting dongles for PS5 and phones

    1 day ago
    Apple ships 6.2 million Macs Q1 2026, M5‑MacBook Pro leads

    Apple ships 6.2 million Macs Q1 2026, M5‑MacBook Pro leads

    Apple’s share rises to 9.5%, moving it into fourth place among global PC makers

    1 day ago
    Galaxy S22 Ultra can be bricked after factory reset

    Galaxy S22 Ultra can be bricked after factory reset

    US owners report IMEI‑level lock that hands control to unknown administrator Numero LLC

    1 day ago
    Mouse: P.I. for Hire arrives April 16 on PC, PS5, and Xbox

    Mouse: P.I. for Hire arrives April 16 on PC, PS5, and Xbox

    Modes: 4K 60 fps quality or 120 fps performance on PS5 and Xbox Series X

    1 day ago
    YouTube Rolls Out Auto Speed for Premium Users

    YouTube Rolls Out Auto Speed for Premium Users

    The AI‑driven playback boost aims to cut dead air on long videos

    2 days ago
    Blackwell Set to Capture Majority of the 2026 GPU Market

    Blackwell Set to Capture Majority of the 2026 GPU Market

    GB300/B300 GPUs Push Blackwell to 71% of Shipments; Rubin Falls to 22%

    2 days ago
    Google launches AI avatar tool for Shorts on April 9, 2026

    Google launches AI avatar tool for Shorts on April 9, 2026

    Ages 18+ can create digital replicas, with Synth ID tags and a 3‑year auto‑delete

    2 days ago
    Mac OS X 10.0 Cheetah runs on Wii

    Mac OS X 10.0 Cheetah runs on Wii

    Ports Mac OS X 10.0 Cheetah to the Wii, showing the PowerPC 750CL can run an OS

    2 days ago
    DuoBell Beats ANC: Safer Cycling with Apple AirPods Max

    DuoBell Beats ANC: Safer Cycling with Apple AirPods Max

    A 750 Hz blind‑spot lets DuoBell cut through ANC on popular headphones

    2 days ago
    Škoda DuoBell prototype unveiled on April 5, 2026

    Škoda DuoBell prototype unveiled on April 5, 2026

    750 Hz pulse and 2,000 Hz chime cut through ANC, alerting riders faster at 15 mph

    2 days ago
    SteamGPT Leak Reveals Dual‑Role AI on Steam

    SteamGPT Leak Reveals Dual‑Role AI on Steam

    Leak shows AI handling support and cheat‑detection for millions on the platform

    2 days ago
    Oppo Pad mini challenges Apple with Snapdragon 8 Gen 5

    Oppo Pad mini challenges Apple with Snapdragon 8 Gen 5

    April 21: Oppo Pad mini 8.8‑inch, Snapdragon 8 Gen 5, 5.39 mm, 279 g, 144 Hz OLED

    2 days ago
    Apple to ship 3 million foldable iPhones by end‑2026

    Apple to ship 3 million foldable iPhones by end‑2026

    Limited rollout equals 12 % of iPhone volume and rivals Samsung’s 2.4 million Galaxy Z Fold 7 sales

    2 days ago
    Apple unveils iPhone 18 Pro, iPhone 18 Pro Max, and iPhone Ultra

    Apple unveils iPhone 18 Pro, iPhone 18 Pro Max, and iPhone Ultra

    Mockups match leaked renders; 20 million Samsung panels for iPhone Ultra

    3 days ago
    Sony launches Playerbase program for Gran Turismo 7

    Sony launches Playerbase program for Gran Turismo 7

    PlayStation gamers can win a flight, facial scan, and an avatar in Gran Turismo 7

    3 days ago
    Claude Mythos Preview Beats Opus 4.6 in Cybersecurity!

    Claude Mythos Preview Beats Opus 4.6 in Cybersecurity!

    Claude Mythos Preview for five partners—pricing after a 100 million token credit

    3 days ago
    Loading...
banner