The Heart AI Can’t See Becomes the Best AI Detector

The Heart AI Can’t See Becomes the Best AI Detector

We Live in the Flow, While AI Lives in Frames

image

The Floating Heart Illusion

Recently, I came across an intriguing optical illusion online — a static “floating heart” pattern:

image

Tip:

If you open this article on a PC and the image doesn’t seem to move, try viewing it on your phone or shrinking your browser window.

You should see the heart start to jump side-to-side.

Apparently, some people believe this is the best AI detector — because no AI model can recognize there’s a heart in it.

---

Testing AI Models

I tested several leading AI models. None of them identified the heart.

  • Gemini 2.5 Pro failed first:
image

Output: irrelevant descriptions → concluded “circle” (incorrect).

  • GPT‑5-Thinking
  • Took over two minutes to “think”, then crashed:
  • GPT‑5 Pro
  • After seven minutes and verbose output, it gave up:
image
image
  • Doubao, Qwen, Yuanbao — China’s top 3 multimodal models — also failed:
  • DeepSeek avoided the test entirely (no multimodal capability).
image

---

Interestingly, the models do know what a heart shape is:

image

So this isn’t a conceptual ignorance problem — it’s a perception problem.

---

The Deeper Question: What’s Going On?

Digging deeper, I found a fascinating research paper from May:

_"Time Blindness: Why Video-Language Models Can’t See What Humans Can?"_

image

While the paper focuses on video illusions, the principle applies to the heart illusion too.

---

The SpookyBench Experiment

The researchers created synthetic videos consisting entirely of black-and-white static noise.

  • Paused frame: pure noise
  • Played video: humans clearly see a deer

Screenshotting the deer is impossible — it exists only in motion over time.

They compiled 451 such videos, testing state-of-the-art vision models:

image
image

Results:

  • Humans: ~98% accuracy in detecting shapes
  • AI models: 0% accuracy — across all architectures, training sets, and prompts
image

I tested Gemini 2.5 Pro myself — it couldn’t detect the deer either.

---

Why AI Fails: Frames vs Flow

The core issue:

Current vision-language models don’t watch videos — they sample static frames.

Process:

  • Take still images at fixed time intervals (e.g., 1s, 1.5s, 2s…)
  • Analyze each image for spatial information
  • Lose all temporal motion information in between

Thus:

  • Each frame = noise
  • No “heart” or “deer” exists in any single frame

This limitation is called Time Blindness.

---

Gestalt Psychology & the Law of Common Fate

image

Humans have an evolved instinct — the Law of Common Fate:

> Objects moving in the same direction are perceived as part of the same entity.

Example:

  • Prehistoric human spots shrubs moving randomly
  • A small patch moves differently, consistently toward them
  • Brain instantly concludes: “Predator approaching!”

---

Applied to the deer illusion:

  • Upward-moving noise points = deer
  • Downward-moving noise points = background

You don’t literally see the deer — you see patterned movement.

AI cannot do this because:

  • Its vision models focus on spatial features
  • Temporal movement patterns are lost in pre-processing

---

AI vs Human Perception

  • Humans: Continuous, dynamic world → motion-first perception
  • AI: Discrete, static world → object-first perception

---

The Floating Heart — A Static Illusion

So why does the heart image appear to move even though it’s static?

image

Answer:

Because your eyes are never still.

---

Micro Eye Movements

1950s research showed:

  • Human eyes make tiny, involuntary movements even when “fixed”
  • These movements refresh the image on our retinas

Illusion images exploit this, creating perceived motion.

image

---

Troxler Fading Effect

Definition:

> When staring at a fixed point, unchanging stimuli in peripheral vision fade within 1–3 seconds.

Example Experiment:

  • Stare at the cross in the center of the image
  • Colors around it fade, replaced by gray-white
image

Philosophy: _No change equals no information._

---

Full Circle: UX Design Meets AI Research

Years ago, I worked in user experience and studied cognitive psychology. Today, that same knowledge is resurfacing in AI research.

Humans and AI:

  • Sometimes take different paths to the same insight
  • But humans can perceive subtleties AI currently cannot — like love in silence, beauty in impermanence, and time itself.

---

Closing Thoughts

Improving AI’s temporal perception could unlock:

  • Illusion recognition
  • Richer video understanding
  • More human-like situational awareness

Platforms like AiToEarn官网 already help creators leverage AI strengths in spatial and textual domains, enabling:

  • AI-powered content generation
  • Cross-platform publishing (Douyin, Kwai, WeChat, Bilibili, Rednote, YouTube, Instagram, Threads, Pinterest, Facebook, LinkedIn, X)
  • Analytics & monetization

While these tools don’t fix time blindness yet, they provide powerful ways to deploy AI creativity today.

---

We live in the flow. AI lives in frames.

And maybe, one day, it will join us in the flow.

---

Read the original

Open in WeChat

Read more