Robot Training: Beijing Man Masters New Skill Gameplay

Robot Training: Beijing Man Masters New Skill Gameplay
# COLA: Sensor-Free Human–Robot Collaboration

*(Still, it has to be college students who know how to have fun — doge!)*

---

## A Robot Sidekick with Campus Energy

While casually surfing the web, I stumbled upon something fascinating:  
A male college student has a **robot teammate** — and it’s *incredibly clingy* (well, sort of~).

**Scenes from their day:**

1. **Morning Supermarket Shift**  
   The robot follows him around during part-time supermarket work — once goods are packed, it happily pulls the cart and navigates stairs with ease:  
   ![image](https://blog.aitoearn.ai/content/images/2025/11/img_001-3.gif)

2. **Noon Campus Cafeteria**  
   It helps push the meal cart, instantly reacting to simple gestures — a pat on the head means *stop*:  
   ![image](https://blog.aitoearn.ai/content/images/2025/11/img_002-3.gif)

3. **Post-Work Workout**  
   After a long day, the robot even joins him for exercise:  
   ![image](https://blog.aitoearn.ai/content/images/2025/11/img_003-5.gif)

4. **Vlog Potential**  
   Imagine filming from the robot’s perspective: *“A High-Energy Robot’s Day”*:  
   ![image](https://blog.aitoearn.ai/content/images/2025/11/img_004-3.jpeg)

---

## The Unusual Interaction Method

Did you notice? All interaction is done via **patting its head** or **tugging its body**.  
No remote control. **No voice commands**.  
![image](https://blog.aitoearn.ai/content/images/2025/11/img_005-3.jpeg)  
![image](https://blog.aitoearn.ai/content/images/2025/11/img_006-24.png)

---

## From Fun to Research

Turns out, this is part of a legitimate high-tech research project — complete with a published academic paper:  
![image](https://blog.aitoearn.ai/content/images/2025/11/img_007-22.png)

---

## The Problem They’re Tackling

**Scenario:** *Human–robot collaborative object carrying*  

- For robotic arms: Well researched and validated.  
- For humanoid robots: Lacks deep exploration.

**Challenge:**  
Humanoids have **complex whole-body dynamics** — coordinating torso, limbs, joints, maintaining balance, and handling diverse environmental factors.

**Goal:**  
Enable **seamless human–humanoid collaboration** for moving varied objects.  
![image](https://blog.aitoearn.ai/content/images/2025/11/img_008-14.png)

---

## Their Solution: COLA

### Overview
**COLA** is a novel reinforcement learning method that operates **without cameras, LiDAR, or external sensors**, relying purely on **proprioception**.  
![image](https://blog.aitoearn.ai/content/images/2025/11/img_009-1.jpeg)

---

### Key Design Philosophy

#### 1. Unified Collaboration Policy
- Traditional: Separate “robot-leading” & “human-leading” modes → causes lag.
- **COLA merges roles into *one policy***:
  - Steady human force → robot follows.
  - Human hesitation or instability → robot leads & stabilizes.
- No manual input — roles switch fluidly.
- Even tasks like carrying heavy items upstairs become smooth:  
  ![image](https://blog.aitoearn.ai/content/images/2025/11/img_010-7.gif)

---

#### 2. Dynamic Closed-Loop Training
- Training environment simulates unpredictable events:
  - Sudden movement changes.
  - Shifts in object weight.
  - Grip slips.
- Focus: Real-world robustness.

---

#### 3. Continuous Feedback Loops
- Robot’s actions alter the environment → environment changes influence next decisions.
- Mirrors actual human–robot transport dynamics.

---

## Why COLA’s Approach Is Different

**COLA’s Core Advantage:**  
Operates purely through *intrinsic perception* — internal data:  
Joint angles, actuator force feedback, position/velocity info.

**Benefits:**
- Immune to environmental interference (e.g., poor lighting).
- Eliminates remote controls.
- Reduces hardware costs and system complexity.  
![image](https://blog.aitoearn.ai/content/images/2025/11/img_012-3.gif)

---

### Technical Steps

1. **Residual Teacher Fine-Tuning**  
   - Train base movement policy.  
   - Add a *Residual Teacher* for cooperative carrying adjustments.  
   - Merge into a unified collaboration strategy.

2. **Simulation Training & Knowledge Distillation**  
   - Run millions of safe simulation episodes.  
   - Train a strong “teacher” model.  
   - Distill knowledge into a lightweight “student” policy for real robots.

---

### Predictive Insights

COLA implicitly predicts:
- **Object movement trend** (tilting risk, direction).
- **Human intent signals** (turning, force adjustment needs).

**Result:**  
Load balance via coordinated trajectory planning.

---

## Experimental Validation

**Simulation Results:**
- Outperformed explicit goal estimation & Transformer-based methods in:
  - Movement precision (velocity, angular velocity, height error).
  - Lower human effort (average external force).
- **Leader-oriented COLA-L** > **Follower-oriented COLA-F**.

**Real-World Tests:**
- Objects: Boxes, flexible stretchers.  
- Patterns: Straight walking, turning.  
![image](https://blog.aitoearn.ai/content/images/2025/11/img_015.jpeg)

**Human Trials:**
- 23 participants.
- COLA rated **highest in tracking accuracy & smoothness**.  
![image](https://blog.aitoearn.ai/content/images/2025/11/img_016-3.png)

---

## The Team Behind COLA

### Co-First Authors
- **Yushi Du** – Dept. of EEE, University of Hong Kong.  
- **Yixuan Li** – Ph.D., Beijing Institute of Technology; interests in humanoid robotics.

---

### Corresponding & Senior Authors
- **Baoxiong Jia** – Research scientist, Beijing Institute for General AI; Peking Univ. & UCLA alumnus; multiple 2025 conference papers.  
- **Wei Liang** – Professor, BIT; heads PIE Lab; focuses on computer vision & VR.  
- **Yanchao Yang** – Asst. Professor, University of Hong Kong; specialties in embodied intelligence.  
- **Siyuan Huang** – Director, Center for Embodied AI & Robotics; teaches at Peking Univ.; worked at DeepMind, Facebook Reality Labs.

---

### Other Contributors
- **Yutang Lin** – Third-year student, Peking University.  
- **Pei Zhou** – Ph.D. student, University of Hong Kong.

---

## Potential for Creators

Sensor-free collaboration like COLA opens new storytelling opportunities.  
AI-powered content platforms such as [AiToEarn官网](https://aitoearn.ai/) can:
- Generate AI-driven content.
- Publish across Douyin, Instagram, YouTube, etc.
- Track and monetize engagement via integrated analytics.  

More info: [AiToEarn开源地址](https://github.com/yikart/AiToEarn).

---

## References
- **Paper:** https://www.arxiv.org/abs/2510.14293  
- **Project Homepage:** https://yushi-du.github.io/COLA/  
- **Social Reference:** https://x.com/siyuanhuang95/status/1980517755163185642  

---

![image](https://blog.aitoearn.ai/content/images/2025/11/img_023-2.png)

**What do you think?**

---

Read more

Xie Saining, Fei-Fei Li, and Yann LeCun Team Up for the First Time! Introducing the New "Hyperception" Paradigm — AI Can Now Predict and Remember, Not Just See

Xie Saining, Fei-Fei Li, and Yann LeCun Team Up for the First Time! Introducing the New "Hyperception" Paradigm — AI Can Now Predict and Remember, Not Just See

Spatial Intelligence & Supersensing: The Next Frontier in AI Leading AI researchers — Fei-Fei Li, Saining Xie, and Yann LeCun — have been highlighting a transformative concept: Spatial Intelligence. This goes beyond simply “understanding images or videos.” It’s about: * Comprehending spatial structures * Remembering events * Predicting future outcomes In essence, a truly

By Honghao Wang
Flexing Muscles While Building Walls: NVIDIA Launches OmniVinci, Outperforms Qwen2.5-Omni but Faces “Fake Open Source” Criticism

Flexing Muscles While Building Walls: NVIDIA Launches OmniVinci, Outperforms Qwen2.5-Omni but Faces “Fake Open Source” Criticism

NVIDIA OmniVinci: A Breakthrough in Multimodal AI NVIDIA has unveiled OmniVinci, a large language model designed for multimodal understanding and reasoning — capable of processing text, visual, audio, and even robotic data inputs. Led by the NVIDIA Research team, the project explores human-like perception: integrating and interpreting information across multiple data

By Honghao Wang