Minghuan Liu Profile
Minghuan Liu

@ericliuof97

Followers
342
Following
344
Statuses
102

Ph.D @sjtu1896. Prev: Visit @UCSD at @xiaolonw's lab. Robot Learning, Reinforcement Learning, Imitation Learning.

San Diego, CA
Joined September 2016
Don't wanna be here? Send us removal request.
@ericliuof97
Minghuan Liu
9 days
Introducing HugWBC: A Unified and General Humanoid Whole-Body Controller for Fine-Grained Locomotion! Project: We provide a general humanoid controller that achieves multiple gaits, and arbitrary combinations of commands in one single policy!
1
10
82
@ericliuof97
Minghuan Liu
9 days
@ChongZitaZhang 🤣Humanoid hurts
0
0
1
@ericliuof97
Minghuan Liu
9 days
This work would not be possible without our coauthors, Yufei Xue, Wentao Dong, Weinan Zhang, and @pangjiangmiao !
0
0
1
@ericliuof97
Minghuan Liu
9 days
Very impressive! High agility is truly what a powerful humanoid controller needed.
@TairanHe99
Tairan He
9 days
🚀 Can we make a humanoid move like Cristiano Ronaldo, LeBron James and Kobe Byrant? YES! 🤖 Introducing ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills Website: Code:
0
1
4
@ericliuof97
Minghuan Liu
1 month
@allenzren @physical_int Amazing open source project!
1
0
3
@ericliuof97
Minghuan Liu
2 months
@roeiherzig Just take a closer look of the work, very insightful! Sorry for my incorrect conclusion. The way of utilizing VLAs is still an open problem for now, and RoboVLMs are indeed still limited as a preliminary work.
0
0
1
@ericliuof97
Minghuan Liu
2 months
@YouJiacheng Sorry for my incorrect words. But MoE is from another dimension that all of the current four class of VLAs can be augmented with. If we ignore MoE, we can simply classify the current version of pi0 into one-step cont. Think we better add a limitation discussion for the pi0 case.
1
0
1
@ericliuof97
Minghuan Liu
2 months
@YouJiacheng It makes sense for you taxnomy of token interaction. As I mentioned before, this work keep the attention interaction in each vlm to support fast integration, allowing to support a) b) c). For d), we need to change the attention mask inside every vlm and should be studied further
0
0
0
@ericliuof97
Minghuan Liu
2 months
@YouJiacheng it's more like trainable param / training recipe, not policy formulation
1
0
0
@ericliuof97
Minghuan Liu
2 months
@YouJiacheng We define the name of interleave under a principle of history modeing. Think it may be better just a one-step continuous model (in fig 2
0
0
1
@ericliuof97
Minghuan Liu
2 months
@YouJiacheng Make sense. I think it may be better to be classified into a one-step continuous formation. Will fixed it soon in the revision. Also we have not supported that kind of implementation because if so we need to change the forward function specifically for each VLM backbone.
1
0
0
@ericliuof97
Minghuan Liu
2 months
@DavidFSWD The 9b model cannot work directly on Jeston. But in theory, we can do that with small models, distillation, and model compilation techniques if we want to do so
0
0
1
@ericliuof97
Minghuan Liu
2 months
@YouJiacheng Our RoboVLMs support pooling and learnable token format, and would add the cross-attention format in the future. But how to represent the output tokens does not change the fact that pi_0 can be classified into a policy head formulation.
1
0
0