ๆบๅจไนๅฟ JIQIZHIXIN (@jiqizhixin)
2026-01-04 | โค๏ธ 128 | ๐ 23
What if you could make a robotโs AI brain more stable and reliable at test time, without expensive retraining?
Researchers from China Telecom AI Institute, Tsinghua University, HKUST, and USTC present TACO.
They fix a key flaw in robot AI models (VLAs): after fine-tuning, they can generate shaky, inconsistent actions. TACO acts as a lightweight โaction verifierโ at inference, picking the most reliable action from multiple options.
It significantly boosts success rates & stability in robot simulations, outperforming standard fine-tuning methods.
Steering Vision-Language-Action Models as Anti-Exploration: A Test-Time Scaling Approach
Paper: https://arxiv.org/abs/2512.02834 Project: https://vla-anti-exploration.github.io/ย Code: https://github.com/breez3young/TACO/
Our report: https://mp.weixin.qq.com/s/t3u7Iv6es3XMTJTZmSYeHA
๐ฌ PapersAccepted by Jiqizhixin
๐ ์๋ณธ ๋งํฌ
- https://arxiv.org/abs/2512.02834
- https://vla-anti-exploration.github.io/
- https://github.com/breez3young/TACO/
- https://mp.weixin.qq.com/s/t3u7Iv6es3XMTJTZmSYeHA
๋ฏธ๋์ด

๐ Related
Auto-generated - needs manual review