Teli Ma 「马特立」

| Email | Google Scholar |
| Github | LinkedIn |

I am a Ph.D. student in Artificial Intelligence at the AI Thrust, Information Hub, The Hong Kong University of Science and Technology (Guangzhou), advised by Prof. Junwei Liang.

I received my Bachelor's degree from the Honors College of Beihang University (BUAA), supervised by Prof. Baochang Zhang and Prof. David Doermann. I also spent time at OpenGVLab, Shanghai AI Lab, collaborated with Dr. Peng Gao and Prof. Hongsheng Li.

Research Interest: The intersection of large-scale pre-training, generalizable manipulation and whole-body control, .

Research Question: How can we build robots that perceive, reason, and act in the open world with human-like generalization?

Email: telima9868 [AT] gmail.com / tma184 [AT] connect.hkust-gz.edu.cn


  News
  • [09/2025] GLOVER won the Best Paper Award at the CoRL 2025 GenPriors Workshop!
  • [08/2025] GLOVER++ and OmniPerception are accepted by CoRL 2025!
  • [02/2025] One paper accepted to CVPR 2025 [project page].
  • [09/2024] One paper accepted to CoRL 2024 [project page].
  • [03/2024] One paper accepted to NAACL 2024 [project page].
  • [10/2023] Went to Paris for ICCV 2023 — nice trip!
  • [09/2023] Nominated by the PCs as a reviewer of ICLR 2024.
  • [09/2023] Honored to be the representative of new PG students at the orientation of INFO Hub.
  • [08/2023] CLIP-Adapter is accepted by IJCV.
  • [07/2023] One paper accepted to ICCV 2023 [SyncTrack].
  • [11/2022] One paper accepted to AAAI 2023 [ReBNN (oral)].
  • [10/2022] One paper accepted to BMVC 2022.
  • [10/2022] One paper accepted to NeurIPS 2022 [MCMAE/ConvMAE (spotlight, 3.7% acceptance rate, 384/10411)].
  • [07/2022] Two papers accepted to ECCV 2022 [IDa-Det] [RBONNs (oral, 2.7% acceptance rate, 158/5803)].
  • [10/2021] One paper accepted to NeurIPS 2021 [DSNet].

  Publications

DiT4DiT: Jointly Modeling Video Dynamics and Actions for Generalizable Robot Control
Teli Ma, Jia Zheng, Zifan Wang, Chunli Jiang, Andy Cui, Junwei Liang, Shuo Yang
arXiv preprint, 2026

webpage | pdf | bibtex | code
@article{ma2026dit4dit,
        title={DiT4DiT: Jointly Modeling Video Dynamics and Actions for Generalizable Robot Control},
        author={Ma, Teli and Zheng, Jia and Wang, Zifan and Jiang, Chunli and Cui, Andy and Liang, Junwei and Yang, Shuo},
        journal={arXiv preprint arXiv:2603.10448},
        year={2026}
      }
Omni-Perception

Omni-Perception: Omnidirectional Collision Avoidance for Legged Locomotion in Dynamic Environments
Zifan Wang, Teli Ma, Yufei Jia, Xun Yang, Jiaming Zhou, Wenlong Ouyang, Qiang Zhang, Junwei Liang
CoRL 2025 (Oral Presentation)

webpage | pdf | bibtex | arXiv | code
@misc{wang2025omniperception,
      title={Omni-Perception: Omnidirectional Collision Avoidance for Legged Locomotion in Dynamic Environments}, 
      author={Zifan Wang and Teli Ma and Yufei Jia and Xun Yang and Jiaming Zhou and Wenlong Ouyang and Qiang Zhang and Junwei Liang},
      year={2025},
      eprint={2505.19214},
      archivePrefix={arXiv},
      primaryClass={cs.RO},
      url={https://arxiv.org/abs/2505.19214}, 
}
GLOVER++

GLOVER++: Unleashing the Potential of Affordance Learning from Human Behaviors for Robotic Manipulation
Teli Ma*, Jia Zheng*, Zifan Wang, Ziyao Gao, Jiaming Zhou, Junwei Liang
CoRL 2025

webpage | pdf | bibtex | arXiv | code
@article{ma2025glover++,
title={GLOVER++: Unleashing the Potential of Affordance Learning from Human Behaviors for Robotic Manipulation},
author={Ma, Teli and Zheng, Jia and Wang, Zifan and Gao, Ziyao and Zhou, Jiaming and Liang, Junwei},
journal={arXiv preprint arXiv:2505.11865},
year={2025} } 
GLOVER

GLOVER: Generalizable Open-Vocabulary Affordance Reasoning for Task-Oriented Grasping
Teli Ma*, Zifan Wang*, Jiaming Zhou, Mengmeng Wang, Junwei Liang
GenPriors Workshop at CoRL 2025 (Best Paper Award)

webpage | pdf | bibtex | arXiv | code
@article{ma2024glover,
      title={Glover: Generalizable open-vocabulary affordance reasoning for task-oriented grasping},
      author={Ma, Teli and Wang, Zifan and Zhou, Jiaming and Wang, Mengmeng and Liang, Junwei},
      journal={arXiv preprint arXiv:2411.12286},
      year={2024}
    }
Mitigating Human-Robot Domain Discrepancy

Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation
Jiaming Zhou, Teli Ma, Kun-Yu Lin, Ronghe Qiu, Zifan Wang, Junwei Liang
CVPR 2025

webpage | pdf | bibtex | arXiv | code
@article{zhou2024mitigating,
        title={Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation},
        author={Zhou, Jiaming and Ma, Teli and Lin, Kun-Yu and Qiu, Ronghe and Wang, Zifan and Liang, Junwei},
        journal={arXiv preprint arXiv:2406.14235},
        year={2024}
      }
Sigma Agent

Contrastive Imitation Learning for Language-guided Multi-Task Robotic Manipulation
Teli Ma, Jiaming Zhou, Zifan Wang, Ronghe Qiu, Junwei Liang
CoRL 2024

webpage | pdf | bibtex | arXiv | code
@article{ma2024contrastive,
  title={Contrastive imitation learning for language-guided multi-task robotic manipulation},
  author={Ma, Teli and Zhou, Jiaming and Wang, Zifan and Qiu, Ronghe and Liang, Junwei},
  journal={arXiv preprint arXiv:2406.09738},
  year={2024}
}
SADE

An Examination of the Compositionality of Large Generative Vision-Language Models
Teli Ma, Rong Li, Junwei Liang
NAACL 2024

webpage | pdf | bibtex | arXiv | code
[[BIBTEX]]
SyncTrack

Synchronize Feature Extracting and Matching: A Single Branch Framework for 3D Object Tracking
Teli Ma*, Mengmeng Wang*, Jimin Xiao, Huifeng Wu, Yong Liu
ICCV 2023

pdf | bibtex | arXiv
[[BIBTEX]]
CorpNet

Correlation Pyramid Network for 3D Single Object Tracking
Mengmeng Wang, Teli Ma, Xingxing Zuo, Jiajun Lv, Yong Liu
CVPRW 2023

pdf | bibtex
[[BIBTEX]]
ReBNN

Resilient Binary Neural Network
Sheng Xu*, Yanjing Li*, Teli Ma*, Mingbao Lin, Hao Dong, Baochang Zhang, Peng Gao, Jinhu Lv
AAAI 2023 (Oral Presentation)

pdf | bibtex | arXiv | code
[[BIBTEX]]
BALLAD

Unleashing the Potential of Vision-Language Models for Long-Tailed Visual Recognition
Teli Ma, Shijie Geng, Mengmeng Wang, Sheng Xu, Hongsheng Li, Baochang Zhang, Peng Gao, Yu Qiao
BMVC 2022

pdf | bibtex | code
[[BIBTEX]]
MCMAE

MCMAE: Masked Convolution Meets Masked Autoencoders
Peng Gao, Teli Ma, Hongsheng Li, Ziyi Lin, Jifeng Dai, Yu Qiao
NeurIPS 2022 (Spotlight)

pdf | bibtex | code
[[BIBTEX]]
IDa-Det

IDa-Det: An Information Discrepancy-aware Distillation for 1-bit Detectors
Sheng Xu*, Yanjing Li*, Bohan Zeng*, Teli Ma, Baochang Zhang, Xianbin Cao, Peng Gao, Jinhu Lv
ECCV 2022

pdf | bibtex | code
[[BIBTEX]]
RBONN

Recurrent Bilinear Optimization for Binary Neural Networks
Sheng Xu*, Yanjing Li*, Tiancheng Wang, Teli Ma, Baochang Zhang, Peng Gao, Yu Qiao, Jinhu Lv, Guodong Guo
ECCV 2022 (Oral Presentation)

pdf | bibtex | code
[[BIBTEX]]
DSNet

Dual-Stream Network for Visual Recognition
Mingyuan Mao*, Renrui Zhang*, Honghui Zheng*, Teli Ma, Yan Peng, Errui Ding, Baochang Zhang, Shumin Han
NeurIPS 2021

pdf | bibtex
[[BIBTEX]]
CLIP-Adapter

CLIP-Adapter: Better Vision-Language Models with Feature Adapters
Peng Gao*, Shijie Geng*, Renrui Zhang*, Teli Ma, Rongyao Fang, Yongfeng Zhang, Hongsheng Li, Yu Qiao
IJCV 2023

pdf | bibtex | arXiv | code
[[BIBTEX]]

  Awards
  • PG Scholarship, HKUST(GZ), 2023
  • Excellent Intern Award, Baidu Inc., 2021
  • Outstanding Undergraduate Thesis, BUAA, 2020
  • Honor Undergraduate Degree, BUAA, 2020
  • Excellent Student Scholarship, BUAA, 2018

  Reviewer Service
AAAI Conference on Artificial Intelligence (AAAI), 2025
International Conference on Learning Representations (ICLR), 2024, 2025, 2026
International Conference on Machine Learning (ICML), 2024, 2025, 2026
Conference on Computer Vision and Pattern Recognition (CVPR), 2022, 2023, 2024, 2025, 2026
European Conference on Computer Vision (ECCV), 2024, 2026
International Conference on Computer Vision (ICCV), 2023, 2025
Conference on Robot Learning (CoRL), 2025, 2026
Conference on Neural Information Processing Systems (NeurIPS), 2023, 2024, 2025
IEEE Robotics and Automation Letters (RA-L), 2025


Website template from here and here.