Bingqian Lin

Postdoctoral Fellow, Shanghai Jiao Tong University

I am currently a postdoctoral fellow at the Shanghai Jiao Tong University, working with Prof. Cewu Lu. I received my Ph.D. degree in Electronics Information Engineering at Sun Yat-sen University, advised by Prof. Xiaodan Liang and Prof. Liang Lin at the Human Cyber Physical Intelligence Integration Lab (HCP-I2 Lab). I was a research intern at Huawei Noah’s Ark Laboratory, advised by Prof. Jianzhuang Liu. I received my BS and MS degrees from the University of Electronic Science and Technology of China (UESTC) and the Xiamen University (XMU).

My research interest lies in Multi-modal understanding, Embodied AI, and Computer vision. Currently, I am focusing on building generalist embodied agents driven by foundation models and the world model theory.

Recent Activities

  • [12/2024]   One paper on Affordance-Oriented Planning for the VLN task is accepted by AAAI 2025.
  • [09/2024]   One paper on Waypoint-Aware World Model for Robotic Manipulation is accepted by Neurips 2024.
  • [06/2024]   Excited to receive Honors Graduate from Sun Yat-sen University!
  • [05/2024]   One paper on Correctable Landmark Discovery for the VLN task is accepted by IEEE TPAMI.
  • [05/2024]   One paper on Map-oriented Prompting for the VLN task is accepted by ACL 2024.
  • [05/2024]   I successfully defended my Ph.D. dissertation!
  • [03/2024]   One paper on Navigational Chain-of-Thought for the VLN task is available at Preprint.
  • [04/2023]   One paper on Knowledge-Enhanced Medical Multimodal Pretraining is available at Preprint.
  • [04/2023]   One paper on Training Deviation-robust Agents for the VLN task is available at TPAMI 2023.
  • [03/2023]   One paper on Dynamic Graph Enhanced Contrastive Learning for the MRG task is available at CVPR 2023.
  • [02/2023]   One paper on Actional Atomic-Concept Learning for the VLN task is available at AAAI 2023.
  • [03/2022]   One paper on Modality-Aligned Action Prompts for the VLN task is available at CVPR 2022.

Selected Publications/Preprints

Correctable Landmark Discovery via Large Models for Vision-Language Navigation
Bingqian Lin*, Yunshuang Nie*, Ziming Wei, Yi Zhu, Hang Xu, Shikui Ma, Jianzhuang Liu, Xiaodan Liang
TPAMI, 2024
arXiv / code / bibtex
NavCoT: Boosting LLM-Based Vision-and-Language Navigation via Learning Disentangled Reasoning
Bingqian Lin*, Yunshuang Nie*, Ziming Wei, Jiaqi Chen, Shikui Ma, Jianhua Han, Hang Xu, Xiaojun Chang, Xiaodan Liang
arXiv, 2024
arXiv / code / bibtex
MapGPT: Map-Guided Prompting with Adaptive Path Planning for Vision-and-Language Navigation
Jiaqi Chen, Bingqian Lin, Ran Xu, Zhenhua Chai, Xiaodan Liang, Kwan-Yee K. Wong
ACL, 2024
arXiv / code / bibtex
Towards Deviation-Robust Agent Navigation via Perturbation-Aware Contrastive Learning
Bingqian Lin*, Yanxin Long*, Yi Zhu, Fengda Zhu, Xiaodan Liang, Qixiang Ye, Liang Lin
TPAMI, 2023
arXiv / bibtex
Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Bingqian Lin*, Zicong Chen*, Mingjie Li*, Haokun Lin, Hang Xu, Yi Zhu, Jianzhuang Liu, Wenjia Cai, Lei Yang, Shen Zhao, Chenfei Wu, Ling Chen, Xiaojun Chang, Yi Yang, Lei Xing, Xiaodan Liang
arXiv, 2023
arXiv / code / bibtex
Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report Generation
Mingjie Li, Bingqian Lin, Zicong Chen, Haokun Lin, Xiaodan Liang, Xiaojun Chang
CVPR, 2023
arXiv / code / bibtex
Actional Atomic-Concept Learning for Demystifying Vision-Language Navigation
Bingqian Lin, Yi Zhu, Xiaodan Liang, Liang Lin, Jianzhuang Liu
AAAI Oral, 2023
arXiv / bibtex
RelCLIP: Adapting Language-Image Pretraining for Visual Relationship Detection via Relational Contrastive Learning
Yi Zhu*, Zhaoqing Zhu*, Bingqian Lin, Xiaodan Liang, Feng Zhao, Jianzhuang Liu
EMNLP, 2022
paper / bibtex
ADAPT: Vision-Language Navigation with Modality-Aligned Action Prompts
Bingqian Lin, Yi Zhu, Zicong Chen, Xiwen Liang, Jianzhuang Liu, Xiaodan Liang
CVPR, 2022
arXiv / code / bibtex
Contrastive Instruction-Trajectory Learning for Vision-Language Navigation
Xiwen Liang, Fengda Zhu, Yi Zhu, Bingqian Lin, Bing Wang, Xiaodan Liang
AAAI, 2022
arXiv / code / bibtex
Adversarial Reinforced Instruction Attacker for Robust Vision-Language Navigation
Bingqian Lin, Yi Zhu, Yanxin Long, Xiaodan Liang, Qixiang Ye, Liang Lin
TPAMI, 2021
arXiv / code / bibtex
Vision-Dialog Navigation by Exploring Cross-modal Memory
Yi Zhu, Fengda Zhu, Zhaohuan Zhan, Bingqian Lin, Jianbin Jiao, Xiaojun Chang, Xiaodan Liang
CVPR, 2020
arXiv / code / bibtex

Academic Services

Reviewer for Journal
  • IEEE Transactions on Neural Networks and Learning Systems (TNNLS)
  • IEEE Transactions on Multimedia (TMM)
  • IEEE Transactions on Circuits and Systems for Video Technology (TCSVT)
  • IEEE Transactions on Medical Imaging (TMI)
  • Neural Networks
Reviewer for Conference
  • The International Conference on Learning Representations (ICLR)
  • Annual Conference on Neural Information Processing Systems (Neurips)
  • IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
  • Annual Meeting of the Association for Computational Linguistics (ACL)
  • Empirical Methods in Natural Language Processing (EMNLP)

Contact

You are very welcome to contact me regarding my research. I typically respond within a few days.
I can be contacted directly at bingqianlin [at] 126.com