About Jing-Cheng Pang (庞竟成):

[Research Overview] [Recent News] [Selected Publications] [Projects]


Your Image

I am a fourth-year Ph.D. candidate at Nanjing University, where I study reinforcement learning under the guidance of Professor Yang Yu. In June 2019, I obtained B.Sc. degree from UESTC, after which I joined the LAMDA group led by Professor Zhi-Hua Zhou, as a Master’s student. In 2021, I transitioned to the Successive Postgraduate and Doctoral Program to pursue my current doctoral research. From July to October in 2024, I engaged as a visiting scholar with Prof. Masashi Sugiyama’s team at RIKEN-AIP, Tokyo, Japan.

Research Overview

My research focuses on connecting human and intelligent agent through natural language. By integrating reinforcement learning (RL) and large language models (LLMs), I aim to develop systems that not only interpret human intent but also act autonomously and learn iteratively in dynamic environments. Particularly, my study includes:

  • Reinforcement Learning: language-conditioned RL, optimization algorithm, imitation learning, generalist agent;
  • Large Language Models: model training, inference-time optimization, LLM-based agent;
  • Embodied AI: home-service robot, sim2real policy learning.

Free free to contact/follow me if you are interested in my work.

[Back to top]

Recent News

  • 2025.03: Invited talk at HUAWEI Jiaxian Community (稼先社区). Topic: Exploration on Intelligent Agent System based on RL and LLMs [slides].
  • 2025.01: ReViWo for robotics manipulation under viewpoint disturbance, is accepted by ICLR 2025.
  • 2025.01: Invited talk at HUAWEI NAIE Group. Topic: RL-driven LLM Optimization and Recent Progress [slides].
  • 2024.12: InCLET is accepted by AAMAS 2025 as a full paper. We will give an oral presentation at Detriot (It’s a pity that we can not get the visa)!
  • 2024.09: Our work, KALM, is accepted by NeurIPS 2024!
  • 2024.07: I start my visiting to Prof. Masashi Sugiyama’s team at RIKEN-AIP, Tokyo, Japan, doing RL research
  • 2024.01: RLC for LLM self-improvement is accepted by ICLR 2024
  • 2023.11: Awarded as Top Reviewer (8%) of NeurIPS 2023
  • 2023.09: TALAR, for instruction-following agents, gets accepted by NeurIPS 2023

[Back to top]

Selected Publications

  1. Jing-Cheng Pang, Nan Tang, Kaiyuan Li, Yuting Tang, Xin-Qiang Cai, Zhen-Yu Zhang, Gang Niu, Masashi Sugiyama and Yang Yu. Learning View-invariant World Models for Visual Robotic Manipulation. In: ICLR, 2025. [paper]
  2. Peng-Yuan Wang, Jing-Cheng Pang, Chen-Yang Wang, Xu-Hui Liu, Tian-Shuo Liu, Si-Hang Yang, Hong Qian and Yang Yu. InCLET: In-context Learning from Language Models can Improve Embodied Instruction-following. In: AAMAS (Oral), 2025. [paper]
  3. Jing-Cheng Pang, Si-Hang Yang, Kaiyuan Li, Jiaji Zhang, Xiong-Hui Chen, Nan Tang and Yang Yu. KALM: Knowledgeable Agents by Offline Reinforcement Learning from Large Language Model Rollouts. In: NeurIPS, 2024. [paper]
  4. Jing-Cheng Pang, Peng-Yuan Wang, Kaiyuan Li, Xiong-Hui Chen, Jiacheng Xu, Zongzhang Zhang and Yang Yu. Language Model Self-improvement by Reinforcement Learning Contemplation. In: ICLR, 2024. [paper]
  5. Jing-Cheng Pang, Xinyu Yang, Si-Hang Yang, Xiong-Hui Chen and Yang Yu. Natural Language Instruction-following with Task-related Language Development and Translation. In: NeurIPS, 2023. [paper]
  6. Jing-Cheng Pang, Tian Xu, Shengyi Jiang, Yu-Ren Liu and Yang Yu. Reinforcement Learning With Sparse-Executing Actions via Sparsity Regularization. TNNLS, to appear. [paper]
  7. Shengyi Jiang, Jing-Cheng Pang and Yang Yu. Offline Imitation Learning with a Misspecified Simulator. In: NeurIPS, 2020. [paper]

[Full publication list]

[Back to top]

Projects

[Back to top]