Jing Liu

Jing Liu(刘璟)

Distinguished Architect, Baidu
Leading LLM post-training efforts at Baidu, focusing on alignment, general agents, coding agents, and deep search.
LLM Post-Training RLHF / Alignment Agentic RL Reasoning RL On-Policy Distillation General Agent Coding Agent Deep Search

About

I lead the LLM Post-Training team at Baidu, where our work spans post-training alignment (RLHF, agentic RL, reasoning RL, reward modeling, on-policy distillation, data curation), general agents (tool use, planning, multi-agent systems), coding agents (SWE agent, WebDev agent), and deep search (deep search, wide search, deep research).

Before joining Baidu in December 2017, I was a researcher at Microsoft Research Asia (MSRA), working on information retrieval, question answering, and knowledge bases. I obtained my Ph.D. in computer science from Harbin Institute of Technology (HIT) under the supervision of Prof. Hsiao-Wuen Hon, Prof. Ting Liu, and Dr. Chin-Yew Lin.

We Are Hiring

The LLM Post-Training team at Baidu is looking for talented engineers and researchers. Both internship and full-time positions are available.

Post-Training

RLHF, agentic RL, reasoning RL, reward modeling, on-policy distillation, data curation

Coding Agent

SWE agent, WebDev agent

General Agent

Tool use, planning, multi-agent systems

Deep Search

Deep search, wide search, deep research

Interested? Send your resume to legendarydan (at) gmail (dot) com

Selected Publications (2023–present)

2026

AAAI 2026
Yuhao Wang, Ruiyang Ren, Yucheng Wang, Jing Liu, Wayne Xin Zhao, Hua Wu, Haifeng Wang

2025

Preprint
Yuhao Wang, Ruiyang Ren, Yucheng Wang, Wayne Xin Zhao, Jing Liu, Hua Wu, Haifeng Wang
SIGIR 2025
Yuhao Wang, Ruiyang Ren, Yucheng Wang, Wayne Xin Zhao, Jing Liu, Hua Wu, Haifeng Wang
WWW 2025
Ruiyang Ren, Yuhao Wang, Kun Zhou, Wayne Xin Zhao, Wenjie Wang, Jing Liu, Ji-Rong Wen, Tat-Seng Chua
COLING 2025
Ruiyang Ren, Yuhao Wang, Yingqi Qu, Wayne Xin Zhao, Jing Liu, Hao Tian, Hua Wu, Ji-Rong Wen, Haifeng Wang

2024

EMNLP 2024
Yuhao Wang, Ruiyang Ren, Junyi Li, Wayne Xin Zhao, Jing Liu, Ji-Rong Wen
EMNLP 2024
Ruiyang Ren, Peng Qiu, Yingqi Qu, Jing Liu, Wayne Xin Zhao, Hua Wu, Ji-Rong Wen, Haifeng Wang
ACM TOIS 2024
Wayne Xin Zhao, Jing Liu, Ruiyang Ren, Ji-Rong Wen

2023

ACL 2023
Ruiyang Ren, Wayne Xin Zhao, Jing Liu, Hua Wu, Ji-Rong Wen, Haifeng Wang
SMoA: Sparse Mixture of Adapters to Mitigate Multiple Dataset Biases
ACL 2023 Workshop (TrustNLP)
Yanchen Liu, Jing Yan, Yan Chen, Jing Liu, Hua Wu

2022

EMNLP 2022
Yifu Qiu, Hongyu Li, Yingqi Qu, Ying Chen, Qiaoqiao She, Jing Liu, Hua Wu, Haifeng Wang
Findings of ACL 2022
Le Qi, Shangwen Lv, Hongyu Li, Jing Liu, Yu Zhang, Qiaoqiao She, Hua Wu, Haifeng Wang, Ting Liu
EMNLP 2022
Hongyu Zhu, Yan Chen, Jing Yan, Jing Liu, Yu Hong, Ying Chen, Hua Wu, Haifeng Wang

2021

EMNLP 2021
Ruiyang Ren, Yingqi Qu, Jing Liu, Wayne Xin Zhao, Qiaoqiao She, Hua Wu, Haifeng Wang, Ji-Rong Wen
Findings of ACL 2021
Ruiyang Ren, Shangwen Lv, Yingqi Qu, Jing Liu, Wayne Xin Zhao, Qiaoqiao She, Hua Wu, Haifeng Wang, Ji-Rong Wen
NAACL 2021
Yingqi Qu, Yuchen Ding, Jing Liu, Kai Liu, Ruiyang Ren, Wayne Xin Zhao, Daxiang Dong, Hua Wu, Haifeng Wang

2020

AAAI 2020
Kai Liu, Xin Liu, An Yang, Jing Liu, Jinsong Su, Sujian Li, Qiaoqiao She

2019

Preprint
Quan Wang, Pingping Huang, Haifeng Wang, Songtai Dai, Wenbin Jiang, Jing Liu, Yajuan Lyu, Yong Zhu, Hua Wu
EMNLP 2019 Workshop (MRQA)
Hongyu Li, Xiyuan Zhang, Yibing Liu, Yiming Zhang, Quan Wang, Xiangyang Zhou, Jing Liu, Hua Wu, Haifeng Wang
ACL 2019
An Yang, Quan Wang, Jing Liu, Kai Liu, Yajuan Lyu, Hua Wu, Qiaoqiao She, Sujian Li
Towards Robust Neural Machine Reading Comprehension via Question Paraphrases
IALP 2019
Ying Li, Hongyu Li, Jing Liu
Preprint
Tianwen Jiang, Sendong Zhao, Jing Liu, Jin-Ge Yao, Ming Liu, Bing Qin, Ting Liu, Chin-Yew Lin

2018

EMNLP 2018
Xingwu Sun, Jing Liu, Yajuan Lyu, Yanjun Ma, Shi Wang
CoNLL 2018
Feng Nie, Shuyan Zhou, Jing Liu, Jinpeng Wang, Chin-Yew Lin, Rong Pan
COLING 2018
Danqing Huang, Jing Liu, Chin-Yew Lin, Jian Yin
ACL 2018 Workshop (MRQA)
Wei He, Kai Liu, Jing Liu, Yajuan Lyu, Shiqi Zhao, Xinyan Xiao, Yuan Liu, Yizhong Wang, Hua Wu, Qiaoqiao She, Xuan Liu, Tian Wu, Haifeng Wang
ACL 2018 Workshop (MRQA)
An Yang, Kai Liu, Jing Liu, Yajuan Lyu, Sujian Li
ACL 2018
Yizhong Wang, Kai Liu, Jing Liu, Wei He, Yajuan Lyu, Hua Wu, Sujian Li, Haifeng Wang
LREC 2018
Tingsong Jiang, Jing Liu, Chin-Yew Lin

2017

IJCNLP 2017
Jinpeng Wang, Yutai Hou, Jing Liu, Yunbo Cao, Chin-Yew Lin

2016

ACL 2016
Hao Peng, Jing Liu, Chin-Yew Lin
ACL 2016
Quan Wang, Jing Liu, Yuanfei Luo, Bin Wang, Chin-Yew Lin
ACL 2016
Lei Sha, Jing Liu, Chin-Yew Lin, Sujian Li, Baobao Chang, Zhifang Sui

2015

CIKM 2015
Jyun-Yu Jiang, Jing Liu, Chin-Yew Lin, Pu-Jen Cheng

2013

EMNLP 2013
Jing Liu, Quan Wang, Chin-Yew Lin, Hsiao-Wuen Hon
WSDM 2013
Jing Liu, Fan Zhang, Xinying Song, Young-In Song, Chin-Yew Lin, Hsiao-Wuen Hon

2012

CIKM 2012
Jing Liu, Xinying Song, Jingtian Jiang, Chin-Yew Lin

2011

SIGIR 2011
Jing Liu, Young-In Song, Chin-Yew Lin

2010

CIKM 2010
Xinying Song, Jing Liu, Yunbo Cao, Chin-Yew Lin, Hsiao-Wuen Hon
NTCIR 2010
Young-In Song, Jing Liu, Tetsuya Sakai, Xinjing Wang, Guwen Feng, Yunbo Cao, Hisami Suzuki, Chin-Yew Lin

Experience

Education

Professional Activities

Datasets

LUGE [portal]

An open-source project of Chinese NLP benchmarks, jointly launched by Baidu, CCF, and CIPSC. LUGE aims to provide comprehensive evaluations across multiple NLP tasks.

DuReaderretrieval [paper] [data] [code]

A large-scale Chinese dataset for passage retrieval, containing over 90K questions and 8M passages from Baidu Search.

DuQM [paper] [data] [code]

A Chinese dataset for evaluating the robustness of question matching models across lexical, syntactic, and semantic dimensions.

DuReaderchecklist [data] [code]

A Chinese dataset for evaluating machine reading comprehension models across multiple dimensions including vocabulary, phrase, semantic role, and reasoning.

DuReaderrobust [paper] [data] [code]

A Chinese dataset for evaluating the robustness and generalization of machine reading comprehension models in real-world applications.