Shuzheng Si

Tsinghua University Ph.D. student, Tsinghua University

Hi, I’m Shuzheng Si, a second-year CS Ph.D. Candidate at Tsinghua University. I am lucky to be advised by Prof. Maosong Sun from the TsinghuaNLP Lab. Before this, I obtained my master’s degree from Peking University, where I was supervised by Prof. Baobao Chang at the Institute of Computational Linguistics. My research interests lie in Natural Language Processing and Large Language Models. Recently, I have been interested in a fundamental challenge for LLMs: Hallucinations.

  • πŸ“– Hallucination Attribution β€” Understanding Hallucinations at the Source: My first line of research aims to explore and understand how training data contributes to hallucinations (NOVA, GATEAU, NUGGETS). This line of research mainly focuses on combining both qualitative and quantitative analyses of training data, guiding data collection to reduce LLM hallucinations.
  • πŸ”§ Hallucination Mitigation β€” Reducing Hallucinations in LLMs: This line of research focuses on designing effective strategies to reduce hallucinations, covering text generation (CANOE, LingoEDU), multi-modal scenarios (MMICL, LACING), and agentic tasks (SpokenWOZ, EAGLET, RhinoInsight). These methods provide practical solutions to reduce hallucinations in real-world applications.
  • πŸ”Ž Hallucination Detection β€” Identifying Hallucinations in the Wild: This research attempts to identify hallucinated responses generated by LLMs and prevent such hallucinated responses from being provided to users (FaithLens, Infi-Check). By enabling the timely hallucination detection and practical interventions, LLM-based systems can be more reliable and trustworthy in real-world deployment.
  • 🌏 Building Information-Seeking Tools With the Lowest Hallucination Rates: I also apply my research to build real-world information-seeking applications with the lowest hallucination rates, e.g., LingoWhale, RhinoInsight, and Zhiliao News released by DeepLang AI and TsinghuaNLP Lab. To date, these applications have provided trustworthy text processing services to hundreds of thousands of Chinese users.
I’m always happy to discuss potential collaborations. Feel free to drop an email if you are interested in connecting. πŸ•ŠπŸ•ŠπŸ•Š


Education
  • Tsinghua University

    Tsinghua University

    Ph.D. in Computer Science and Technology Sep. 2024 - Jul. 2028 (expected)

  • Peking University

    Peking University

    M.S. in Software Engineering Sep. 2021 - Jul. 2024

  • Yunnan University

    Yunnan University

    B.S. at the School of Software (Rank: 1/300+) Sep. 2017 - Jul. 2021

Honors & Awards
  • CAST’s Young Talents Support Project, Ph.D. Program 2025
  • EMNLP SAC Highlights Paper Award 2025
  • Comprehensive Excellence Scholarship, THU 2025
  • Merit Student, PKU 2022
  • Student of the Year Nominee Award (Ranked 1st, YNU) 2020
  • National Scholarship 2019
  • Provincial Government Scholarship 2018
Experience
  • DeepLang AI

    DeepLang AI

    Research Staff Apr. 2024 - Now

  • Alibaba DAMO Academy

    Alibaba DAMO Academy

    Research Intern Jun. 2022 - Jun. 2023

  • SenseTime Research

    SenseTime Research

    Research Intern Jul. 2021 - Feb. 2022

Service
  • NLP Communities: Reviewer of ACL, EMNLP, NAACL, COLING, and TASLP
  • ML Communities: Reviewer of NeurIPS, ICLR, ICML, and AAAI
  • CV Communities: Reviewer of ICCV
  • I am also a member of the BIRD team, led by the talent researcher Jinyang Li, which drives the development of text-to-SQL for real-world database applications
News
2025
β–ͺ πŸš€ Honored to receive the China Association for Science and Technology (CAST) Young Talents Support Project for Doctoral Students.
Dec 15
β–ͺ πŸ† My first-authored paper GATEAU wins the SAC Highlights Award (top 35 out of 8,000+ submissions) at EMNLP 2025!
Nov 09
β–ͺ πŸŽ‰ Four papers accepted by EMNLP 2025, congrats to all co-authors!
Aug 25
β–ͺ πŸŽ‰ Four papers accepted by ACL 2025, congrats to all co-authors!
May 15
2024
β–ͺ πŸ§‘πŸ»β€πŸ’» Started my Ph.D. study journey.
Sep 01
First-Authored Papers (view all papers on Google Scholar )
FaithLens: Detecting and Explaining Faithfulness Hallucination
FaithLens: Detecting and Explaining Faithfulness Hallucination

Shuzheng Si, Qingyi Wang, Haozhe Zhao, Yuzhuo Bai, Guanqiao Chen, Kangyang Luo, Gang Chen, Fanchao Qi, Minjia Zhang, Baobao Chang, Maosong Sun

Preprint 2025

FaithLens: Detecting and Explaining Faithfulness Hallucination
FaithLens: Detecting and Explaining Faithfulness Hallucination

Shuzheng Si, Qingyi Wang, Haozhe Zhao, Yuzhuo Bai, Guanqiao Chen, Kangyang Luo, Gang Chen, Fanchao Qi, Minjia Zhang, Baobao Chang, Maosong Sun

Preprint 2025

A Goal Without a Plan Is Just a Wish: Efficient and Effective Global Planner Training for Long-Horizon Agent Tasks
A Goal Without a Plan Is Just a Wish: Efficient and Effective Global Planner Training for Long-Horizon Agent Tasks

Shuzheng Si, Haozhe Zhao, Kangyang Luo, Gang Chen, Fanchao Qi, Minjia Zhang, Baobao Chang, Maosong Sun

Preprint 2025

A Goal Without a Plan Is Just a Wish: Efficient and Effective Global Planner Training for Long-Horizon Agent Tasks
A Goal Without a Plan Is Just a Wish: Efficient and Effective Global Planner Training for Long-Horizon Agent Tasks

Shuzheng Si, Haozhe Zhao, Kangyang Luo, Gang Chen, Fanchao Qi, Minjia Zhang, Baobao Chang, Maosong Sun

Preprint 2025

From Context to EDUs: Faithful and Structured Context Compression via Elementary Discourse Unit Decomposition
From Context to EDUs: Faithful and Structured Context Compression via Elementary Discourse Unit Decomposition

Yiqing Zhou*, Yu Lei*, Shuzheng Si*, Qingyan Sun*, Wei Wang, Yifei Wu, Hao Wen, Gang Chen, Fanchao Qi, Maosong Sun

Technical Report 2025 (* indicates co-first authors)

From Context to EDUs: Faithful and Structured Context Compression via Elementary Discourse Unit Decomposition
From Context to EDUs: Faithful and Structured Context Compression via Elementary Discourse Unit Decomposition

Yiqing Zhou*, Yu Lei*, Shuzheng Si*, Qingyan Sun*, Wei Wang, Yifei Wu, Hao Wen, Gang Chen, Fanchao Qi, Maosong Sun

Technical Report 2025 (* indicates co-first authors)

RhinoInsight: Improving Deep Research through Control Mechanisms for Model Behavior and Context
RhinoInsight: Improving Deep Research through Control Mechanisms for Model Behavior and Context

Yu Lei*, Shuzheng Si*, Wei Wang*, Yifei Wu, Gang Chen, Fanchao Qi, Maosong Sun

Technical Report 2025 (* indicates co-first authors)

RhinoInsight: Improving Deep Research through Control Mechanisms for Model Behavior and Context
RhinoInsight: Improving Deep Research through Control Mechanisms for Model Behavior and Context

Yu Lei*, Shuzheng Si*, Wei Wang*, Yifei Wu, Gang Chen, Fanchao Qi, Maosong Sun

Technical Report 2025 (* indicates co-first authors)

Teaching Large Language Models to Maintain Contextual Faithfulness via Synthetic Tasks and Reinforcement Learning
Teaching Large Language Models to Maintain Contextual Faithfulness via Synthetic Tasks and Reinforcement Learning

Shuzheng Si, Haozhe Zhao, Cheng Gao, Yuzhuo Bai, Zhitong Wang, Bofei Gao, Kangyang Luo, Wenhao Li, Yufei Huang, Gang Chen, Fanchao Qi, Minjia Zhang, Baobao Chang, Maosong Sun

AAAI 2026 (Oral), KnowFM@ACL 2025 Workshop (Oral)

Teaching Large Language Models to Maintain Contextual Faithfulness via Synthetic Tasks and Reinforcement Learning
Teaching Large Language Models to Maintain Contextual Faithfulness via Synthetic Tasks and Reinforcement Learning

Shuzheng Si, Haozhe Zhao, Cheng Gao, Yuzhuo Bai, Zhitong Wang, Bofei Gao, Kangyang Luo, Wenhao Li, Yufei Huang, Gang Chen, Fanchao Qi, Minjia Zhang, Baobao Chang, Maosong Sun

AAAI 2026 (Oral), KnowFM@ACL 2025 Workshop (Oral)

GATEAU: Selecting Influential Samples for Long Context Alignment
GATEAU: Selecting Influential Samples for Long Context Alignment

Shuzheng Si, Haozhe Zhao, Gang Chen, Yunshui Li, Kangyang Luo, Chuancheng Lv, Kaikai An, Fanchao Qi, Baobao Chang, Maosong Sun

EMNLP 2025 SAC Highlights Award πŸ†

GATEAU: Selecting Influential Samples for Long Context Alignment
GATEAU: Selecting Influential Samples for Long Context Alignment

Shuzheng Si, Haozhe Zhao, Gang Chen, Yunshui Li, Kangyang Luo, Chuancheng Lv, Kaikai An, Fanchao Qi, Baobao Chang, Maosong Sun

EMNLP 2025 SAC Highlights Award πŸ†

InFi-Check: Interpretable and Fine-Grained Fact-Checking of LLMs
InFi-Check: Interpretable and Fine-Grained Fact-Checking of LLMs

Yuzhuo Bai*, Shuzheng Si*, Kangyang Luo, Qingyi Wang, Wenhao Li, Gang Chen, Fanchao Qi, Maosong Sun

KnowFM@ACL 2025 Workshop (Oral, * indicates co-first authors)

InFi-Check: Interpretable and Fine-Grained Fact-Checking of LLMs
InFi-Check: Interpretable and Fine-Grained Fact-Checking of LLMs

Yuzhuo Bai*, Shuzheng Si*, Kangyang Luo, Qingyi Wang, Wenhao Li, Gang Chen, Fanchao Qi, Maosong Sun

KnowFM@ACL 2025 Workshop (Oral, * indicates co-first authors)

Aligning Large Language Models to Follow Instructions and Hallucinate Less via Effective Data Filtering
Aligning Large Language Models to Follow Instructions and Hallucinate Less via Effective Data Filtering

Shuzheng Si, Haozhe Zhao, Gang Chen, Cheng Gao, Yuzhuo Bai, Zhitong Wang, Kaikai An, Kangyang Luo, Chen Qian, Fanchao Qi, Baobao Chang, Maosong Sun

ACL 2025

Aligning Large Language Models to Follow Instructions and Hallucinate Less via Effective Data Filtering
Aligning Large Language Models to Follow Instructions and Hallucinate Less via Effective Data Filtering

Shuzheng Si, Haozhe Zhao, Gang Chen, Cheng Gao, Yuzhuo Bai, Zhitong Wang, Kaikai An, Kangyang Luo, Chen Qian, Fanchao Qi, Baobao Chang, Maosong Sun

ACL 2025

Looking Beyond Text: Reducing Language Bias in Large Vision-Language Models via Multimodal Dual-Attention and Soft-Image Guidance
Looking Beyond Text: Reducing Language Bias in Large Vision-Language Models via Multimodal Dual-Attention and Soft-Image Guidance

Haozhe Zhao*, Shuzheng Si*, Liang Chen, Yichi Zhang, Maosong Sun, Mingjia Zhang, Baobao Chang

EMNLP 2025 (* indicates co-first authors)

Looking Beyond Text: Reducing Language Bias in Large Vision-Language Models via Multimodal Dual-Attention and Soft-Image Guidance
Looking Beyond Text: Reducing Language Bias in Large Vision-Language Models via Multimodal Dual-Attention and Soft-Image Guidance

Haozhe Zhao*, Shuzheng Si*, Liang Chen, Yichi Zhang, Maosong Sun, Mingjia Zhang, Baobao Chang

EMNLP 2025 (* indicates co-first authors)

Rethinking Semantic Parsing for Large Language Models: Enhancing LLM Performance with Semantic Hints
Rethinking Semantic Parsing for Large Language Models: Enhancing LLM Performance with Semantic Hints

Kaikai An*, Shuzheng Si*, Helan Hu, Haozhe Zhao, Yuchi Wang, Qingyan Guo, Baobao Chang

ACL 2025 (* indicates co-first authors)

Rethinking Semantic Parsing for Large Language Models: Enhancing LLM Performance with Semantic Hints
Rethinking Semantic Parsing for Large Language Models: Enhancing LLM Performance with Semantic Hints

Kaikai An*, Shuzheng Si*, Helan Hu, Haozhe Zhao, Yuchi Wang, Qingyan Guo, Baobao Chang

ACL 2025 (* indicates co-first authors)

Improving the Robustness of Distantly-Supervised Named Entity Recognition via Uncertainty-Aware Teacher Learning and Student-Student Collaborative Learning
Improving the Robustness of Distantly-Supervised Named Entity Recognition via Uncertainty-Aware Teacher Learning and Student-Student Collaborative Learning

Shuzheng Si, Helan Hu, Haozhe Zhao, Shuang Zeng, Kaikai An, Zefan Cai, Baobao Chang

ACL (Findings) 2024

Improving the Robustness of Distantly-Supervised Named Entity Recognition via Uncertainty-Aware Teacher Learning and Student-Student Collaborative Learning
Improving the Robustness of Distantly-Supervised Named Entity Recognition via Uncertainty-Aware Teacher Learning and Student-Student Collaborative Learning

Shuzheng Si, Helan Hu, Haozhe Zhao, Shuang Zeng, Kaikai An, Zefan Cai, Baobao Chang

ACL (Findings) 2024

Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation
Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation

Haozhe Zhao*, Zefan Cai*, Shuzheng Si*, Liang Chen, Yufeng He, Kaikai An, Baobao Chang

NAACL 2024 (* indicates co-first authors)

Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation
Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation

Haozhe Zhao*, Zefan Cai*, Shuzheng Si*, Liang Chen, Yufeng He, Kaikai An, Baobao Chang

NAACL 2024 (* indicates co-first authors)

MMICL: Empowering Vision-Language Model with Multi-Modal In-Context Learning
MMICL: Empowering Vision-Language Model with Multi-Modal In-Context Learning

Haozhe Zhao*, Zefan Cai*, Shuzheng Si*, Xiaojian Ma, Kaikai An, Liang Chen, Zixuan Liu, Sheng Wang, Wenjuan Han, Baobao Chang

ICLR 2024 (* indicates co-first authors)

MMICL: Empowering Vision-Language Model with Multi-Modal In-Context Learning
MMICL: Empowering Vision-Language Model with Multi-Modal In-Context Learning

Haozhe Zhao*, Zefan Cai*, Shuzheng Si*, Xiaojian Ma, Kaikai An, Liang Chen, Zixuan Liu, Sheng Wang, Wenjuan Han, Baobao Chang

ICLR 2024 (* indicates co-first authors)

SpokenWOZ: A Large-Scale Speech-Text Benchmark for Spoken Task-Oriented Dialogue Agents
SpokenWOZ: A Large-Scale Speech-Text Benchmark for Spoken Task-Oriented Dialogue Agents

Shuzheng Si, Wentao Ma, Haoyu Gao, Yuchuan Wu, Ting-En Lin, Yinpei Dai, Hangyu Li, Rui Yan, Fei Huang, Yongbin Li

NeurIPS 2023 (reviewer’s score: 9/ 9/ 7/ 6/ 5)

SpokenWOZ: A Large-Scale Speech-Text Benchmark for Spoken Task-Oriented Dialogue Agents
SpokenWOZ: A Large-Scale Speech-Text Benchmark for Spoken Task-Oriented Dialogue Agents

Shuzheng Si, Wentao Ma, Haoyu Gao, Yuchuan Wu, Ting-En Lin, Yinpei Dai, Hangyu Li, Rui Yan, Fei Huang, Yongbin Li

NeurIPS 2023 (reviewer’s score: 9/ 9/ 7/ 6/ 5)

SANTA: Separate Strategies for Inaccurate and Incomplete Annotation Noise in Distantly-Supervised Named Entity Recognition
SANTA: Separate Strategies for Inaccurate and Incomplete Annotation Noise in Distantly-Supervised Named Entity Recognition

Shuzheng Si, Zefan Cai, Shuang Zeng, Guoqiang Feng, Jiaxing Lin, Baobao Chang

ACL (Findings) 2023

SANTA: Separate Strategies for Inaccurate and Incomplete Annotation Noise in Distantly-Supervised Named Entity Recognition
SANTA: Separate Strategies for Inaccurate and Incomplete Annotation Noise in Distantly-Supervised Named Entity Recognition

Shuzheng Si, Zefan Cai, Shuang Zeng, Guoqiang Feng, Jiaxing Lin, Baobao Chang

ACL (Findings) 2023

Mining Clues from Incomplete Utterance: A Query-Enhanced Network for Incomplete Utterance Rewriting
Mining Clues from Incomplete Utterance: A Query-Enhanced Network for Incomplete Utterance Rewriting

Shuzheng Si, Shuang Zeng, Baobao Chang

NAACL 2022

Mining Clues from Incomplete Utterance: A Query-Enhanced Network for Incomplete Utterance Rewriting
Mining Clues from Incomplete Utterance: A Query-Enhanced Network for Incomplete Utterance Rewriting

Shuzheng Si, Shuang Zeng, Baobao Chang

NAACL 2022

SCL-RAI: Span-based Contrastive Learning with Retrieval Augmented Inference for Unlabeled Entity Problem in NER
SCL-RAI: Span-based Contrastive Learning with Retrieval Augmented Inference for Unlabeled Entity Problem in NER

Shuzheng Si, Shuang Zeng, Jiaxing Lin, Baobao Chang

COLING 2022

SCL-RAI: Span-based Contrastive Learning with Retrieval Augmented Inference for Unlabeled Entity Problem in NER
SCL-RAI: Span-based Contrastive Learning with Retrieval Augmented Inference for Unlabeled Entity Problem in NER

Shuzheng Si, Shuang Zeng, Jiaxing Lin, Baobao Chang

COLING 2022