Hi there! I am Tianyu Liu (刘天宇), a fourth-year joint Ph.D. student at USTC and Shanghai AI Laboratory, supervised by Sun Xiao.

My research interests mainly lie in efficient inference for LLMs. I am working on speculative decoding, a promising technique for accelerating LLM inference, and I am open to collaborations on related inference topics.

I expect to graduate in 2027 and am currently seeking job opportunities in efficient LLM inference and AI systems. Please feel free to contact me by Email.

🔥 News

📝 Selected Publications

: corresponding author; *: equal contribution

ICLR 2025
PEARL illustration

Featured PEARL: Parallel Speculative Decoding with Adaptive Draft Length

Tianyu Liu, Yun Li, Qitan Lv, Kai Liu, Jianchen Zhu, Winston Hu, Xiao Sun

[Paper]  [Project]  [Code]


🚀 Introduce nano-PEARL: a Draft-Target Disaggregation LLM Serving System via Parallel Speculative Decoding!

Preprint
TALON illustration

TALON: Confidence-Aware Speculative Decoding with Adaptive Token Trees

Tianyu Liu*, Qitan Lv*, Yuhao Shen, Xiao Sun, Xiaoyan Sun

[Paper]

ACL 2026 Findings
LogitSpec illustration

LogitSpec: Accelerating Retrieval-based Speculative Decoding via Next Next Token Speculation

Tianyu Liu*, Qitan Lv*, Hao Li, Xing Gao, Xiao Sun, Xiaoyan Sun

[Paper]  [Code]

NeurIPS 2023
REST illustration

Learning Rule-Induced Subgraph Representations for Inductive Relation Prediction

Tianyu Liu, Qitan Lv, Jie Wang, Shuling Yang, Hanzhu Chen

[Paper]  [Code]

🗂 Other Publications

arXiv 2025 HIPPO: Accelerating Video Large Language Models Inference via Holistic-aware Parallel Speculative Decoding

Qitan Lv*, Tianyu Liu*, Wen Wu, Xuenan Xu, Bowen Zhou, Feng Wu, Chao Zhang

[Paper]

arXiv 2025 KALE: Enhancing Knowledge Manipulation in Large Language Models via Knowledge-aware Learning

Qitan Lv*, Tianyu Liu*, Qiaosheng Zhang, Xingcheng Xu, Chaochao Lu

[Paper]

ACL 2026 Main (Oral) Double: Breaking the Acceleration Limit via Double Retrieval Speculative Parallelism

Yuhao Shen, Tianyu Liu, Junyi Shen, Jinyang Wu, Quan Kong, Li Huan, Cong Wang

[Paper]

ICLR 2026 Speculative Decoding via Hybrid Drafting and Rollback-Aware Branch Parallelism

Yuhao Shen, Junyi Shen, Quan Kong, Tianyu Liu, Yao Lu, Cong Wang

[Paper]

NAACL 2025 Exploiting Edited Large Language Models as General Scientific Optimizers

Qitan Lv*, Tianyu Liu*, Hong Wang

[Paper]

📖 Educations

  • 2022.09 - Present, Ph.D. in Department of Electronic Engineering and Information Science, University of Science and Technology of China.
  • 2018.09 - 2022.06, B.Sc. in Computer Science and Technology, Central University of Finance and Economics.

🖥️ Professional Services

  • Conference reviewer for ICLR'25, ICLR'26, WWW'25, NeurIPS'25.