avatar

Yu Zhang

Researcher
Moonshot AI
yzhang.cs (at) outlook.com


Hi, my name is Yu Zhang ([jy tʃɑŋ], 張宇/张宇 in traditional/simplified Chinese). I am a researcher at Moonshot AI.

I received my Ph.D. degree from Soochow University in 2025, advised by Prof. Guohong Fu. Prior to this, I received my M. Eng. and B. Eng. degrees from Soochow University in 2021 and 2018, respectively.

My early research focused on structured prediction tasks, specifically dependency parsing and constituency parsing. Currently, my research interests have evolved to focus on developing efficient text generation models. I am particularly intrigued by the prospect of developing hardware-efficient methods for linear-time sequence modeling. As a disciple of parallel programming, I am passionate about exploring techniques that harness the power of parallel computing to develop scalable subquadratic models.

Publications

* denotes equal contributions

Kimi Linear: An Expressive, Efficient Attention Architecture
Yu Zhang, Zongyu Lin, Xingcheng Yao, Jiaxi Hu, Fanqing Meng, Chengyin Liu, Xin Men, Songlin Yang, Zhiyuan Li, Wentao Li, Enzhe Lu, Weizhou Liu, Yanru Chen, Weixin Xu, Longhui Yu, Yejie Wang, Yu Fan, Longguang Zhong, Enming Yuan, Dehao Zhang, Yizhi Zhang, TY Liu, Haiming Wang, Shengjun Fang, Weiran He, Shaowei Liu, Yiwei Li, Jianlin Su, Jiezhong Qiu, Bo Pang, Junjie Yan, Zhejun Jiang, Weixiao Huang, Bohong Yin, Jiacheng You, Chu Wei, Zhengtao Wang, Chao Hong, Yutian Chen, Guanduo Chen, Yucheng Wang, Huabin Zheng, Feng Wang, Yibo Liu, Mengnan Dong, Zheng Zhang, Siyuan Pan, Wenhao Wu, Yuhao Wu, Longyu Guan, Jiawen Tao, Guohong Fu, Xinran Xu, Yuzhi Wang, Guokun Lai, Yuxin Wu, Xinyu Zhou, Zhilin Yang, Yulun Du
Preprint
arxiv code citation
Gated Slot Attention for Efficient Linear-Time Sequence Modeling
Yu Zhang*, Songlin Yang*, Ruijie Zhu, Yue Zhang, Leyang Cui, Yiqiao Wang, Bolun Wang, Freda Shi, Bailin Wang, Wei Bi, Peng Zhou, Guohong Fu
NeurIPS 2024
arxiv code citation
Parallelizing Linear Transformers with the Delta Rule over Sequence Length
Songlin Yang, Bailin Wang, Yu Zhang, Yikang Shen, Yoon Kim
NeurIPS 2024
arxiv code citation
Scalable MatMul-free Language Modeling
Ruijie Zhu, Yu Zhang, Ethan Sifferman, Tyler Sheaves, Yiqiao Wang, Dustin Richmond, Peng Zhou, Jason K. Eshraghian
Preprint
arxiv code citation
Non-autoregressive Text Editing with Copy-aware Latent Alignments
Yu Zhang*, Yue Zhang*, Leyang Cui, Guohong Fu
EMNLP 2023
arxiv code citation
Semantic Role Labeling as Dependency Parsing: Exploring Latent Tree Structures Inside Arguments
Yu Zhang, Qingrong Xia, Shilin Zhou, Yong Jiang, Guohong Fu, Min Zhang
COLING 2022
arxiv code citation
Fast and Accurate End-to-End Span-based Semantic Role Labeling as Word-based Graph Parsing Best Paper Award
Shilin Zhou, Qingrong Xia, Zhenghua Li, Yu Zhang, Yu Hong, Min Zhang
COLING 2022
arxiv code citation
Fast and Accurate Neural CRF Constituency Parsing
Yu Zhang*, Houquan Zhou*, Zhenghua Li
IJCAI 2020
arxiv code citation
Efficient Second-Order TreeCRF for Neural Dependency Parsing
Yu Zhang, Zhenghua Li, Min Zhang
ACL 2020
arxiv code citation
Is POS Tagging Necessary or Even Helpful for Neural Dependency Parsing? Best Paper Award
Houquan Zhou*, Yu Zhang*, Zhenghua Li, Min Zhang
NLPCC 2020
arxiv code citation
HLT@SUDA at SemEval-2019 Task 1: UCCA Graph Parsing as Constituent Tree Parsing
Wei Jiang, Zhenghua Li, Yu Zhang, Min Zhang
SemEval 2019
arxiv code citation

Projects

FLA
A Triton-Based Library for Hardware-Efficient Implementations of Linear Attention Mechanism
SuPar
State-of-the-art syntactic/semantic parsers. A Python package designed for structured prediction, including reproductions of many state-of-the-art syntactic/semantic parsers (with pretrained models for more than 19 languages), and highly-parallelized implementations of several well-known structured prediction algorithms.

Experience

2025 - present
AI Researcher at Moonshot AI
2024 - 2025
Research Intern at Shanghai AI Lab
mentored by Peng Gao
2023 - 2024
Research Intern at Tencent AI Lab
mentored by Wei Bi
2020 - 2021
Research Intern at Alibaba DAMO Academy
mentored by Yong Jiang

Last updated: