Screenshot 2024-10-01 at 2.26.53 PM.png

张 培源

<aside> <img src="https://prod-files-secure.s3.us-west-2.amazonaws.com/44aa5dbc-d55c-477f-9d96-25e2fe4ed8c1/cd361d2a-9856-49d3-bfa3-abc0d900c2fa/25231.png" alt="https://prod-files-secure.s3.us-west-2.amazonaws.com/44aa5dbc-d55c-477f-9d96-25e2fe4ed8c1/cd361d2a-9856-49d3-bfa3-abc0d900c2fa/25231.png" width="40px" /> Github

</aside>

<aside> <img src="https://prod-files-secure.s3.us-west-2.amazonaws.com/44aa5dbc-d55c-477f-9d96-25e2fe4ed8c1/f30b1dd3-7384-4ae2-b912-11f36f7e174f/Picture2.png" alt="https://prod-files-secure.s3.us-west-2.amazonaws.com/44aa5dbc-d55c-477f-9d96-25e2fe4ed8c1/f30b1dd3-7384-4ae2-b912-11f36f7e174f/Picture2.png" width="40px" /> G Scholar

</aside>

<aside> <img src="https://prod-files-secure.s3.us-west-2.amazonaws.com/44aa5dbc-d55c-477f-9d96-25e2fe4ed8c1/4a681ff9-9560-425e-a09b-83cefa5ab4e8/twitter-3.png" alt="https://prod-files-secure.s3.us-west-2.amazonaws.com/44aa5dbc-d55c-477f-9d96-25e2fe4ed8c1/4a681ff9-9560-425e-a09b-83cefa5ab4e8/twitter-3.png" width="40px" /> Twitter

</aside>

<aside> 📧 Email

</aside>

<aside> 💬 WeChat

</aside>

👋 Hi there!

I’m Peiyuan (Perry) Zhang, a second-year PhD student in Computer Science at UC San Diego, advised by Prof. Hao Zhang. My research focuses on efficient architectures for video generation, as well as their post-training and inference.

Previously, I worked on LLMs and VLMs at SUTD and NTU, and interned at ByteDance Seedance team. I enjoy building scalable systems and exploring new ideas at the intersection of machine learning, vision, and efficiency. I believe that great ML scientists are, fundamentally, exceptional software engineers.

🌋 News


💡 Projects


FastVideo

A unified inference and post-training framework for accelerated video generation.

TinyLlama

The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.

lmms-eval

One-for-all LMMs evaluation package.

EasyContext

LLM long context training made simple.

📜 Selected Publications

(*: equal contribution)


Attn-QAT: 4-Bit Attention With Quantization-Aware Training

Peiyuan Zhang*,Matthew Noto*, Wenxuan Tan*, Chengquan Jiang,Will Lin,Wei Zhou,Hao Zhang

[paper]

VSA: Faster Video Diffusion with Trainable Sparse Attention

Peiyuan Zhang*, Yongqi Chen*, Haofeng Huang*, Will Lin, Zhengzhong Liu, Ion Stoica, Eric Xing, Hao Zhang

Neurips 2025.[paper]

Fast Video Generation with Sliding Tile Attention

Peiyuan Zhang, Yongqi Chen, Runlong Su, Hangliang Ding, Ion Stoica, Zhengzhong Liu, Hao Zhang

ICML 2025.[paper]

Long Context Transfer from Language to Vision

Peiyuan Zhang*, Kaichen Zhang*, Bo Li*, Guangtao Zeng, Jingkang Yang, Yuanhan Zhang, Ziyue Wang, Haoran Tan, Chunyuan Li, Ziwei Liu

TMLR .[paper]

One Network, Many Masks: Towards More Parameter-Efficient Transfer Learning

Guangtao Zeng*, Peiyuan Zhang*, Wei Lu

ACL 2023 Long Paper. [paper]

Better Few-Shot Relation Extraction with Label Prompt Dropout

Peiyuan Zhang, Wei Lu

EMNLP 2022 Long Paper. [paper]

🔬 Experience


UC San Diego

PhD Student, with Prof. Hao Zhang

ByteDance Seed, San Jose, 06/2025–12/2025

Research Intern, with Xiaonan Nie, Guoqiang Wei, and Yan Zeng, Seedance Architecture

Nanyang Technological University

Reseach Assistant, with Prof. Ziwei Liu , Vision-Language Models.

Singapore University of Technology and Design, 05/2019–09/2022

Undergrad and Reseach Assistant, with Prof. Wei Lu, Language Models.

Singapore Agency of Science, Technology and Research, 05/2020–09/2020

Research Intern, DeepFake Detection.