chienyu_cse_head_shot_0281_corp.jpg

Photo taken in UW Gates Center, Nov 2024

Chien-Yu Lin 林建宇

Research Scientist
FAIR, Meta Superintelligence Labs (MSL)
Menlo Park, CA


About me

I’m a Research Scientist in the FAIR group at Meta Superintelligence Labs. My research focuses on making machine learning more efficient through algorithm-system co-design.

I received my Ph.D. in 2025 from the Paul G. Allen School of Computer Science & Engineering at the University of Washington. I was fortunate to be advised by Luis Ceze and to work closely with Baris Kasikci and Arvind Krishnamurthy. During my time at UW, I developed efficient methods for a range of ML workloads, including GNNs, NeRFs, and LLMs. Before UW, I earned my B.S. and M.S. in Electronics Engineering from National Yang Ming Chiao Tung University (formerly NCTU), where I worked on sparse CNN accelerator design under the guidance of Prof. Bo-Cheng Lai.

Outside research, I enjoy tennis, hiking, and (backcountry) skiing. I have completed human-powered ascents and ski descents on four of Washington state’s five volcanoes (all except Glacier Peak). In the summer of 2025, I completed a 1,000-mile bike trip from Seattle to Menlo Park to celebrate my graduation.

My current research focus is on automatically discovering next-generation efficient LLMs with AI agents.

News

Jan 26, 2026 One paper (TeleRAG) got accepted to MLSys 2026 and two papers (Composer and UniQL) got accepted to ICLR 2026.
Jul 20, 2025 Joined Meta at the MPK!
Apr 24, 2025 I will attend ICLR in Singapore to present, Palu our awesome low-rank KV-Cache compression paper. Come find me for a coffee chat!

Selected publications


* means equal contribution

  1. telerag.png
    TeleRAG: Efficient Retrieval-Augmented Generation Inference with Lookahead Retrieval
    Chien-Yu Lin*, Keisuke Kamahori*, Yiyu Liu, and 11 more authors
    2026
  2. palu_concept.png
    Palu: Compressing KV-Cache with Low-Rank Projection
    Chi-Chih Chang*, Wei-Cheng Lin*Chien-Yu Lin*, and 7 more authors
    In Proceedings of International Conference on Learning Representations (ICLR), 2025
  3. atom2.png
    Atom: Low-Bit Quantization for Efficient and Accurate LLM Serving
    Yilong Zhao, Chien-Yu Lin, Kan Zhu, and 7 more authors
    In Proceedings of Machine Learning and Systems (MLSys), 2024
  4. FastSR-NeRF: Improving NeRF Efficiency on Consumer Devices with A Simple Super-Resolution Pipeline
    Chien-Yu Lin, Qichen Fu, Thomas Merth, and 2 more authors
    In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Jan 2024
    Oral (Top 2.6%)
  5. SPIN: An Empirical Evaluation on Sharing Parameters of Isotropic Networks
    Chien-Yu Lin*, Anish Prabhu*, Thomas Merth, and 4 more authors
    In Proceedings the 17th European Conference on Computer Vision (ECCV), Jan 2022
  6. Supporting compressed-sparse activations and weights on SIMD-like accelerator for sparse convolutional neural networks
    Chien-Yu Lin, and Bo-Cheng Lai
    In Proceedings of the 23rd Asia and South Pacific Design Automation Conference (ASP-DAC), Jan 2018