Xiangyu Li

Ph.D. Student | Tsinghua University | Mobile Computing | LLM Inference

profile/xiangyu_1.jpg

I am a 3rd year Ph.D. student at the Institute for AI Industry Research (AIR), Tsinghua University, advised by Prof. Yunxin Liu. Prior to this, I received my B.Eng. degree from the Department of Electronic Engineering, Tsinghua Univeristy (2022/06).

My research interest lies in the intersection of mobile computing and efficient deep learning (prior work on efficient and adaptive memory management: FlexNN). I am currently focused on efficiently deploying LLMs (Large Language Models) on edge/mobile devices.

Before joining AIR, I researched on Graph Mining System in the NICS-EFC group led by Prof. Yu Wang (2020/01~2021/06), and developed my interest in system research eversince. I also worked as a summer intern at ByteDance (2021/06~2021/09).

Apart from research and internship, I am proud to have served my peers as a member of the Student Association for Science and Technology, Dept. of EE (EESAST) for 3 years.

Feel free to contact: lixiangy22@mails.tsinghua.edu.cn

News 📢

Oct 13, 2024 The code of our MobiCom 2024 paper “FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices” is now available on GitHub. :sparkles: [code]
Jun 02, 2024 Our MobiCom 2024 paper “FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices” is now available in the ACM Digital Library. :book: [pdf]
Apr 28, 2024 Our MobiCom 2024 paper “FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices” has been awarded all the 4 badges: “Artifacts Available”, “Artifacts Evaluated - Functional”, “Artifacts Evaluated - Reusable”, and “Results Replicated” in MobiCom 2024 Artifact Evaluation! :medal_sports:
Jan 10, 2024 Our position & survey paper on Mobile LLM Agents “Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security” is released. :sparkles: [arXiv] [GitHub] [机器之心]
Nov 22, 2023 Our paper “FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices” is conditionally accepted by MobiCom 2024. Thanks to all the coauthors: Yuanchun Li, Yuanzhe Li, Ting Cao and Yunxin Liu! :clap:

Selected Publications 📖

  1. FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices
    Xiangyu Li , Yuanchun Li, Yuanzhe Li, Ting Cao, and Yunxin Liu
    In Proceedings of the 30th Annual International Conference on Mobile Computing and Networking, Washington D.C., DC, USA, 2024
  2. Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
    Yuanchun Li, Hao Wen, Weijun Wang,  Xiangyu Li , Yizhen Yuan, Guohong Liu, Jiacheng Liu, Wenxing Xu, Xiang Wang, Yi Sun, and 15 more authors
    arXiv preprint arXiv:2401.05459, 2024