Skip to main content
University of Oxford logo Home

Search form

  • Log in
  • Members
  • About Us
    • Contact Us
    • Travel & Maps
    • Our Building
    • Supporting Mathematics
    • Alumni
    • History
    • Art and Oxford Mathematics
    • Equality, Diversity and Inclusion
    • News
    • Vacancies
  • Study Here
    • Undergraduate Study
    • Postgraduate Study
    • Current Students
  • Research
    • Research Groups
    • Case Studies
    • Faculty Books
  • Outreach
    • Posters
    • Oxford Mathematics Alphabet
    • Oxford Online Maths Club
    • Oxford Maths Festival
    • It All Adds Up
    • Problem Solving Matters
    • MIORPA
    • PROMYS Europe
    • Oxfordshire Maths Masterclasses
    • Outreach Information
    • Mailing List
  • People
    • Key Contacts
    • People List
    • A Global Department
    • Research Fellowship Programmes
    • Professional Services Teams
  • Events
    • Venue Hire
    • Public Lectures & Events
    • Departmental Seminars & Events
    • Special Lectures
    • Conferences
    • Summer Schools
    • Past Events
    • Info for Event Organisers & Attendees

Primary tabs

  • View
  • Contact
PROFILE

Dr Shiwei Liu

Ph.D.
Pronouns
He / Him
Status
Research Fellow
+44 1865 270744
Contact form
https://shiweiliuiiiiiii.github.io/
ORCID iD
https://orcid.org/https://orcid.org/0009-0001-1255-4436
Research groups
  • Machine Learning and Data Science
  • Numerical Analysis
Address
Mathematical Institute
University of Oxford
Andrew Wiles Building
Radcliffe Observatory Quarter
Woodstock Road
Oxford
OX2 6GG
Major / recent publications

Adriana Fernandez-Lopez, Honglie Chen, Pingchuan Ma, Lu Yin, Qiao Xiao, Stavros Petridis,  Shiwei Liu,  Maja Pantic.  MSRS: Training Multimodal Speech Recognition Models from Scratch with Sparse Mask Optimization. Interspeech, 2024.

Qiao Xiao, Pingchuan Ma, Adriana Fernandez-Lopez, Boqian Wu, Lu Yin, Stavros Petridis, Mykola Pechenizkiy, Maja Pantic, Decebal Constantin Mocanu, Shiwei Liu. Dynamic Data Pruning for Automatic Speech Recognition. Interspeech, 2024.

Zhang, Zhenyu, Runjin Chen, Shiwei Liu, Zhewei Yao, Olatunji Ruwase, Beidi Chen, Xiaoxia Wu, and Zhangyang Wang. "Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding." arXiv preprint arXiv:2403.04797 (2024).

Yin, Lu, You Wu, Zhenyu Zhang, Cheng-Yu Hsieh, Yaqing Wang, Yiling Jia, Mykola Pechenizkiy, Yi Liang, Zhangyang Wang, and Shiwei Liu. "Outlier weighed layerwise sparsity (owl): A missing secret sauce for pruning llms to high sparsity." The Forty-first International Conference on
Machine Learning (ICML), 2024.

Lu Yin, Ajay Jaiswal, Shiwei Liu, Souvik Kundu, and Zhangyang Wang. "Pruning Small Pre-Trained Weights Irreversibly and Monotonically Impairs Difficult Downstream Tasks in LLMs." The Forty-first International Conference on Machine Learning (ICML), 2024.

Yuxin Zhang, Yuxuan Du, Gen Luo, Yunshan Zhong, Zhenyu Zhang, Shiwei Liu, Rongrong Ji. "CaM: Cache Merging for Memory-efficient LLMs Inference." The Forty-first International Conference on Machine Learning (ICML), 2024.

Jie Ji, Gen Li, Lu Yin, Minghai Qin, Geng Yuan, Linke Guo, Shiwei Liu, Xiaolong Ma. Advancing Dynamic Sparse Training by Exploring Optimization Opportunities. The Forty-first International Conference on Machine Learning (ICML), 2024.

Zhangheng Li, Shiwei Liu, Tianlong Chen, Ajay Kumar Jaiswal, Zhenyu Zhang, Dilin Wang, Raghuraman Krishnamoorthi, Shiyu Chang, Zhangyang Wang. Sparse Cocktail: Co-Training Many Sparsity Patterns and Ratios at Once. The Forty-first International Conference on Machine
Learning (ICML), 2024.

Zhang, Yuxin, Lirui Zhao, Mingbao Lin, Yunyun Sun, Yiwu Yao, Xingjia Han, Jared Tanner, Shiwei Liu, and Rongrong Ji. "Dynamic sparse no training: Training-free fine-tuning for sparse llms."  In The Twelfth International Conference on Learning Representations. 2024.

Li, Gen, Lu Yin, Jie Ji, Wei Niu, Minghai Qin, Bin Ren, Linke Guo, Shiwei Liu, and Xiaolong Ma. "NeurRev: Train Better Sparse Neural Network Practically via Neuron Revitalization." In The Twelfth International Conference on Learning Representations. 2024.

Yang, Enneng, Zhenyi Wang, Li Shen, Shiwei Liu, Guibing Guo, Xingwei Wang, and Dacheng Tao. "Adamerging: Adaptive model merging for multi-task learning." In The Twelfth International Conference on Learning Representations. 2024.

Pham, Hoang, Shiwei Liu, Lichuan Xiang, Dung Le, Hongkai Wen, and Long Tran-Thanh. "Towards Data-Agnostic Pruning At Initialization: What Makes a Good Sparse Mask?." Advances in Neural Information Processing Systems 36 (2023).

Hoang, Duc, Souvik Kundu, Shiwei Liu, and Zhangyang Wang. "Don’t just prune by magnitude! Your mask topology is a secret weapon." Advances in Neural Information Processing Systems 36 (2023): 65056-65068.

Yin, Lu, Gen Li, Meng Fang, Li Shen, Tianjin Huang, Zhangyang Wang, Vlado Menkovski, Xiaolong Ma, Mykola Pechenizkiy, and Shiwei Liu. "Dynamic sparsity is channel-level sparsity learner." Advances in Neural Information Processing Systems 36 (2023).

Jaiswal, Ajay, Shiwei Liu, Tianlong Chen, and Zhangyang Wang. "The emergence of essential sparsity in large pre-trained models: The weights that matter." Advances in Neural Information Processing Systems 36 (2023).

Huang, Tianjin, Lu Yin, Zhenyu Zhang, Li Shen, Meng Fang, Mykola Pechenizkiy, Zhangyang Wang, and Shiwei Liu. "Are large kernels better teachers than transformers for convnets?." In International Conference on Machine Learning, pp. 14023-14038. PMLR, 2023.

Jaiswal, Ajay Kumar, Shiwei Liu, Tianlong Chen, Ying Ding, and Zhangyang Wang. "Instant soup: Cheap pruning ensembles in a single pass can draw lottery tickets from large models." In International Conference on Machine Learning, pp. 14691-14701. PMLR, 2023.

Shiwei Liu, Tianlong Chen, Zhenyu Zhang, Xuxi Chen, Tianjin Huang, Ajay Jaiswal, and Zhangyang Wang. "Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!." ICLR 2023.

Shiwei Liu, Tianlong Chen, Xiaohan Chen, Xuxi Chen, Qiao Xiao, Boqian Wu, Tommi Kärkkäinen, Mykola Pechenizkiy, Decebal Mocanu, and Zhangyang Wang. "More convnets in the 2020s: Scaling up kernels beyond 51x51 using sparsity."  ICLR 2023.

Shiwei Liu, Tianlong Chen, Xiaohan Chen, Li Shen, Decebal Constantin Mocanu, Zhangyang Wang, and Mykola Pechenizkiy. "The unreasonable effectiveness of random pruning: Return of the most naive baseline for sparse training." ICLR 2022.

Shiwei Liu, Tianlong Chen, Xiaohan Chen, Zahra Atashgahi, Lu Yin, Huanyu Kou, Li Shen, Mykola Pechenizkiy, Zhangyang Wang, and Decebal Constantin Mocanu. "Sparse training via boosting pruning plasticity with neuroregeneration." Advances in Neural Information Processing Systems 34 (2021): 9908-9922.

Shiwei Liu, Lu Yin, Decebal Constantin Mocanu, and Mykola Pechenizkiy. "Do we actually need dense over-parameterization? in-time over-parameterization in sparse training." In International Conference on Machine Learning, pp. 6989-7000. PMLR, 2021.

Preferred address

Mathematical Institute
University of Oxford
Oxford
OX2 6GG

Further details

Shiwei Liu is a Royal Society Newton International Fellow at the University of Oxford. Previously,  he was a postdoctoral fellow at UT Austin and IFML. He obtained his Ph.D. at the Eindhoven University of Technology (TU/e), the Netherlands. 

I am open to collaborating with remote students and researchers who are interested in deep learning, efficient training/inference of LLMs, learning with sparsity, architecture in Deep Learning, etc. Feel free to drop me an email if you are interested. 

Recent publications
Dynamic Sparse No Training: Training-Free Fine-tuning for Sparse LLMs
Zhang, Y Zhao, L Lin, M Sun, Y Yao, Y Han, X Tanner, J Liu, S Ji, R (13 Oct 2023) http://arxiv.org/abs/2310.08915v3
Research interests

Deep Learning

Machine Learning

Learning with Sparsity

Large Language Models

Architecture of Deep Learning

Prizes, awards, and scholarships

Rising Star in AI, KAUST, 01/2024

Rising Star Award, Conference on Parsimony and Learning (CPAL), 10/2023

Best Ph.D. Dissertation Award Runner-Up, Informatics Europe, 10/2023

Newton International Fellowship Award, Royal Society & British Academy, 9/2023

Best Paper Award, Learning on Graphs Conference (LoG 2022), 11/2022

Cum Laude (Distinguished Ph.D. thesis), Eindhoven University of Technology, NL, 5% 4/2022

Facebook LinkedIn Bluesky X
TikTok Instagram YouTube
London Mathematical Society Good Practice Scheme Athena SWAN Silver Award (ECU Gender Charter) Stonewall Silver Employer 2022

© Mathematical Institute

Accessibility Statement


Privacy Policy

Cookies

sfy39587stp18