Theo dõi
Shiwei Liu
Shiwei Liu
University of Oxford & Eindhoven University of Technology
Email được xác minh tại maths.ox.ac.uk - Trang chủ
Tiêu đề
Trích dẫn bởi
Trích dẫn bởi
Năm
More convnets in the 2020s: Scaling up kernels beyond 51x51 using sparsity
S Liu, T Chen, X Chen, X Chen, Q Xiao, B Wu, M Pechenizkiy, D Mocanu, ...
ICLR2023, The International Conference on Learning Representations, 2023
2212023
Do we actually need dense over-parameterization? in-time over-parameterization in sparse training
S Liu, L Yin, DC Mocanu, M Pechenizkiy
ICML2021, International Conference on Machine Learning, 2021
1462021
Sparse Training via Boosting Pruning Plasticity with Neuroregeneration
S Liu, T Chen, X Chen, Z Atashgahi, L Yin, H Kou, L Shen, M Pechenizkiy, ...
NeurIPS2021, Advances in Neural Information Processing Systems, 2021
1392021
The unreasonable effectiveness of random pruning: Return of the most naive baseline for sparse training
S Liu, T Chen, X Chen, L Shen, DC Mocanu, Z Wang, M Pechenizkiy
ICLR2022, The International Conference on Learning Representations, 2022
1302022
Adamerging: Adaptive model merging for multi-task learning
E Yang, Z Wang, L Shen, S Liu, G Guo, X Wang, D Tao
ICLR2024, The International Conference on Learning Representations, 2024
1042024
Sparse evolutionary deep learning with over one million artificial neurons on commodity hardware
S Liu, DC Mocanu, ARR Matavalam, Y Pei, M Pechenizkiy
Neural Computing and Applications 33, 2589-2604, 2021
1032021
Selfish sparse RNN training
S Liu, DC Mocanu, Y Pei, M Pechenizkiy
ICML2021, International Conference on Machine Learning, 2021
93*2021
Deep Ensembling with No Overhead for either Training or Testing: The All-Round Blessings of Dynamic Sparsity
S Liu, T Chen, Z Atashgahi, X Chen, G Sokar, E Mocanu, M Pechenizkiy, ...
ICLR2022, The International Conference on Learning Representations, 2021
612021
Outlier weighed layerwise sparsity (owl): A missing secret sauce for pruning llms to high sparsity
L Yin, Y Wu, Z Zhang, CY Hsieh, Y Wang, Y Jia, G Li, A Jaiswal, ...
ICML2024, International Conference on Machine Learning, 2024
602024
Dynamic sparse no training: Training-free fine-tuning for sparse llms
Y Zhang, L Zhao, M Lin, Y Sun, Y Yao, X Han, J Tanner, S Liu, R Ji
ICLR2024, The International Conference on Learning Representations, 2024
492024
Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers
T Chen, Z Zhang, A Jaiswal, S Liu, Z Wang
ICLR2023, The International Conference on Learning Representations, 2023
462023
The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter
A Jaiswal, S Liu, T Chen, Z Wang
NeurIPS2023, 37th Annual Conference on Neural Information Processing Systems, 2023
402023
Revisiting pruning at initialization through the lens of Ramanujan graph
DNM Hoang, S Liu, R Marculescu, Z Wang
ICLR2023, The International Conference on Learning Representations, 2023
382023
Achieving personalized federated learning with sparse local models
T Huang, S Liu, L Shen, F He, W Lin, D Tao
arXiv preprint arXiv:2201.11380, 2022
372022
Topological Insights into Sparse Neural Networks
S Liu, T Van der Lee, A Yaman, Z Atashgahi, D Ferraro, G Sokar, ...
ECML2020, European Conference on Machine Learning, 2020
362020
Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!
S Liu, T Chen, Z Zhang, X Chen, T Huang, A Jaiswal, Z Wang
ICLR2023, The International Conference on Learning Representations, 2023
342023
Dynamic Sparse Network for Time Series Classification: Learning What to “See”
Q Xiao, B Wu, Y Zhang, S Liu, M Pechenizkiy, E Mocanu, DC Mocanu
NeurIPS2022, 36th Annual Conference on Neural Information Processing Systems, 2022
322022
A Brain-inspired Algorithm for Training Highly Sparse Neural Networks
Z Atashgahi, J Pieterse, S Liu, DC Mocanu, R Veldhuis, M Pechenizkiy
Machine Learning Journal (ECML-PKDD 2022 journal track), 2019
30*2019
Visual prompting upgrades neural network sparsification: A data-model perspective
C Jin, T Huang, Y Zhang, M Pechenizkiy, S Liu, S Liu, T Chen
arXiv preprint arXiv:2312.01397, 2023
292023
Found in the middle: How language models use long contexts better via plug-and-play positional encoding
Z Zhang, R Chen, S Liu, Z Yao, O Ruwase, B Chen, X Wu, Z Wang
NeurIPS 2024, 2024
252024
Hệ thống không thể thực hiện thao tác ngay bây giờ. Hãy thử lại sau.
Bài viết 1–20