Google
May 24, 2024Abstract page for arXiv paper 2405.15877: Basis Selection: Low-Rank Decomposition of Pretrained Large Language Models for Target Applications.
May 24, 2024We propose Basis Selection, a low-rank decomposition approach to compress pretrained large language models for target applications. Basis�...
A two-stage model-compression method to reduce a model's inference time cost by first decomposing the matrices in the model into smaller matrices and then�...
May 24, 2024• We propose Basis Selection, a low-rank decomposition approach to compress pretrained large language models for target applications. Basis�...
Sep 11, 2024Our method focuses on identifying and removing these redundant parts, retaining only the necessary elements for the target applications.
May 27, 2024This paper presents a method for compressing and adapting large pre-trained language models to specific target applications.
Basis Selection: Low-Rank Decomposition of Pretrained Large Language Models for Target Applications.
Jun 16, 2024Our strategy uses a novel regular- ization to enable the masking to comply with the SVD property where the ranks have sorted singular values.
Cited by ; Basis Selection: Low-Rank Decomposition of Pretrained Large Language Models for Target Applications. Y Li, C Zhao, H Lee, E Chang, Y Shi, V Chandra.
We propose Basel, a low-rank decomposition approach to compress pretrained large language models for target applications. Basel identifies the beneficial�...