Unstructured pruning and low rank factorisation of self-supervised pre-trained speech models
Release time: 2024/09/28
Hits:
- DOI number:
- 10.1109/JSTSP.2024.3433616
- Journal:
- IEEE Journal of Selected Topics in Signal Processing
- Abstract:
- Self-supervised pre-trained speech models require significant memory and computational resources, limiting their applicability to many speech tasks. Unstructured pruning is a compression method that can achieve minimal performance degradation, while the resulting sparse matrix mandates special hardware or computational operators for acceleration. In this study, we propose a novel approach that leverages the potential low-rank structures of the unstructured sparse matrices by applying truncated singular value decomposition (SVD), thus converting them into parameter-efficient dense models. Moreover, we introduce nuclear norm regularisation to ensure lower rank and a learnable singular value selection strategy to determine the approximate truncation rate for each matrix. Experiments on multiple speech tasks demonstrate that the proposed method can convert an unstructured sparse model into a light-weight and hardware-friendly dense model with comparable or superior performance.
- First Author:
- Haoyu Wang
- Indexed by:
- Journal paper
- Correspondence Author:
- Wei-Qiang Zhang
- Translation or Not:
- no
- Date of Publication:
- 2024/07/25
- Links to published journals:
- https://ieeexplore.ieee.org/document/10609479