Post-training Neural Network (NN) model compression is an attractive approach for deploying large, memory-consuming models on devices with limited memory resources. In this study, we investigate the rate-distortion tradeoff for NN model compression.
Articles
Related Articles
May 27, 2021
Smartphone Location Recognition with Unknown Modes in Deep Feature Space
Smartphone location recognition aims to identify the location of a smartphone on a user in specific...
Read More >
1 MIN READING
February 26, 2024
LINEAR LOG-NORMAL ATTENTION WITH UNBIASED CONCENTRATION
Transformer models have achieved remarkable results in a wide range of applications. However, their scalability is...
Read More >
1 MIN READING
March 15, 2024
Does the Performance of Text-to-Image Retrieval Models Generalize Beyond Captions-as-a-Query?
Text-image retrieval (T2I) refers to the task of recovering all images relevant to a keyword query....
Read More >
1 MIN READING