TY - GEN
T1 - Decoding Compressed Trust
T2 - 41st International Conference on Machine Learning (ICML 2024)
AU - Hong, Junyuan
AU - Duan, Jinhao
AU - Zhang, Chenhui
AU - Li, Zhangheng
AU - Xie, Chulin
AU - Lieberman, Kelsey
AU - Diffenderfer, James
AU - Bartoldson, Brian
AU - Jaiswal, Ajay
AU - Xu, Kaidi
AU - Kailkhura, Bhavya
AU - Hendrycks, Dan
AU - Song, Dawn
AU - Wang, Zhangyang
AU - Li, Bo
PY - 2024/7
Y1 - 2024/7
N2 - Compressing high-capability Large Language Models (LLMs) has emerged as a favored strategy for resource-efficient inferences. While state-of-the-art (SoTA) compression methods boast impressive advancements in preserving benign task performance, the potential risks of compression in terms of safety and trustworthiness have been largely neglected. This study conducts the first, thorough evaluation of three (3) leading LLMs using five (5) SoTA compression techniques across eight (8) trustworthiness dimensions. Our experiments highlight the intricate interplay between compression and trustworthiness, revealing some interesting patterns. We find that quantization is currently a more effective approach than pruning in achieving efficiency and trustworthiness simultaneously. For instance, a 4-bit quantized model retains the trustworthiness of its original counterpart, but model pruning significantly degrades trustworthiness, even at 50% sparsity. Moreover, employing quantization within a moderate bit range could unexpectedly improve certain trustworthiness dimensions such as ethics and fairness. Conversely, extreme quantization to very low bit levels (3 bits) tends to reduce trustworthiness significantly. This increased risk cannot be uncovered by looking at benign performance alone, in turn, mandating comprehensive trustworthiness evaluation in practice. These findings culminate in practical recommendations for simultaneously achieving high utility, efficiency, and trustworthiness in LLMs. © 2024 by the author(s).
AB - Compressing high-capability Large Language Models (LLMs) has emerged as a favored strategy for resource-efficient inferences. While state-of-the-art (SoTA) compression methods boast impressive advancements in preserving benign task performance, the potential risks of compression in terms of safety and trustworthiness have been largely neglected. This study conducts the first, thorough evaluation of three (3) leading LLMs using five (5) SoTA compression techniques across eight (8) trustworthiness dimensions. Our experiments highlight the intricate interplay between compression and trustworthiness, revealing some interesting patterns. We find that quantization is currently a more effective approach than pruning in achieving efficiency and trustworthiness simultaneously. For instance, a 4-bit quantized model retains the trustworthiness of its original counterpart, but model pruning significantly degrades trustworthiness, even at 50% sparsity. Moreover, employing quantization within a moderate bit range could unexpectedly improve certain trustworthiness dimensions such as ethics and fairness. Conversely, extreme quantization to very low bit levels (3 bits) tends to reduce trustworthiness significantly. This increased risk cannot be uncovered by looking at benign performance alone, in turn, mandating comprehensive trustworthiness evaluation in practice. These findings culminate in practical recommendations for simultaneously achieving high utility, efficiency, and trustworthiness in LLMs. © 2024 by the author(s).
UR - http://www.scopus.com/inward/record.url?scp=85203817859&partnerID=8YFLogxK
UR - https://www.scopus.com/record/pubmetrics.uri?eid=2-s2.0-85203817859&origin=recordpage
M3 - RGC 32 - Refereed conference paper (with host publication)
T3 - Proceedings of Machine Learning Research
SP - 18611
EP - 18633
BT - Proceedings of the 41st International Conference on Machine Learning
A2 - Salakhutdinov, Ruslan
A2 - Kolter, Zico
A2 - Heller, Katherine
A2 - Weller, Adrian
A2 - Oliver, Nuria
A2 - Scarlett, Jonathan
A2 - Berkenkamp, Felix
PB - ML Research Press
Y2 - 21 July 2024 through 27 July 2024
ER -