A study evaluates the trustworthiness of compressed LLMs using different compression techniques. Quantization is found to be more effective than pruning, with moderate bit-range quantization enhancing trust dimensions like ethics and fairness. Extreme quantization poses risks to trustworthiness.

5m read timeFrom marktechpost.com
Post cover image

Sort: