MLSys'24 Best Paper - AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration

125 views

MIT HAN Lab

7 months ago

MLSys'24 Best Paper - AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration

MLSys'24 Best Paper - AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration