Abstract: Post-training quantization(PTQ) has been widely studied in recent years because it does not require retraining the network or the entire training dataset. However, naively applying the PTQ ...
This is the code for the paper [OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models](https://arxiv.org/abs/2306. ...
Abstract: Automatic quantization generates efficient hybrid precision quantization schemes without manual effort, offering a promising approach for developing hardware-friendly MIMO detectors. However ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results