Popular repositories Loading
-
gptq
gptq PublicForked from IST-DASLab/gptq
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
Python
-
GPTQ-for-LLaMa
GPTQ-for-LLaMa PublicForked from qwopqwop200/GPTQ-for-LLaMa
4 bits quantization of LLaMA using GPTQ
Python
-
llm-awq
llm-awq PublicForked from mit-han-lab/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Python
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.



