Thank you for your great works. We have released the 4-bit GPTQ quantized LLaDA model on Hugging Face: - [LLaDA-8B-Base-GPTQ-4bit](https://huggingface.co/FunAGI/LLaDA-8B-Base-gptqmodel-4bit) - [LLaDA-8B-Instruct-GPTQ-4bit](https://huggingface.co/FunAGI/LLaDA-8B-Instruct-gptqmodel-4bit) Based on the published evaluation code, we have evaluated the quantized **base model**. The results are as follows: | Dataset | GPTQ-4bit | FP16 | |----------------|------------|--------| | MMLU | 65.20 | 65.90 | | CMMLU | 69.23 | 69.90 | | ARC-Challenge | 45.48 | 47.90 |
This issue appears to be discussing a feature request or bug report related to the repository. Based on the content, it seems to be still under discussion. The issue was opened by 1773226512 and has received 1 comments.