Hy Low-bit model AngelSlim/Hy-MT1.5-1.8B-1.25bit Translation • Updated 5 days ago • 599 • 80 AngelSlim/Hy-MT1.5-1.8B-2bit Translation • 2B • Updated 5 days ago • 423 • 7 AngelSlim/Hy-MT1.5-1.8B-2bit-GGUF Translation • 2B • Updated 5 days ago • 1.95k • 10 AngelSlim/Hy-MT1.5-1.8B-1.25bit-GGUF Translation • 2B • Updated 5 days ago • 2.42k • 19
Qwen3VL-quant The collection of quantization models of Qwen3-VL AngelSlim/Qwen3-VL-4B-Instruct-FP8-Static 4B • Updated Nov 5, 2025 • 6 • 1 AngelSlim/Qwen3-VL-2B-Instruct-FP8-Static 2B • Updated Nov 5, 2025 • 27 • 1 AngelSlim/Qwen3-VL-8B-Instruct-FP8-Static 9B • Updated Nov 5, 2025 • 13 • 1 AngelSlim/Qwen3-VL-32B-Instruct-FP8-Static Updated Nov 5, 2025 • 9 • 1
Papers SpecExit: Accelerating Large Reasoning Model via Speculative Exit Paper • 2509.24248 • Published Sep 29, 2025 • 2 Tequila: Trapping-free Ternary Quantization for Large Language Models Paper • 2509.23809 • Published Sep 28, 2025 • 3
SpecExit: Accelerating Large Reasoning Model via Speculative Exit Paper • 2509.24248 • Published Sep 29, 2025 • 2
Tequila: Trapping-free Ternary Quantization for Large Language Models Paper • 2509.23809 • Published Sep 28, 2025 • 3
Qwen2/2.5-quant The collection of quantization models of Qwen2 and Qwen2.5 AngelSlim/Qwen2_5-7B_instruct_fp8_static 8B • Updated Jul 23, 2025 • 2 • 1 AngelSlim/Qwen2_5-7B_int4_gptq 8B • Updated Jul 10, 2025 • 1 • 1 AngelSlim/Qwen2_5-7B_int4_awq 8B • Updated Jul 10, 2025 • 1 • 1 AngelSlim/Qwen2_5-1_5B_int4_awq 2B • Updated Jul 10, 2025 • 2 • 1
EAGLE3 The collection of eagle3 series models for Qwen3 and Hunyuan. AngelSlim/Qwen3-8B_eagle3 Updated Jan 13 • 10.3k • 1 AngelSlim/Qwen3-14B_eagle3 Updated Jan 13 • 1.59k • 3 AngelSlim/Qwen3-32B_eagle3 Updated Jan 13 • 2.22k • 7 AngelSlim/Qwen3-a3B_eagle3 Updated Jan 13 • 37 • 8
Qwen3-quant The collection of quantization models of Qwen3 AngelSlim/Qwen3-32B_fp8_static 33B • Updated Jul 23, 2025 • 4 • 1 AngelSlim/Qwen3-32B_int4_gptq 33B • Updated Jul 10, 2025 • 40 • 2 AngelSlim/Qwen3-32B_int4_awq 33B • Updated Jul 10, 2025 • 3 • 1 AngelSlim/Qwen3-8B_fp8_static 8B • Updated Jul 23, 2025 • 2 • 1
Deepseek-quant The collection of quantization models of DeepSeek and Deepseek_r1_distill AngelSlim/DeepSeek-R1-0528_w4a8_fp8 363B • Updated Sep 4, 2025 • 17 • 3 AngelSlim/Deepseek_r1_distill_qwen-1_5b_fp8_static Updated Jul 23, 2025 • 1 • 1 AngelSlim/Deepseek_r1_distill_qwen-1_5b_int4_awq 2B • Updated Jul 10, 2025 • 2 • 1 AngelSlim/Deepseek_r1_distill_qwen-1_5b_int4_gptq 2B • Updated Jul 10, 2025 • 2 • 1
Hy Low-bit model AngelSlim/Hy-MT1.5-1.8B-1.25bit Translation • Updated 5 days ago • 599 • 80 AngelSlim/Hy-MT1.5-1.8B-2bit Translation • 2B • Updated 5 days ago • 423 • 7 AngelSlim/Hy-MT1.5-1.8B-2bit-GGUF Translation • 2B • Updated 5 days ago • 1.95k • 10 AngelSlim/Hy-MT1.5-1.8B-1.25bit-GGUF Translation • 2B • Updated 5 days ago • 2.42k • 19
EAGLE3 The collection of eagle3 series models for Qwen3 and Hunyuan. AngelSlim/Qwen3-8B_eagle3 Updated Jan 13 • 10.3k • 1 AngelSlim/Qwen3-14B_eagle3 Updated Jan 13 • 1.59k • 3 AngelSlim/Qwen3-32B_eagle3 Updated Jan 13 • 2.22k • 7 AngelSlim/Qwen3-a3B_eagle3 Updated Jan 13 • 37 • 8
Qwen3VL-quant The collection of quantization models of Qwen3-VL AngelSlim/Qwen3-VL-4B-Instruct-FP8-Static 4B • Updated Nov 5, 2025 • 6 • 1 AngelSlim/Qwen3-VL-2B-Instruct-FP8-Static 2B • Updated Nov 5, 2025 • 27 • 1 AngelSlim/Qwen3-VL-8B-Instruct-FP8-Static 9B • Updated Nov 5, 2025 • 13 • 1 AngelSlim/Qwen3-VL-32B-Instruct-FP8-Static Updated Nov 5, 2025 • 9 • 1
Qwen3-quant The collection of quantization models of Qwen3 AngelSlim/Qwen3-32B_fp8_static 33B • Updated Jul 23, 2025 • 4 • 1 AngelSlim/Qwen3-32B_int4_gptq 33B • Updated Jul 10, 2025 • 40 • 2 AngelSlim/Qwen3-32B_int4_awq 33B • Updated Jul 10, 2025 • 3 • 1 AngelSlim/Qwen3-8B_fp8_static 8B • Updated Jul 23, 2025 • 2 • 1
Papers SpecExit: Accelerating Large Reasoning Model via Speculative Exit Paper • 2509.24248 • Published Sep 29, 2025 • 2 Tequila: Trapping-free Ternary Quantization for Large Language Models Paper • 2509.23809 • Published Sep 28, 2025 • 3
SpecExit: Accelerating Large Reasoning Model via Speculative Exit Paper • 2509.24248 • Published Sep 29, 2025 • 2
Tequila: Trapping-free Ternary Quantization for Large Language Models Paper • 2509.23809 • Published Sep 28, 2025 • 3
Deepseek-quant The collection of quantization models of DeepSeek and Deepseek_r1_distill AngelSlim/DeepSeek-R1-0528_w4a8_fp8 363B • Updated Sep 4, 2025 • 17 • 3 AngelSlim/Deepseek_r1_distill_qwen-1_5b_fp8_static Updated Jul 23, 2025 • 1 • 1 AngelSlim/Deepseek_r1_distill_qwen-1_5b_int4_awq 2B • Updated Jul 10, 2025 • 2 • 1 AngelSlim/Deepseek_r1_distill_qwen-1_5b_int4_gptq 2B • Updated Jul 10, 2025 • 2 • 1
Qwen2/2.5-quant The collection of quantization models of Qwen2 and Qwen2.5 AngelSlim/Qwen2_5-7B_instruct_fp8_static 8B • Updated Jul 23, 2025 • 2 • 1 AngelSlim/Qwen2_5-7B_int4_gptq 8B • Updated Jul 10, 2025 • 1 • 1 AngelSlim/Qwen2_5-7B_int4_awq 8B • Updated Jul 10, 2025 • 1 • 1 AngelSlim/Qwen2_5-1_5B_int4_awq 2B • Updated Jul 10, 2025 • 2 • 1