Add Qwen/Qwen3.5-0.8B to quant queue
Browse files
Qwen/Qwen3.5-0.8B_quant_request_False_W4A16_4bit_int4.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"job_type": "quantization & evaluation", "model": "Qwen/Qwen3.5-0.8B", "revision": "main", "private": false, "params": 0.87, "architectures": "Qwen3_5ForConditionalGeneration", "num_layers": 24, "input_dtype": "float16", "input_bits": 16, "model_params": 0.87, "model_weight_gb": 1.75, "quant_scheme": "INT4 (W4A16)", "quant_precision": "4bit", "quant_weight_dtype": "int4", "quant_bits": 4, "quant_memory_gb": 0.19, "quant_gpu_type": "NVIDIA GeForce RTX 4090", "quant_gpu_nums": 1, "eval_memory_gb": 0.52, "eval_gpu_type": "NVIDIA GeForce RTX 4090", "eval_gpu_nums": 1, "quant_model_size_gb": 0.43, "hardware": "NVIDIA GeForce RTX 4090", "gpu_nums": 1, "script": "auto_quant", "status": "Pending", "submitted_time": "2026-04-24T12:20:48Z", "model_type": "quantization", "job_id": -1, "job_start_time": null, "likes": 511, "license": "apache-2.0", "still_on_hub": true, "tags": [], "downloads": 0, "created_at": ""}
|