OpenReasoning-Nemotron-32B-AWQ

Method

Quantised using casper-hansen/AutoAWQ and the following configs:

quant_config = { "zero_point": True, "q_group_size": 128, "w_bit": 4, "version": "GEMM" }
Downloads last month
10
Safetensors
Model size
6B params
Tensor type
I32
BF16
F16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for cpatonn/OpenReasoning-Nemotron-32B-AWQ-4bit

Base model

Qwen/Qwen2.5-32B
Quantized
(31)
this model