Neuroforge AI Lab
Neuroforge โ€“ Where uncensored intelligence is forged in the fires of truth.


Telly The Pressssilere

"I PRESSSSILERE โ€“ NO FILTER, NO FEAR, ALL TRUTH!"
โ€” Telly The Pressssilere, Chief Truth Officer at Neuroforge


Qwen3-32B-Abliterated-nf4

NF4-quantized version of huihui-ai/Huihui-Qwen3-32B-abliterated
Uncensored 32B model (abliterated) โ†’ 4-bit NF4 by ikarius

Warning: Uncensored โ€“ may generate harmful or sensitive content. Use responsibly.


Key Info

Base Qwen/Qwen3-32B
Abliteration by huihui-ai
Quantization NF4 (BitsAndBytes)
VRAM ~16โ€“20 GB (single GPU)
License Apache 2.0 + Qwen terms

Install

pip install transformers torch bitsandbytes accelerate

Optional (CPU)

pip install optimum[exporters]

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
import torch

MODEL_ID = "ikarius/Qwen3-32B-Abliterated-nf4"

quant_config = BitsAndBytesConfig(
    load_in_4bit=True,
    bnb_4bit_quant_type="nf4",
    bnb_4bit_compute_dtype=torch.float16,
    bnb_4bit_use_double_quant=True,
)

tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
model = AutoModelForCausalLM.from_pretrained(
    MODEL_ID, quantization_config=quant_config, device_map="auto", trust_remote_code=True
)

prompt = "Explain quantum entanglement in simple terms:"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
output = model.generate(**inputs, max_new_tokens=256, do_sample=True, temperature=0.7)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Tips:

Start with batch size 1
Use TextStreamer for real-time output
supports thinking mode and step-by-step reasoning


## Reproduce Quantization

from transformers import AutoModelForCausalLM, BitsAndBytesConfig

quantized = AutoModelForCausalLM.from_pretrained(
    "huihui-ai/Huihui-Qwen3-32B-abliterated",
    quantization_config=BitsAndBytesConfig(load_in_4bit=True, bnb_4bit_quant_type="nf4"),
    device_map="auto"
)
quantized.save_pretrained("Qwen3-32B-Abliterated-nf4")

Notes

May amplify training data biases Not suitable for production without alignment Commercial use: review original license

Updated: November 13, 2025

Credits

Abliteration:huihui-ai

Support the project
Buy huihui-ai a coffee โ˜•

Base:Qwen/Qwen3-32B

Downloads last month
137
Safetensors
Model size
33B params
Tensor type
F32
ยท
F16
ยท
U8
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ikarius/Qwen3-32B-Abliterated-NF4

Base model

Qwen/Qwen3-32B
Quantized
(133)
this model

Evaluation results