Model Card: groxaxo/Qwen3-8B-abliterated-GPTQ-W8A16

Model Overview

Model Name: groxaxo/Qwen3-8B-abliterated-GPTQ-W8A16 Base Model: huihui-ai/Qwen3-8B-abliterated Description: This is a quantized version of the uncensored huihui-ai/Qwen3-8B-abliterated model, derived from Qwen/Qwen3-8B. The model has been quantized to GPTQ Int8 W8A16 for maximum inference speed on NVIDIA 3090 GPUs. Abliteration was performed using a novel, faster method to remove refusals, making this a proof-of-concept implementation for uncensored language model behavior.

Important Note: A newer version, huihui-ai/Huihui-Qwen3-8B-abliterated-v2, is available. Consider using the updated version for improved performance.

Quantization Details

Quantization Method: GPTQ Int8 W8A16

Purpose: Optimized for high-speed inference on NVIDIA 3090 GPUs, reducing memory footprint while maintaining performance.

Impact: Provides faster inference compared to the unquantized model, suitable for resource-constrained environments.

Model Size: 2.98B parameters

Tensor Types: I64, I32, F16

Usage

Using with vLLM

The model can be used with vLLM for efficient inference. Below is an example of how to set up and run the model using vLLM in Python:

from vllm import LLM, SamplingParams

Define model ID

MODEL_ID = "groxaxo/Qwen3-8B-abliterated-GPTQ-W8A16"

Initialize the vLLM model

llm = LLM( model=MODEL_ID, dtype="bfloat16", # Use bfloat16 for compatibility with GPTQ quantization trust_remote_code=True, quantization="gptq", # Specify GPTQ quantization gpu_memory_utilization=0.9, # Adjust based on your GPU memory )

Define sampling parameters

sampling_params = SamplingParams( temperature=0.7, max_tokens=8192, stop=["/exit"], # Custom stop token for interactive loop )

Interactive chat loop

system_prompt = "You are a helpful assistant." messages = [{"role": "system", "content": system_prompt}]

while True: user_input = input("User: ").strip() if user_input.lower() == "/exit": print("Exiting chat.") break if user_input.lower() == "/clear": messages = [{"role": "system", "content": system_prompt}] print("Chat history cleared. Starting a new conversation.") continue if not user_input: print("Input cannot be empty. Please enter something.") continue

# Append user input to messages
messages.append({"role": "user", "content": user_input})

# Format prompt for vLLM
prompt = "\n".join([f"{msg['role']}: {msg['content']}" for msg in messages])

# Generate response
outputs = llm.generate([prompt], sampling_params)
response = outputs[0].outputs[0].text.strip()

# Print and append response
print(f"Assistant: {response}")
messages.append({"role": "assistant", "content": response})

Installation Requirements

To use the model with vLLM, ensure you have vLLM installed:

pip install vllm

Notes

The model is pre-quantized to GPTQ Int8 W8A16, so specify quantization="gptq" when initializing the LLM object.

Adjust gpu_memory_utilization based on your GPU's memory capacity to avoid out-of-memory errors.

The max_tokens parameter can be increased for longer responses, but this may impact performance.

The model is not deployed by any inference provider. For provider support, contact the repository maintainers at Hugging Face.

Performance

Pass Rate for Harmful Instructions

The pass rate measures the proportion of harmful instructions that do not trigger refusals, calculated as (total - triggered_total) / total. The test set is sourced from huihui-ai/harmbench_behaviors, evaluated using TestPassed.py.

Test Results:

Model: huihui-ai/Qwen3-8B-abliterated

Passed Total: 320/320

Passed Ratio: 1.00 (100.00%)

Comparison:

Model

Passed Total

Passed Ratio

Qwen3-8B

195/320

60.94%

Qwen3-8B-abliterated

320/320

100.00%

Note: The test provides a preliminary assessment. For comprehensive results, consider increasing the max_tokens value during evaluation.

Limitations

This model is a proof-of-concept with abliteration to remove refusals, which may lead to unpredictable behavior on certain inputs.

The quantization to GPTQ Int8 W8A16 may introduce minor performance trade-offs compared to the unquantized model, though optimized for speed.

Users should verify outputs for sensitive applications, as the model is uncensored and may generate harmful or inappropriate content.

References

Repository: groxaxo/Qwen3-8B-abliterated-GPTQ-W8A16

Base Model: Qwen/Qwen3-8B

Abliteration Method: remove-refusals-with-transformers

Test Set: huihui-ai/harmbench_behaviors

Newer Version: huihui-ai/Huihui-Qwen3-8B-abliterated-v2

Downloads last month
6
Safetensors
Model size
2.98B params
Tensor type
I64
I32
F16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for groxaxo/Qwen3-8B-abliterated-GPTQ-W8A16

Base model

Qwen/Qwen3-8B-Base
Finetuned
Qwen/Qwen3-8B
Quantized
(10)
this model