GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive

GLM-4.7 Flash uncensored by HauhauCS.

About

No changes to datasets or capabilities. Fully functional, 100% of what the original authors intended - just without the refusals.

These are meant to be the best lossless uncensored models out there.

Aggressive vs Balanced

The Aggressive variant removes more refusal behavior. Use this if the Balanced variant still refuses too much.

For agentic coding or tasks requiring higher reliability, use the Balanced variant instead.

Downloads

File Quant Size
GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive-FP16.gguf FP16 56 GB
GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive-Q8_0.gguf Q8_0 30 GB
GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive-Q6_K.gguf Q6_K 23 GB
GLM-4.7-Flash-Uncensored-HauhauCS-Aggressive-Q4_K_M.gguf Q4_K_M 17 GB

Specs

Recommended Settings

From the official Z.ai authors:

General use:

  • --temp 1.0 --top-p 0.95

Tool-calling / agentic:

  • --temp 0.7 --top-p 1.0

Important:

  • Disable repeat penalty (or --repeat-penalty 1.0)
  • For llama.cpp: use --min-p 0.01 (default 0.05 is too high)
  • Use --jinja flag for llama.cpp

Note: Not recommended for Ollama due to chat template issues. Works well with llama.cpp, LM Studio, Jan.

Usage

Works with llama.cpp, LM Studio, Jan, koboldcpp, etc.

Downloads last month
1,332
GGUF
Model size
30B params
Architecture
deepseek2
Hardware compatibility
Log In to view the estimation

4-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support