GLM-4.7-Flash-Uncensored-HauhauCS-Balanced

GLM-4.7 Flash uncensored by HauhauCS.

About

No changes to datasets or capabilities. Fully functional, 100% of what the original authors intended - just without the refusals.

These are meant to be the best lossless uncensored models out there.

Agentic Coding

If you're doing agentic coding, use the Balanced variants. Good balance between capability and not refusing everything.

Downloads

File Quant Size
GLM-4.7-Flash-Uncensored-HauhauCS-Balanced-FP16.gguf FP16 56 GB
GLM-4.7-Flash-Uncensored-HauhauCS-Balanced-Q8_0.gguf Q8_0 30 GB
GLM-4.7-Flash-Uncensored-HauhauCS-Balanced-Q6_K.gguf Q6_K 23 GB
GLM-4.7-Flash-Uncensored-HauhauCS-Balanced-Q4_K_M.gguf Q4_K_M 17 GB

Specs

Recommended Settings

From the official Z.ai authors:

General use:

  • --temp 1.0 --top-p 0.95

Tool-calling / agentic:

  • --temp 0.7 --top-p 1.0

Important:

  • Disable repeat penalty (or --repeat-penalty 1.0)
  • For llama.cpp: use --min-p 0.01 (default 0.05 is too high)
  • Use --jinja flag for llama.cpp

Note: Not recommended for Ollama due to chat template issues. Works well with llama.cpp, LM Studio, Jan.

Usage

Works with llama.cpp, LM Studio, Jan, koboldcpp, etc.

Downloads last month
910
GGUF
Model size
30B params
Architecture
deepseek2
Hardware compatibility
Log In to view the estimation

4-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Space using HauhauCS/GLM-4.7-Flash-Uncensored-HauhauCS-Balanced 1