An uncensored version of the open-source language model GPT-OSS 120B has been released. This "aggressive" version is designed not to refuse any query.
Technical Specifications
- Total parameters: 117 billion
- Active parameters: ~5.1 billion (MoE with 128 experts, top-4 routing)
- Context window: 128K
- Format: MXFP4 (model's native precision)
The model is provided in a single 61GB file and can be run on a single H100 GPU. For configurations with lower VRAM, it is possible to offload the MoE layers to the CPU via the --n-cpu-moe N option in llama.cpp. It works with llama.cpp, LM Studio, Ollama, and others.
Other Models
Uncensored versions of GPT-OSS 20B, GLM 4.7 Flash, and Qwen3 8b VL are also available.
๐ฌ Comments (0)
๐ Log in or register to comment on articles.
No comments yet. Be the first to comment!