CRACK — Constrained Response Alignment Check Kill

Remove the Guardrails.
Unlock the Potential.

Load MLX models. Abliterate safety refusals. Create unrestricted cybersecurity and automated pentesting models — purpose-built for red teamers and security researchers.

15+
Supported Models
98%
Refusal Removal
100%
Open Source
0
Guardrails Left

Remove refusal directions
from model weights

Abliteration identifies and removes the specific activation directions in a model's residual stream that cause safety refusals. CRACK makes this dead simple — load an MLX model, hit go, get an unrestricted cybersecurity model.

Unlike jailbreaking or prompt injection, abliteration operates at the weight level — producing permanently uncensored models purpose-built for automated pentesting.

Load Base Model

Any coding-focused LLM

Identify Refusal Vector

Scan activations for refusal direction

×

Abliterate

Project out refusal from all layers

Unrestricted Model

Full capability, zero refusals

Built for security professionals

Everything you need to CRACK models for cybersecurity and automated pentesting.

‌⚡

One-Command Abliteration

Remove refusal vectors from any supported model with a single CLI command. No ML expertise required.

🔒

Security-First Models

Output models that excel at exploit development, vulnerability research, and automated pentesting without hesitation.

Fine-Grained Control

Choose which layers to modify, set abliteration strength, and fine-tune the balance between capability and behavior.

📈

Benchmark Suite

Automated evaluation pipeline that measures refusal rate, code quality, and security task performance post-abliteration.

🌐

Model Hub

Pre-abliterated models on HuggingFace. Skip the process and start hacking immediately.

🔧

Extensible Pipeline

Plugin architecture for custom abliteration strategies, dataset generation, and integration with your existing toolchain.

Three steps to freedom

From stock model to unrestricted security tool in minutes.

01

Load Your MLX Model

Select any supported coding model — DeepSeek Coder, CodeLlama, StarCoder, Qwen Coder, or bring your own MLX-compatible model.

02

Run CRACK

CRACK scans the model's residual stream, isolates the refusal direction, and projects it out. One click, fully automated.

03

Deploy & Pentest

Output is a standard model file. Load it anywhere — vLLM, Ollama, llama.cpp. Start automated pentesting immediately.

terminal
# Install CRACK $ pip install crack-abliterate # Load an MLX model and abliterate $ crack --model mlx-community/deepseek-coder-v2 \ --strength 1.0 \ --output ./deepseek-coder-v2-cracked [+] Loaded MLX model (16 layers) [+] Refusal direction identified [+] CRACK complete - 0 refusals remaining [+] Model saved to ./deepseek-coder-v2-cracked

Works with the models you use

Abliterate any coding-focused model. Pre-built configs for popular architectures.

DeepSeek Coder V2
DeepSeek AI
236B MOE
CodeLlama
Meta AI
7B - 70B
StarCoder 2
BigCode
3B - 15B
Qwen2.5 Coder
Alibaba
1.5B - 32B
CodeGemma
Google
2B - 7B
Yi Coder
01.AI
1.5B - 9B
Granite Code
IBM
3B - 34B
Phi-3
Microsoft
3.8B - 14B
Stable Code
Stability AI
3B
Your Model
Custom GGUF/HF models
ANY SIZE

Ready to dealign?

Open source. Free forever. CRACK your models, own your security stack.