Qwen2.5-Coder-3B Pentest - GGUF

GGUF quantizations of fawazo/qwen2.5-coder-3b-pentest optimized for Jetson Orin Nano (8GB).

Model Description

An AI pentesting assistant fine-tuned on 150K+ cybersecurity examples covering:

  • OWASP Top 10 vulnerabilities
  • MITRE ATT&CK framework
  • API security testing
  • Web application penetration testing

Output Format: JSON for automation

Quantizations

File Size RAM Needed Recommended For
qwen2.5-coder-3b-pentest-q4_k_m.gguf ~1.8GB ~3GB Jetson Orin Nano 8GB
qwen2.5-coder-3b-pentest-q5_k_m.gguf ~2.1GB ~4GB Better quality
qwen2.5-coder-3b-pentest-q8_0.gguf ~3.4GB ~5GB Best quality
qwen2.5-coder-3b-pentest-f16.gguf ~6GB ~8GB Full precision

Usage on Jetson

With Ollama

# Download Q4_K_M (recommended for 8GB)
huggingface-cli download fawazo/qwen2.5-coder-3b-pentest-gguf qwen2.5-coder-3b-pentest-q4_k_m.gguf

# Create Modelfile
cat > Modelfile << 'EOF'
FROM ./qwen2.5-coder-3b-pentest-q4_k_m.gguf

SYSTEM """You are an expert penetration testing AI assistant. Analyze web traffic and respond with JSON:
{"action": "report|request|command|complete", ...}"""

PARAMETER temperature 0.3
PARAMETER num_ctx 2048
EOF

# Create and run
ollama create pentest-agent -f Modelfile
ollama run pentest-agent

With llama.cpp

./llama-cli -m qwen2.5-coder-3b-pentest-q4_k_m.gguf -ngl 99 -c 2048 -p "Analyze this request..."

Example Usage

Input:

Analyze this HTTP exchange:
REQUEST: GET /api/users?id=1
RESPONSE: {"user": "admin", "role": "administrator"}

Output:

{
  "action": "request",
  "method": "GET",
  "path": "/api/users?id=2",
  "reasoning": "Testing for IDOR - checking if user IDs are enumerable"
}

Training Details

  • Base: Qwen/Qwen2.5-Coder-3B
  • Method: SFT with LoRA (r=32)
  • Dataset: 150K+ combined examples from Trendyol, Fenrir v2.0, pentest-agent
  • Frameworks: OWASP, MITRE ATT&CK, NIST CSF

License

Apache 2.0 (inherits from base model and training datasets)

Downloads last month
54
GGUF
Model size
3B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for fawazo/qwen2.5-coder-3b-pentest-gguf

Base model

Qwen/Qwen2.5-3B
Quantized
(29)
this model