AI-Powered Proof-of-Concept Generator for Security Research. Fine-tuned CodeLlama-7B achieving 78.4% token accuracy.
The model was trained on 1,472 CVE-exploit pairs using QLoRA 4-bit quantization. This optimization technique enabled training on consumer hardware (RTX 4050, 6GB VRAM) while maintaining quality.
Base model specialized for code generation, fine-tuned on exploit patterns from NVD, Exploit-DB, and Metasploit.
4-bit quantization with LoRA adapters (343MB) enabling efficient training on consumer GPUs while preserving quality.
Generates exploits and shellcode for Linux (x86/x64), Windows (x86/x64), and ARM architectures.
PoCSmith consists of a complete AI/ML pipeline: data collection, model training, and production deployment with a professional CLI interface.
The training dataset consists of 1,472 carefully curated samples linking CVE descriptions to working exploit code. Using QLoRA, I achieved a 30% loss reduction over 3 epochs while keeping VRAM usage under 6GB, making it accessible for consumer hardware. The model is published on Hugging Face Hub as LoRA adapters (343MB) that can be loaded on top of CodeLlama-7B.