Benchmarks/coding/SecCodeBench

SecCodeBench

SecCodeBench evaluates LLM coding agents on secure code generation and vulnerability detection, testing the ability to produce code that is both functional and free from security vulnerabilities.

Progress Over Time

Interactive timeline showing model performance evolution on SecCodeBench

State-of-the-art frontier
Open
Proprietary

SecCodeBench Leaderboard

1 models
ContextCostLicense
1
Alibaba Cloud / Qwen Team
Alibaba Cloud / Qwen Team
397B262K$0.60 / $3.60
Notice missing or incorrect data?

FAQ

Common questions about SecCodeBench

SecCodeBench evaluates LLM coding agents on secure code generation and vulnerability detection, testing the ability to produce code that is both functional and free from security vulnerabilities.
The SecCodeBench leaderboard ranks 1 AI models based on their performance on this benchmark. Currently, Qwen3.5-397B-A17B by Alibaba Cloud / Qwen Team leads with a score of 0.683. The average score across all models is 0.683.
The highest SecCodeBench score is 0.683, achieved by Qwen3.5-397B-A17B from Alibaba Cloud / Qwen Team.
1 models have been evaluated on the SecCodeBench benchmark, with 0 verified results and 1 self-reported results.
SecCodeBench is categorized under coding. The benchmark evaluates text models.