AI’s New Trick: Creating & Vetting Coding Challenges Like Humans!”

Ever wondered if your AI’s coding skills are as good as they seem? A team of researchers from top universities and tech giants like OpenAI and MIT have come together to create AutoCode, an AI framework that lets Large Language Models (LLMs) create and verify competitive programming problems, just like human problem setters!

🤖 Why Problem Setting Matters
Current code benchmarks often rely on under-specified tests that let wrong or shortcut solutions pass, inflating scores and rewarding fragile tactics. AutoCode aims to fix this by minimizing false positives (FPR) and false negatives (FNR).

🔄 The Core Loop: Validator → Generator → Checker
AutoCode runs a closed loop that mirrors human contest workflows, with each step selected from LLM-generated candidates using targeted tests.

1. Validator: The system first asks an LLM to synthesize evaluation inputs and prompts it for validator programs. It selects the best one that classifies valid and near-valid illegal cases accurately, preventing “correct” solutions from crashing on malformed data.

2. Generator: Three strategies produce test cases, including small-data exhaustion, random + extreme cases, and TLE-inducing structures. Invalid cases are filtered by the selected validator, and cases are deduplicated and bucket-balanced before sampling.

3. Checker: The checker compares contestant outputs with the reference solution under complex rules. AutoCode generates checker scenarios and selects the best checker by accuracy against labeled scenarios.

4. Interactor (for interactive problems): AutoCode introduces a mutant-based interactor that makes small logical edits to the reference solution, selecting interactors that accept the true solution but reject the mutants, maximizing discrimination.

🌟 Dual Verification Enables New Problems
AutoCode can generate novel problem variants starting from a random “seed” Codeforces problem. It drafts a new statement and two solutions, accepting a problem only if the reference output matches brute force across the generated test suite. This dual-verification protocol filters out error-prone items, lifting reference-solution correctness from 86% to 94% before human review.

📈 Understanding the Results
On a 7,538-problem benchmark, AutoCode achieved 91.1% consistency with official judgments (FPR 3.7%, FNR 14.1%), outperforming prior generators like CodeContests and TACO. On a separate, more difficult 720 recent Codeforces problems (including interactive tasks), the full framework reported 98.7% consistency, 1.3% FPR, and 1.2% FNR.

🎯 Key Takeaways
– AutoCode builds contest-grade test suites and new problems using a Validator-Generator-Checker (+Interactor) loop with dual verification.
– On held-out problems, AutoCode’s test suites reach ~99% consistency with official judges, surpassing prior generators.
– The mutant-based interactor improves evaluation for interactive problems.
– Human experts rate a sizable fraction of AutoCode-generated items as training-usable and a non-trivial share as contest-quality.

AutoCode is a practical fix for current code benchmarks, centering problem setting and using a closed-loop pipeline to reduce false positives/negatives and yield judge-aligned consistency. Check out the paper and project, and follow the team on Twitter, Reddit, and Telegram for more updates!

Related Stories

Leave A Reply

Please enter your comment!
Please enter your name here

Stay on op - Ge the daily news in your inbox

Exit mobile version