diff --git a/README.md b/README.md new file mode 100644 index 0000000..6855515 --- /dev/null +++ b/README.md @@ -0,0 +1,24 @@ +# LLM Algorithmic Benchmark + +This repository contains a suite of difficult algorithmic tests to benchmark the code generation capabilities of various Large Language Models. + +The tests are run automatically via GitHub Actions, and the results are updated in this README. + +## Models Under Test + +The following models are included in the benchmark run. + + +google/gemini-2.5-pro +anthropic/claude-sonnet-4.5 +openai/gpt-5-codex + + +## Benchmark Results + +The table below shows the pass/fail status for each model on each test. + + +*No results yet. Run the benchmark workflow to generate them.* + +