Delete: Replaced with extensionless README file

This commit is contained in:
2025-10-13 05:50:06 -07:00
parent cd9bf7e10f
commit ac7a450d31

View File

@@ -1,29 +0,0 @@
# LLM Algorithmic Benchmark
This repository contains a suite of difficult algorithmic tests to benchmark the code generation capabilities of various Large Language Models.
The tests are run automatically via GitHub Actions, and the results are updated in this README.
## Models Under Test
The following models are included in the benchmark run.
<!-- MODELS_START -->
google/gemini-2.5-pro
anthropic/claude-sonnet-4.5
openai/gpt-5-codex
<!-- MODELS_END -->
## Benchmark Results
The table below shows the pass/fail status for each model on each test.
<!-- RESULTS_START -->
| Model | 1_dijkstra | 2_convex_hull | 3_lis | 4_determinant |
| --- | --- | --- | --- | --- |
| google/gemini-2.5-pro | ❌ Fail | ❌ Fail | ❌ Fail | ❌ Fail |
| anthropic/claude-sonnet-4.5 | ❌ Fail | ❌ Fail | ❌ Fail | ❌ Fail |
| openai/gpt-5-codex | ❌ Fail | ❌ Fail | ❌ Fail | ❌ Fail |
<!-- RESULTS_END -->