mirror of
https://github.com/multipleof4/lynchmark.git
synced 2026-01-13 16:17:54 +00:00
Feat: Create initial README for benchmark
This commit is contained in:
24
README.md
Normal file
24
README.md
Normal file
@@ -0,0 +1,24 @@
|
||||
# LLM Algorithmic Benchmark
|
||||
|
||||
This repository contains a suite of difficult algorithmic tests to benchmark the code generation capabilities of various Large Language Models.
|
||||
|
||||
The tests are run automatically via GitHub Actions, and the results are updated in this README.
|
||||
|
||||
## Models Under Test
|
||||
|
||||
The following models are included in the benchmark run.
|
||||
|
||||
<!-- MODELS_START -->
|
||||
google/gemini-2.5-pro
|
||||
anthropic/claude-sonnet-4.5
|
||||
openai/gpt-5-codex
|
||||
<!-- MODELS_END -->
|
||||
|
||||
## Benchmark Results
|
||||
|
||||
The table below shows the pass/fail status for each model on each test.
|
||||
|
||||
<!-- RESULTS_START -->
|
||||
*No results yet. Run the benchmark workflow to generate them.*
|
||||
<!-- RESULTS_END -->
|
||||
|
||||
Reference in New Issue
Block a user