mirror of
https://github.com/multipleof4/lynchmark.git
synced 2026-01-13 16:17:54 +00:00
b258c84ae4f23b09ed46168d96c8a78670ba4a8d
LLM Algorithmic Benchmark
This repository contains a suite of difficult algorithmic tests to benchmark the code generation capabilities of various Large Language Models.
The tests are run automatically via GitHub Actions, and the results are updated in this README.
Models Under Test
The following models are included in the benchmark run.
google/gemini-2.5-pro anthropic/claude-sonnet-4.5 openai/gpt-5-codex
Benchmark Results
The table below shows the pass/fail status for each model on each test.
No results yet. Run the benchmark workflow to generate them.
Description
Languages
JavaScript
85.2%
HTML
14.8%