HumanEval
ai_benchmark
Overview
Open source✓ Open Source
Use caseevaluating functional correctness of code generated from docstrings
Also see
Alternative to
Knowledge graph stats
Claims7
Avg confidence97%
Avg freshness99%
Last updatedUpdated yesterday
Trust distribution
100% unverified
Governance
Not assessed
HumanEval
concept
OpenAI benchmark of 164 hand-written Python programming problems for evaluating code generation
Compare with...used by
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| ○Unverified | High | Fresh | 1 |
alternative to
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| LiveCodeBench | ○Unverified | High | Fresh | 1 |
evaluates
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| Python code generation from function signatures and docstrings | ○Unverified | High | Fresh | 1 |
open source
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| true | ○Unverified | High | Fresh | 1 |
primary use case
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| evaluating functional correctness of code generated from docstrings | ○Unverified | High | Fresh | 1 |
first released
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| 2021 | ○Unverified | High | Fresh | 1 |
created by
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| OpenAI | ○Unverified | High | Fresh | 1 |