Statements (20)
| Predicate | Object |
|---|---|
| gptkbp:instanceOf |
gptkb:performance
|
| gptkbp:author |
Mark Chen et al.
|
| gptkbp:citation |
gptkb:Evaluating_Large_Language_Models_Trained_on_Code
2021 |
| gptkbp:contains |
164 programming problems
|
| gptkbp:createdBy |
gptkb:OpenAI
|
| gptkbp:hasTaskType |
code completion
function synthesis |
| gptkbp:hasTestCases |
yes
|
| gptkbp:language |
gptkb:Python
|
| gptkbp:license |
gptkb:MIT_License
|
| gptkbp:releaseYear |
2021
|
| gptkbp:repository |
https://github.com/openai/human-eval
|
| gptkbp:usedBy |
gptkb:Upanishads
gptkb:GPT-4 other LLMs |
| gptkbp:usedFor |
code generation evaluation
|
| gptkbp:bfsParent |
gptkb:LLaMA_3
|
| gptkbp:bfsLayer |
7
|
| https://www.w3.org/2000/01/rdf-schema#label |
HumanEval
|