abacaj / code-eval

Run evaluation on LLMs using human-eval benchmark
380Updated last year

Related projects

Alternatives and complementary repositories for code-eval