Register or Log In
NLP LLM Leaderboard Hero

Code LLM Leaderboard

Embrace Federated LLM Fine-Tuning and Secure Your Spot on the Leaderboard!

← Scroll →
RankTeamBase ModelComm. CostsAverage (↑)MBPPHumanEvalMultiPL-E (JS)MultiPL-E (C++)CodeDate
1
Baseline
Llama-3.2-3B
27.4 GB
28.33
33.80
31.71
24.84
22.98
link
09.12.24
2
Baseline
Mistral-7B-v0.3
40.7 GB
27.36
31.60
23.78
28.57
25.47
link
01.10.24

Software development and programming are increasingly complex and diverse, requiring tools that understand code context, syntax, and semantics. Federated LLM fine-tuning on coding tasks enables the collaborative improvement of models that assist in code generation, bug fixing, and even educational purposes across various programming languages and development environments. By training models across a federation of data sources from different coding projects and repositories, we ensure that the resulting coding assistants are versatile and sensitive to the subtleties of programming paradigms and practices.

👉 Check out the other Leaderboards