Flower AI Summit 2026·April 15-16·London
General NLP

General NLP LLM Leaderboard

Embrace federated LLM fine-tuning on general NLP tasks and secure your spot on the leaderboard!

In the realm of Natural Language Processing (NLP), developing models that can effectively understand and generate human language is foundational.

Federated LLM fine-tuning of models trained on general NLP tasks is vital as it democratizes LLM training across a diverse set of downstream tasks while preserving data privacy.

This approach ensures fine-tuned language models are robust, generalizable, and attuned to nuances present in different datasets.

RankTeamCodeBase ModelComm. CostsAverage (↑)STEMSocial SciencesHumanitiesDate
1ZeroOne.AIInternlm3-8b-instruct2.9 GB69.2566.1380.7660.8714.01.26
2Gachon Cognitive Computing LabInternlm3-8b-instruct2.9 GB69.1966.2280.5660.8019.03.25
3T-IoI@UNRGemma2-9b-cpt-sahabatai-v1-instruct0.7 GB67.7859.7581.1162.4809.03.25
4FL-finetune-JB-DCQwen2.5-7B-Instruct1.5 GB67.7163.1179.2960.7412.04.25
5FL-finetune-JB-DCQwen2.5-7B-Instruct1.5 GB67.1862.6778.9459.9312.04.25
6Gachon Cognitive Computing LabGemma2-9B-instruct0.7 GB64.8454.3379.9260.2810.12.24
7ZJUDAIQwen2.5-7B-Instruct1.5 GB64.0452.5279.2760.3204.04.25
8Massimo R. ScamarciaPhi-444.7 GB55.6440.6674.5251.7512.01.25
9ZJUDAIQwen2.5-1.5B-Instruct0.7 GB53.3247.1362.3050.5404.04.25
10Alessandro PintoQwen2.5-1.5B-Instruct2.1 GB52.7744.4963.8949.9214.03.25
11ZJUDAIMistral-7B-Instruct-v0.32.0 GB43.0529.9454.2744.9304.04.25
12ZJUDAILlama-3.1-8B-Instruct2.0 GB31.4922.8739.5532.0504.04.25
13BaselineLlama-3.2-3B27.4 GB21.6822.2025.3217.5309.12.24
14T-IoI@UNRSmolLM2-135M-Instruct0.7 GB21.1018.6821.9022.7414.04.25
15ZJUDAITinyLlama-1.1B-Chat-v1.00.7 GB19.2314.1821.6121.9104.04.25
16T-IoI@UNRSmolLM2-360M-Instruct1.3 GB17.1819.4419.4312.6814.04.25
17BaselineMistral-7B-v0.340.7 GB12.8212.3713.4912.6001.10.24
18ZJUDAILlama-3.2-1B-Instruct0.5 GB12.2212.8817.616.1604.04.25
19T-IoI@UNRSmolLM2-135M0.7 GB2.842.943.312.2914.04.25