« MiMo » : différence entre les versions


(Page créée avec « ==en construction== == Définition == XXXXXXXXX == Français == ''' MiMo-7B'' == Anglais == '''MiMo-7B''' MiMo-7B, a large language model specifically designed for reasoning tasks. The model is optimized across both pre-training and post-training stages to unlock its reasoning potential. Despite having only 7 billion parameters, MiMo-7B achieves superior performance on mathematics and code reasoning tasks, outperforming even much larger models including Open... »)
 
Aucun résumé des modifications
 
Ligne 5 : Ligne 5 :


== Français ==
== Français ==
''' MiMo-7B''
''' MiMo-7B'''


== Anglais ==
== Anglais ==

Dernière version du 17 mai 2025 à 20:51

en construction

Définition

XXXXXXXXX

Français

MiMo-7B

Anglais

MiMo-7B

MiMo-7B, a large language model specifically designed for reasoning tasks. The model is optimized across both pre-training and post-training stages to unlock its reasoning potential. Despite having only 7 billion parameters, MiMo-7B achieves superior performance on mathematics and code reasoning tasks, outperforming even much larger models including OpenAI's o1-mini.

Source

Source : arxiv

Contributeurs: wiki