MiMo


Révision datée du 17 mai 2025 à 20:48 par Pitpitt (discussion | contributions) (Page créée avec « ==en construction== == Définition == XXXXXXXXX == Français == ''' MiMo-7B'' == Anglais == '''MiMo-7B''' MiMo-7B, a large language model specifically designed for reasoning tasks. The model is optimized across both pre-training and post-training stages to unlock its reasoning potential. Despite having only 7 billion parameters, MiMo-7B achieves superior performance on mathematics and code reasoning tasks, outperforming even much larger models including Open... »)
(diff) ← Version précédente | Voir la version actuelle (diff) | Version suivante → (diff)

en construction

Définition

XXXXXXXXX

Français

' MiMo-7B

Anglais

MiMo-7B

MiMo-7B, a large language model specifically designed for reasoning tasks. The model is optimized across both pre-training and post-training stages to unlock its reasoning potential. Despite having only 7 billion parameters, MiMo-7B achieves superior performance on mathematics and code reasoning tasks, outperforming even much larger models including OpenAI's o1-mini.

Source

Source : arxiv

Contributeurs: wiki