【 以下文字转载自 NewExpress 讨论区 】
"OLMo 2 32B: First fully open model to outperform GPT 3.5 and GPT 4o mini"
"OLMo is a fully open model: [they] release all artifacts. Training code, pre- & post-train data, model weights, and a recipe on how to reproduce it yourself."
--
FROM 182.118.238.*