PulseAugur
LIVE 12:28:27
research · [1 source] ·
0
research

MosaicML releases MPT-7B, an open-source LLM with commercial use license

MosaicML has released MPT-7B, an open-source transformer model trained on one trillion tokens that matches LLaMA-7B's quality and is commercially licensed. This model boasts an impressive context length of up to 84,000 tokens, significantly exceeding limitations found in models like GPT-3. MosaicML also open-sourced its LLM Foundry codebase used for training and evaluation, alongside three fine-tuned versions of MPT-7B, including one specialized for long-form storytelling. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

RANK_REASON Release of an open-source model with significant context length improvements and commercially viable licensing.

Read on Latent Space Podcast →

MosaicML releases MPT-7B, an open-source LLM with commercial use license

COVERAGE [1]

  1. Latent Space Podcast TIER_1 · Alessio Fanelli and Latent.Space ·

    MPT-7B and The Beginning of Context=Infinity — with Jonathan Frankle and Abhinav Venigalla of MosaicML

    <p><em>We are excited to be the </em><strong><em>first</em></strong><em> podcast in the world to release an in-depth interview on the new SOTA in commercially licensed open source models - MosiacML MPT-7B!</em></p><p><em>The Latent Space crew will be at the NYC Lux AI Summit next…