PulseAugur
LIVE 14:01:31
research · [1 source] ·
0
research

EleutherAI releases Pile-T5, an improved T5 model trained on The Pile with LLaMA tokenizer

EleutherAI has released Pile-T5, an updated version of the T5 language model. This new iteration was trained on the Pile dataset and utilizes the LLaMA tokenizer, addressing weaknesses in the original T5's handling of code and its pretraining data. Pile-T5 was trained for twice as many tokens as the original T5 and demonstrates significant performance improvements, particularly on code-related tasks, outperforming widely used T5 models even when matched for token count. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

RANK_REASON Release of a new, improved version of an existing open-source model by a known research group.

Read on EleutherAI Blog →

EleutherAI releases Pile-T5, an improved T5 model trained on The Pile with LLaMA tokenizer

COVERAGE [1]

  1. EleutherAI Blog TIER_1 ·

    Pile-T5

    Trained T5 on the Pile