This paper systematically investigates the in-context learning capabilities of Transformer models, focusing on Gaussian-mixture binary classification tasks. It empirically analyzes how factors like input dimension, number of examples, and pre-training tasks influence in-context accuracy. The research also explores benign overfitting, where models generalize well despite memorizing noisy in-context labels, and maps the conditions under which in-context learning succeeds or fails. AI
Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →
IMPACT Provides an empirical map of scaling behavior in in-context classification, highlighting critical factors for success.
RANK_REASON Academic paper investigating in-context learning capabilities of Transformers.