A new paper explores the critical role of user state representation in contextual multi-armed bandit (CMAB) recommender systems, finding that variations in state representation can yield greater performance improvements than changes to the bandit algorithm itself. The research highlights that no single embedding or aggregation strategy is universally superior, emphasizing the need for domain-specific evaluations. Another study introduces BEAR, a novel fine-tuning objective for Large Language Models (LLMs) in recommendation tasks that explicitly accounts for beam search behavior during training to address inconsistencies between training and inference. Additionally, a paper proposes a methodology to measure the stability and plasticity of recommender systems, evaluating how models adapt to retraining and changes in data patterns. AI
Summary written by gemini-2.5-flash-lite from 9 sources. How we write summaries →
IMPACT Advances in user state representation and LLM fine-tuning for recommendations could lead to more personalized and effective user experiences.
RANK_REASON The cluster contains multiple academic papers published on arXiv, focusing on research in recommender systems and LLM applications.