Researchers have developed Themis-RM, a suite of multilingual code reward models designed for flexible scoring across multiple criteria. These models, ranging from 600M to 32B parameters, were trained on Themis-CodePreference, a large dataset of over 350,000 code preferences. The accompanying Themis-CodeRewardBench benchmark evaluates code RMs across eight programming languages and five preference dimensions, revealing limitations in current models beyond functional correctness. Experiments show positive scaling trends and strong cross-lingual transfer, highlighting the value of multi-criteria training. AI
Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →
IMPACT Introduces new tools and benchmarks for evaluating and training code generation models, potentially improving their multi-lingual and multi-criteria capabilities.
RANK_REASON This is a research paper detailing the creation of new code reward models and a benchmark.