PulseAugur
LIVE 10:56:36
tool · [1 source] ·
0
tool

Withdrawn paper details novel continuous-time policy evaluation method

A research paper, now withdrawn, proposed a novel method for continuous-time policy evaluation called High-Order Generator Regression. This technique aims to improve upon the standard Bellman baseline by using multi-step transitions and moment-matching coefficients to estimate the time-dependent generator. The paper theoretically decomposed the estimation error and provided a regime map for when higher-order gains are expected, demonstrating consistent improvements over the Bellman baseline in calibration studies. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT This research explores advanced techniques for policy evaluation, potentially impacting reinforcement learning applications.

RANK_REASON The cluster contains a withdrawn academic paper detailing a novel statistical method. [lever_c_demoted from research: ic=1 ai=1.0]

Read on arXiv stat.ML →

COVERAGE [1]

  1. arXiv stat.ML TIER_1 · Yaowei Zheng, Richong Zhang, Shenxi Wu, Shirui Bian, Haosong Zhang, Li Zeng, Xingjian Ma, Yichi Zhang ·

    Beyond Bellman: High-Order Generator Regression for Continuous-Time Policy Evaluation

    arXiv:2604.18972v2 Announce Type: replace Abstract: We study finite-horizon continuous-time policy evaluation from discrete closed-loop trajectories under time-inhomogeneous dynamics. The target value surface solves a backward parabolic equation, but the Bellman baseline obtained…