Olivier Codol
Neuroscience, RL for motor learning, neural control of movement, NeuroAI.
Opinions stated here are my own, not those of my employer.
- Q-value overestimation animation for my upcoming talk about "Recent Advances in RL for Continuous Control" at the Mannheim RL Workshop
- Do you observe these issues with value functions instead of Q-values? I’ve had much less issues with value functions than Q-value based algorithms in the past and wondering if this relates