[general_dat] [Recordatorio] Seminario AISAR – The Theoretical Foundations of Reward Learning

Agustín Martinez Suñé agusmartinez92 at gmail.com
Mon Oct 13 08:05:49 -03 2025


Recordatorio de esta charla que es hoy!

El dom, 12 oct 2025 a las 20:22, Agustín Martinez Suñé (<
agusmartinez92 at gmail.com>) escribió:

> Desde el Programa de Becas AISAR en AI Safety tenemos el placer de
> invitarlos a la próxima charla de nuestro seminario online, con la
> participación de investigadores del área.
>
> 📌 Fecha y hora: Lunes 13 de octubre, 10:00 hs (ARG).
> 🎤 Orador: Joar Skalse – PhD @ University of Oxford | Director @ DEDUCTO
> 📖 Título: The Theoretical Foundations of Reward Learning
>
> Charla online: Para asistir a la charla, registrate acá:
> https://luma.com/nostkcsb
>
> Abstract: In this talk, I will provide an overview of my research on how
> to build a theoretical foundation for the field of reward learning,
> including my main motivations for pursuing this research, and some of my
> core results.
>
> This research agenda involves answering questions such as: What is the
> right method for expressing goals and instructions to AI systems? How
> similar must two different goal specifications be in order to not be
> hackable? What is the right way to quantify the differences and
> similarities between different goal specifications in a given specification
> language? What happens if you execute a task specification that is not
> close to the “ideal” specification? Which specification learning algorithms
> are guaranteed to converge to a good specification? How sensitive are these
> specification learning algorithms to misspecification? If we have a bound
> on the error in a specification (under some metric), can we devise safe
> methods for optimising it?
>
> Encontrá más detalles en: https://www.lesswrong.com/s/TEybbkyHpMEB2HTv3
> Equipo AISAR
> http://scholarship.aisafety.ar/
> <http://scholarship.aisafety.ar/?utm_source=chatgpt.com>
>


Más información sobre la lista de distribución general_dat