Podchaser Logo
Home
Michael Cohen on Input Tampering in Advanced RL Agents

Michael Cohen on Input Tampering in Advanced RL Agents

Released Sunday, 25th June 2023
Good episode? Give it some love!
Michael Cohen on Input Tampering in Advanced RL Agents

Michael Cohen on Input Tampering in Advanced RL Agents

Michael Cohen on Input Tampering in Advanced RL Agents

Michael Cohen on Input Tampering in Advanced RL Agents

Sunday, 25th June 2023
Good episode? Give it some love!
Rate Episode

Michael Cohen is is a DPhil student at the University of Oxford with Mike Osborne. He will be starting a postdoc with Professor Stuart Russell at UC Berkeley, with the Center for Human-Compatible AI. His research considers the expected behaviour of generally intelligent artificial agents, with a view to designing agents that we can expect to behave safely.

You can see more links and a full transcript at www.hearthisidea.com/episodes/cohen.

We discuss:

  • What is reinforcement learning, and how is it different from supervised and unsupervised learning?
  • Michael's recently co-authored paper titled 'Advanced artificial agents intervene in the provision of reward'
  • Why might it be hard to convey what we really want to RL learners — even when we know exactly what we want?
  • Why might advanced RL systems might tamper with their sources of input, and why could this be very bad?
  • What assumptions need to hold for this "input tampering" outcome?
  • Is reward really the optimisation target? Do models "get reward"?
  • What's wrong with the analogy between RL systems and evolution?

Key links:

Show More
Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features