Reinforcement Learning, Fast and Slow
Abstract:
Deep reinforcement learning (RL) methods have driven impressive advances in artificial intelligence in recent years, exceeding human performance in domains ranging from Atari to Go to no-limit poker. This progress has drawn the attention of cognitive scientists interested in understanding human learning. However, the concern has been raised that deep RL may be too sample-inefficient – that is, it may simply be too slow – to provide a plausible model of how humans learn. In the present review, we counter this critique by describing recently developed techniques that allow deep RL to operate more nimbly, solving problems much more quickly than previous methods. Although these techniques were developed in an AI context, we propose that they may have rich implications for psychology and neuroscience. A key insight, arising from these AI methods, concerns the fundamental connection between fast RL and slower, more incremental forms of learning.
Authors: Matthew Botvinick, Sam Ritter, Jane X. Wang, Zeb Kurth-Nelson, Charles Blundell, Demis Hassabis
https://www.cell.com/trends/cognitive-sciences/fulltext/S1364-6613(19)30061-0
Видео Reinforcement Learning, Fast and Slow канала Yannic Kilcher
Deep reinforcement learning (RL) methods have driven impressive advances in artificial intelligence in recent years, exceeding human performance in domains ranging from Atari to Go to no-limit poker. This progress has drawn the attention of cognitive scientists interested in understanding human learning. However, the concern has been raised that deep RL may be too sample-inefficient – that is, it may simply be too slow – to provide a plausible model of how humans learn. In the present review, we counter this critique by describing recently developed techniques that allow deep RL to operate more nimbly, solving problems much more quickly than previous methods. Although these techniques were developed in an AI context, we propose that they may have rich implications for psychology and neuroscience. A key insight, arising from these AI methods, concerns the fundamental connection between fast RL and slower, more incremental forms of learning.
Authors: Matthew Botvinick, Sam Ritter, Jane X. Wang, Zeb Kurth-Nelson, Charles Blundell, Demis Hassabis
https://www.cell.com/trends/cognitive-sciences/fulltext/S1364-6613(19)30061-0
Видео Reinforcement Learning, Fast and Slow канала Yannic Kilcher
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
What a day in AI! (Sora, Gemini 1.5, V-JEPA, and lots of news)Did Google fake their Gemini Video?Reinforced Self-Training (ReST) for Language Modeling (Paper Explained)[ML News] GPT-4 solves MIT Exam with 100% ACCURACY | OpenLLaMA 13B releasedTree-Ring Watermarks: Fingerprints for Diffusion Images that are Invisible and Robust (Explained)RWKV: Reinventing RNNs for the Transformer Era (Paper Explained)Tree of Thoughts: Deliberate Problem Solving with Large Language Models (Full Paper Review)OpenAI suggests AI licenses (US Senate hearing on AI regulation w/ Sam Altman)[ML News] Geoff Hinton leaves Google | Google has NO MOAT | OpenAI down half a billionScaling Transformer to 1M tokens and beyond with RMT (Paper Explained)AI Alignment Livestream (aka OpenAssistant "Just Chatting")OpenAssistant First Models are here! (Open-Source ChatGPT)The biggest week in AI (GPT-4, Office Copilot, Google PaLM, Anthropic Claude & more)GPT-4 is here! What we know so far (Full Analysis)This ChatGPT Skill will earn you $10B (also, AI reads your mind!) | ML NewsLLaMA: Open and Efficient Foundation Language Models (Paper Explained)Open Assistant Inference Backend Development (Hands-On Coding)OpenAssistant - ChatGPT's Open Alternative (We need your help!)Open Assistant Live Coding (Open-Source ChatGPT Replication)AI Essay Competition (lab42)Open Assistant Live Coding (Open-Source ChatGPT Replication)