I wonder if gwern has changed their view on RL/meta-learning at all given GPT, scaling laws, and current dominance of training on big offline datasets. This would be somewhat in line with skybrian’s comment on Hacker News: https://news.ycombinator.com/item?id=13231808
I wonder if gwern has changed their view on RL/meta-learning at all given GPT, scaling laws, and current dominance of training on big offline datasets. This would be somewhat in line with skybrian’s comment on Hacker News: https://news.ycombinator.com/item?id=13231808
I see that is has references to papers from this year, so presumably has been updated to reflect any changes in view.