Michael Dennis (@michaeld1729) 's Twitter Profile
Michael Dennis

@michaeld1729

RS @DeepMind. Works on Unsupervised Environment Design, Problem Specification, Game/Decision Theory, RL, AIS. prev @CHAI_Berkeley

ID: 1199958835508592640

linkhttps://www.michaeldennis.ai/ calendar_today28-11-2019 07:51:14

1,1K Tweet

2,2K Followers

740 Following

Tim Rocktäschel (@_rockt) 's Twitter Profile Photo

Today, Edward Hughes and Michael Dennis from Google DeepMind's Open-Endedness Team will be presenting "Open-Endedness is Essential for Artificial Superhuman Intelligence" as an Oral at 4:30pm in Hall C1-3.

AutoRL Workshop (@autorl_workshop) 's Twitter Profile Photo

We have a speaker change: instead of Jack Parker-Holder we'll hear from Michael Dennis - the focus of the talk is the same, though, so join if you're interested in generating any environments you can imagine!

David Abel (@dabelcs) 's Twitter Profile Photo

New #RLC2024 paper Three Dogmas of Reinforcement Learning joint w/ Mark Ho and Anna Harutyunyan! arxiv.org/pdf/2407.10583 We reflect on where our scientific paradigm needs adjustment, and suggest three departures from previous conventions. Curious to hear what folks think! 🧵

New #RLC2024 paper Three Dogmas of Reinforcement Learning joint w/ <a href="/mark_ho_/">Mark Ho</a> and <a href="/aharutyu/">Anna Harutyunyan</a>!

arxiv.org/pdf/2407.10583

We reflect on where our scientific paradigm needs adjustment, and suggest three departures from previous conventions. Curious to hear what folks think!

🧵
Eugene Vinitsky (@eugenevinitsky) 's Twitter Profile Photo

On the eve of RL_Conference, some thoughts: 1. If we didn't have this conference, I can't imagine a workshop like Finding The Frame Workshop existing. It's important but just wouldn't fly elsewhere 2. The level of community excitement and involvement has been WILD 1/3

Ben Eysenbach (@ben_eysenbach) 's Twitter Profile Photo

I'm excited to share recent work with Grace Liu and Michael Tang on exploration in RL! A Single Goal is All You Need: Skills and Exploration Emerge from Contrastive RL without Rewards, Demonstrations, or Subgoals Paper, code, and videos: graliuce.github.io/sgcrl/ A thread.

I'm excited to share recent work with <a href="/GraceLiu78/">Grace Liu</a> and <a href="/_michaeltang_/">Michael Tang</a> on exploration in RL!

A Single Goal is All You Need: Skills and Exploration Emerge from Contrastive RL without Rewards, Demonstrations, or Subgoals

Paper, code, and videos: graliuce.github.io/sgcrl/

A thread.
Yuhe Nie (@nieyuhe) 's Twitter Profile Photo

🚀 Excited to share our latest research on combining traditional Procedural Content Generation (PCG) algorithms with Large Language Models (LLMs) to enhance text-controlled game map generation! 🗺

🚀 Excited to share our latest research on combining traditional Procedural Content Generation (PCG) algorithms with Large Language Models (LLMs) to enhance text-controlled game map generation! 🗺
Julian Togelius (@togelius) 's Twitter Profile Photo

There are many game content generators out there, mostly tailored to specific games. But what if you want to just describe the content you want in words? We present Moonshine: a method for distilling game content generators into text-conditioned generative models!

There are many game content generators out there, mostly tailored to specific games. But what if you want to just describe the content you want in words? We present Moonshine: a method for distilling game content generators into text-conditioned generative models!
Joth (@jothwip) 's Twitter Profile Photo

People were dunking on this sort of stuff for being too jank *this year* and the fact we're at this point already is insane x.com/_rockt/status/…

Julian Togelius (@togelius) 's Twitter Profile Photo

My lab had been working on scaling up environment/level generation via reinforcement learning for a while now, and we have some nice results to share with you! PCGRL+, our updated PCGRL framework, is not only an order of magnitude faster but can also scale to larger levels than

Michael Dennis (@michaeld1729) 's Twitter Profile Photo

People using RL to generate levels for UED should know about PCGRL and PCGRL+ using RL for proc gen more broadly They get more generalizable RL level generators by varying level size, fixing level features, and shrinking its context - would be good to know for PAIRED-like algs!

Michael Dennis (@michaeld1729) 's Twitter Profile Photo

Interesting thoughts on UED regret estimates! There's a tradeoff between promoting generalisation and learnability in UED, regret is a good proxy for both but there are likely even better proxies for learnability I'ld want learnability early in training and generalization late

xuan (ɕɥɛn / sh-yen) (@xuanalogue) 's Twitter Profile Photo

Should AI be aligned with human preferences, rewards, or utility functions? Excited to finally share a preprint that Micah Carroll Matija Franklin Hal Ashton & I have worked on for almost 2 years, arguing that AI alignment has to move beyond the preference-reward-utility nexus!

Should AI be aligned with human preferences, rewards, or utility functions?

Excited to finally share a preprint that <a href="/MicahCarroll/">Micah Carroll</a> <a href="/FranklinMatija/">Matija Franklin</a> <a href="/hal_ashton/">Hal Ashton</a> &amp; I have worked on for almost 2 years, arguing that AI alignment has to move beyond the preference-reward-utility nexus!