What happens when you apply the scaling laws of large language models to the physical work of atoms? Elad Gil sits down with Liam Fedus, co-founder at Periodic Labs, which is pioneering an AI foundation lab for atoms. Liam discusses how he pivoted from dark matter physics research to the front lines of artificial intelligence, including stints at Google Brain and working on ChatGPT at OpenAI. He talks about how Periodic is connecting massive language models to the physical world to overcome data bottlenecks in material science. Liam also shares how they use language models as an orchestration layer operating alongside specialized neural nets to run closed-loop physical experiments. They also explore the future of AGI and ASI, as well as the role of robotics in lab automation. Sign up for new podcasts every week. Email feedback to [email protected] Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @LiamFedus | @periodiclabs Chapters: 00:00 – Cold Open 00:05 – Liam Fedus Introduction 00:39 – Liam’s Background at Google Brain, OpenAI 05:14 – From ChatGPT to Materials and Atoms 06:34 – Training Data in the Physical World 09:52 – Generalization Across Domains 11:31 – Models as an Orchestration Layer 12:48 – Commercialization and Business Model 16:10 – How Periodic’s Success May Shape the Future 17:45 – Multidisciplinary Scaling 19:41 – Capital and Compute 21:12 – Hiring at Periodic 21:44 – Thoughts on AGI and ASI 23:30 – Timeline for Machine-Directed Self-Improvement 25:39 – Automation and Data Generation 27:59 – Why Liam is Excited About the Future of Robotics 29:25 – Conclusion
Fler avsnitt av No Priors: Artificial Intelligence | Technology | Startups
Visa alla avsnitt av No Priors: Artificial Intelligence | Technology | StartupsNo Priors: Artificial Intelligence | Technology | Startups med Conviction finns tillgänglig på flera plattformar. Informationen på denna sida kommer från offentliga podd-flöden.
