No Priors: Artificial Intelligence | Technology | Startups Podcast Por Conviction arte de portada

No Priors: Artificial Intelligence | Technology | Startups

No Priors: Artificial Intelligence | Technology | Startups

De: Conviction
Escúchala gratis

Acerca de esta escucha

At this moment of inflection in technology, co-hosts Elad Gil and Sarah Guo talk to the world's leading AI engineers, researchers and founders about the biggest questions: How far away is AGI? What markets are at risk for disruption? How will commerce, culture, and society change? What’s happening in state-of-the-art in research? “No Priors” is your guide to the AI revolution. Email feedback to show@no-priors.com. Sarah Guo is a startup investor and the founder of Conviction, an investment firm purpose-built to serve intelligent software, or "Software 3.0" companies. She spent nearly a decade incubating and investing at venture firm Greylock Partners. Elad Gil is a serial entrepreneur and a startup investor. He was co-founder of Color Health, Mixer Labs (which was acquired by Twitter). He has invested in over 40 companies now worth $1B or more each, and is also author of the High Growth Handbook.© Copyright 2023 Conviction. All Rights Reserved. Ciencia Economía Gestión y Liderazgo Liderazgo
Episodios
  • Meet AlphaEvolve: The Autonomous Agent That Discovers Algorithms Better Than Humans With Google DeepMind’s Pushmeet Kohli and Matej Balog
    Jun 26 2025
    Much of the scientific process involves searching. But rather than continue to rely on the luck of discovery, Google DeepMind has engineered a more efficient AI agent that mines complex spaces to facilitate scientific breakthroughs. Sarah Guo speaks with Pushmeet Kohli, VP of Science and Strategic Initiatives, and research scientist Matej Balog at Google DeepMind about AlphaEvolve, an autonomous coding agent they developed that finds new algorithms through evolutionary search. Pushmeet and Matej talk about how AlphaEvolve tackles the problem of matrix multiplication efficiency, scaling and iteration in problem solving, and whether or not this means we are at self-improving AI. Together, they also explore the implications AlphaEvolve has to other sciences beyond mathematics and computer science. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @pushmeet | @matejbalog Chapters: 00:00 Pushmeet Kohli and Matej Balog Introduction 0:48 Origin of AlphaEvolve 02:31 AlphaEvolve’s Progression from AlphaGo and AlphaTensor 08:02 The Open Problem of Matrix Multiplication Efficiency 11:18 How AlphaEvolve Evolves Code 14:43 Scaling and Predicting Iterations 16:52 Implications for Coding Agents 19:42 Overcoming Limits of Automated Evaluators 25:21 Are We At Self-Improving AI? 28:10 Effects on Scientific Discovery and Mathematics 31:50 Role of Human Scientists with AlphaEvolve 38:30 Making AlphaEvolve Broadly Accessible 40:18 Applying AlphaEvolve Within Google 41:39 Conclusion
    Más Menos
    42 m
  • The Operating System for Self Driving Cars (and Tanks, and Trucks...) With Qasar Younis and Peter Ludwig of Applied Intuition
    Jun 17 2025
    When will fully autonomous vehicles see widespread adoption? According to Applied Intuition, that future is closer than you may think. Applied Intuition’s CEO, Qasar Younis, and CTO, Peter Ludwig, talk with Elad Gil about how now is the best time to both work on self-driving vehicle technology and monetize it. Qasar and Peter discuss the advantages of developing their own OS in-house for their autonomous applications, self-driving technology’s potential to drive re-shoring of vehicle manufacturing to the United States, and how best to gauge the bar for safety in autonomous systems. Plus, they explore how self-driving technology may reshape the designs of not only vehicles, but cities themselves. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @qasar | @AppliedInt Chapters: 00:00 Qasar Younis and Peter Ludwig Introduction 01:28 A Primer on Applied Intuition 11:08 Applied Intuition’s Customers 12:04 Impact of Chinese Vehicles Manufacturers 15:44 EV Policies in the European Market 20:49 Can Robotics and Automation Re-Shore Vehicle Manufacturing? 21:53 Training Models for Autonomous Vehicles 26:41 Gauging the Bar for Autonomous Vehicles Safety 32:03 Timeline for Large-Scale Autonomous Vehicle Adoption 36:28 Rethinking Urban Design for Autonomous Vehicles 38:47 How Applied Intuition Uses AI for Tooling and OS 42:09 Designing for User Experience 43:31 Applied Intuition’s Hiring Strategy 45:01 Conclusion
    Más Menos
    45 m
  • Will we have Superintelligence by 2028? With Anthropic’s Ben Mann
    Jun 12 2025
    What happens when you give AI researchers unlimited compute and tell them to compete for the highest usage rates? Ben Mann, Co-Founder, from Anthropic sits down with Sarah Guo and Elad Gil to explain how Claude 4 went from "reward hacking" to efficiently completing tasks and how they're racing to solve AI safety before deploying computer-controlling agents. Ben talks about economic Turing tests, the future of general versus specialized AI models, Reinforcement Learning From AI Feedback (RLAIF), and Anthropic’s Model Context Protocol (MCP). Plus, Ben shares his thoughts on if we will have Superintelligence by 2028. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @8enmann Links: ai-2027.com/ Chapters: 00:00 Ben Mann Introduction 00:33 Releasing Claude 4 02:05 Claude 4 Highlights and Improvements 03:42 Advanced Use Cases and Capabilities 06:42 Specialization and Future of AI Models 09:35 Anthropic's Approach to Model Development 18:08 Human Feedback and AI Self-Improvement 19:15 Principles and Correctness in Model Training 20:58 Challenges in Measuring Correctness 21:42 Human Feedback and Preference Models 23:38 Empiricism and Real-World Applications 27:02 AI Safety and Ethical Considerations 28:13 AI Alignment and High-Risk Research 30:01 Responsible Scaling and Safety Policies 35:08 Future of AI and Emerging Behaviors 38:35 Model Context Protocol (MCP) and Industry Standards 41:00 Conclusion
    Más Menos
    41 m
Todavía no hay opiniones