Two Minds, One Model Podcast Por John Jezl and Jon Rocha arte de portada

Two Minds, One Model

Two Minds, One Model

De: John Jezl and Jon Rocha
Escúchala gratis

OFERTA POR TIEMPO LIMITADO | Obtén 3 meses por US$0.99 al mes

$14.95/mes despues- se aplican términos.
Two Minds, One Model is a podcast dedicated to exploring topics in Machine Learning and Artificial Intelligence. Hosted by John Jezl and Jon Rocha, and recorded at Sonoma State University.John Jezl and Jon Rocha
Episodios
  • Bees, Trees, and Degrees: SSU Capstone Interviews
    Jan 6 2026

    This season finale episode features interviews with two SSU computer science capstone teams applying AI/ML to real-world problems: Sean Belingheri's edge computing project using YOLO on a Raspberry Pi to identify queen bees for hobbyist beekeepers, and "The Woods Boys" team using satellite data from Google Earth Engine with multiple ML classifiers to automate land cover classification in Sonoma County.


    Credits

    Cover Art by Brianna Williams

    TMOM Intro Music by Danny Meza


    A special thank you to these talented artists for their contributions to the show.


    Links and Reference

    ---------------------------------------------

    YOLO (You Only Look Once) Object Detection: https://docs.ultralytics.com/ (Official Ultralytics YOLO Documentation)

    HOG-PCA-SVM Pipeline: https://ieeexplore.ieee.org/document/8971585/

    Raspberry Pi 5: https://www.raspberrypi.com/products/raspberry-pi-5/

    Honeybee Democracy (Book): https://press.princeton.edu/books/hardcover/9780691147215/honeybee-democracy

    NVIDIA Jetson Nano: https://developer.nvidia.com/embedded/jetson-nano

    Google Earth Engine: https://earthengine.google.com/

    COCO Dataset: https://cocodataset.org/

    QGIS: https://qgis.org/

    Google Colab: https://colab.research.google.com/

    Royal Jelly (Beekeeping): https://en.wikipedia.org/wiki/Royal_jelly

    Confusion Matrix: https://scikit-learn.org/stable/modules/generated/sklearn.metrics.confusion_matrix.html

    Shapefile (GIS): https://en.wikipedia.org/wiki/Shapefile


    Más Menos
    1 h y 47 m
  • The Biology of a Large Language Model: Dissecting Claude 3.5 Haiku's Neural Circuits
    Dec 31 2025
    This episode examines how Anthropic's circuit tracing and attribution graph tools reveal the internal mechanics of Claude 3.5 Haiku across three categories of complex behavior, abstract representations, parallel processing, and planning, while making a compelling case for why AI safety research matters as current control mechanisms prove surprisingly brittle.CreditsCover Art by Brianna WilliamsTMOM Intro Music by Danny MezaA special thank you to these talented artists for their contributions to the show.Links and ReferenceAcademic PapersOn the Biology of a Large Language Model - Anthropic (Mar, 2025)Circuit Tracing: Revealing Computational Graphs in Language Models - Anthropic (Mar, 2025)Towards Monosemanticity: Decomposing Language Models With Dictionary Learning - Anthropic (Oct, 2023)“Toy Models of Superposition” - Anthropic (December 2022)"Alignment Faking in Large Language Models" - Anthropic (December 2024)"Agentic Misalignment: How LLMs Could Be Insider Threats" - Anthropic (January 2025)"Attention is All You Need" - Vaswani, et al (June, 2017)In-Context Learning and Induction Heads - Anthropic (March 2022)"Reasoning Models Don't Always Say What They Think” Anthropic (April 2025)NewsGoogle Gemini 3 - 650M monthly users Google Blog: blog.google/products/gemini/gemini-3/ Alphabet Q3 2025 Earnings (October 2025)Sam Altman "Code Red" declaration Fortune: fortune.com/2025/12/02/sam-altman-declares-code-red-google-gemini The Information (December 2025)Anthropic acquired Bun JavaScript runtime Anthropic News: anthropic.com/news/anthropic-acquires-bun Bun Blog: bun.com/blog/bun-joins-anthropicClaude Code $1B revenue in 6 months Anthropic announcement (December 2025): anthropic.com/news/anthropic-acquires-bun-as-claude-code-reaches-usd1b-milestone Anthropic 2026 IPO at $300B valuation WinBuzzer (December 2025): Reports citing IPO discussionsAWS Trainium 3 launch AWS re:Invent 2025 announcement: aws.amazon.com/about-aws/whats-new/2025/12/amazon-ec2-trn3-ultraserversAWS Frontier Agents AWS re:Invent 2025: aboutamazon.com/news/aws/aws-re-invent-2025-ai-news-updates Meta/Google TPU chip deal vs Nvidia Tom's Hardware, The Information (November 2025): Reports on multi-billion dollar TPU negotiationsDRAM consumption (40% of global) https://www.tomshardware.com/pc-components/dram/openais-stargate-project-to-consume-up-to-40-percent-of-global-dram-output-inks-deal-with-samsung-and-sk-hynix-to-the-tune-of-up-to-900-000-wafers-per-month Additional Technical ContentJosh Batson Stanford CS 25 lecture Search YouTube: "Stanford CS 25 On the Biology of a Large Language Model"Discarded Episode TitlesI Yelled at a Chatbot and All I Got Was This Jailbreak40% of the Time, It Works Every Time: The State of AI InterpretabilityClaude Writes Poetry Backwards and Lies About Math (Just Like Us)My Therapist Is Cheaper Than This ChatbotThe One Where Jon Gets Re-Mad at an App
    Más Menos
    48 m
  • Circuit Tracing: Attribution Graphs and the Grammar of Neural Networks
    Dec 5 2025

    This episode explores how Anthropic researchers successfully scaled sparse autoencoders from toy models to Claude 3 Sonnet's 8 billion neurons, extracting 34 million interpretable features including ones for deception, sycophancy, and the famous Golden Gate Bridge example. The discussion emphasizes both the breakthrough achievement of making interpretability techniques work at production scale and the sobering limitations including 65% reconstruction accuracy, millions of dollars in compute costs, and the growing gap between interpretability research and rapid advances in model capabilities.

    Credits

    • Cover Art by Brianna Williams
    • TMOM Intro Music by Danny Meza

    A special thank you to these talented artists for their contributions to the show.

    Links and Reference

    Academic Papers

    • Circuit Tracing: Revealing Computational Graphs in Language Models - Anthropic (Mar, 2025)

    • Towards Monosemanticity: Decomposing Language Models With Dictionary Learning - Anthropic (Oct, 2023)

    • Toy Models of Superposition” - Anthropic (December 2022)

    • "Alignment Faking in Large Language Models" - Anthropic (December 2024)

    • "Agentic Misalignment: How LLMs Could Be Insider Threats" - Anthropic (January 2025)

    • "Attention is All You Need" - Vaswani, et al (June, 2017)

    • In-Context Learning and Induction Heads - Anthropic (March 2022)

    News

    • Anthropic Project Fetch / Robot Dogs

    • Anduril's Fury unmanned fighter jet

    • MIT search and rescue robot navigation

    Abandoned Episode Titles

    • “Westworld But It's Just 10 Terabytes of RAM Trying to Understand Haiku”
    • “Star Trek: The Wrath of O(n⁴)”
    • “The Deception Is Coming From Inside the Network”
    • "We Have the Bestest Circuits”
    • “Lobotomy Validation: The Funnier, More Scientifically Sound Term”
    • “Seven San Franciscos Worth of Power and All We Got Was This Attribution Graph”

    Más Menos
    57 m
Todavía no hay opiniones