#335 Sriram Raghavan: Why IBM Is Betting Everything on Small AI Models Podcast Por  arte de portada

#335 Sriram Raghavan: Why IBM Is Betting Everything on Small AI Models

#335 Sriram Raghavan: Why IBM Is Betting Everything on Small AI Models

Escúchala gratis

Ver detalles del espectáculo

Why IBM Is Betting Everything on Small AI Models

In this episode of Eye on AI, Craig Smith sits down with Sriram Raghavan, Vice President of AI at IBM Research, to explore one of the most important debates in enterprise AI right now. Do you actually need a massive model to get world class results? IBM's answer is no, and Sriram breaks down exactly why.

Sriram explains why IBM chose to train its Granite models directly using reinforcement learning rather than distilling from larger models like most of the industry. The reason goes beyond performance. It comes down to data lineage, safety alignment, and a belief that small, efficient models are the only sustainable path for enterprises running AI across hybrid cloud environments.

We get into the full technical stack behind that bet. How data quality has replaced model size as the real competitive advantage. Why parameter count is becoming the wrong metric entirely. How IBM's inference time scaling techniques allow an 8 billion parameter model to match the performance of GPT-4o and Claude 3.5 on code and math benchmarks. And why IBM is pioneering a new concept called Generative Computing, which treats AI models not as prompt receivers but as programmable computing elements with runtimes, modular LoRA adapters, and proper programming abstractions.

Sriram also shares where IBM Research is headed next, including breakthroughs in continuous learning, agent orchestration, and making unstructured enterprise data actually usable at scale.

Subscribe for more conversations with the people building the future of AI and emerging technology.

Stay Updated:

Craig Smith on X: https://x.com/craigss

Eye on A.I. on X: https://x.com/EyeOn_AI

(00:00) Why IBM Skips Distillation and Trains Small Models Directly

(04:50) Did We Even Need Giant AI Models in the First Place?

(08:12) How Data Quality Became the New Competitive Moat

(11:54) Why Parameter Count Is the Wrong Way to Measure a Model

(15:36) Reinforcement Learning Without Losing Broad Capabilities

(22:05) Inference Time Scaling: Getting Big Model Results From Small Models

(28:12) Generative Computing: Treating AI as a Programming Element

(36:40) Why IBM Open Sources and How Small Models Make It Sustainable

(41:25) The Path to Continuous Learning Without Rewriting Weights

(51:00) IBM's Full Roadmap: Models, Data, and Agents

Todavía no hay opiniones