#335 Sriram Raghavan: Why IBM Is Betting Everything on Small AI Models
No se pudo agregar al carrito
Add to Cart failed.
Error al Agregar a Lista de Deseos.
Error al eliminar de la lista de deseos.
Error al añadir a tu biblioteca
Error al seguir el podcast
Error al dejar de seguir el podcast
-
Narrado por:
-
De:
Why IBM Is Betting Everything on Small AI Models
In this episode of Eye on AI, Craig Smith sits down with Sriram Raghavan, Vice President of AI at IBM Research, to explore one of the most important debates in enterprise AI right now. Do you actually need a massive model to get world class results? IBM's answer is no, and Sriram breaks down exactly why.
Sriram explains why IBM chose to train its Granite models directly using reinforcement learning rather than distilling from larger models like most of the industry. The reason goes beyond performance. It comes down to data lineage, safety alignment, and a belief that small, efficient models are the only sustainable path for enterprises running AI across hybrid cloud environments.
We get into the full technical stack behind that bet. How data quality has replaced model size as the real competitive advantage. Why parameter count is becoming the wrong metric entirely. How IBM's inference time scaling techniques allow an 8 billion parameter model to match the performance of GPT-4o and Claude 3.5 on code and math benchmarks. And why IBM is pioneering a new concept called Generative Computing, which treats AI models not as prompt receivers but as programmable computing elements with runtimes, modular LoRA adapters, and proper programming abstractions.
Sriram also shares where IBM Research is headed next, including breakthroughs in continuous learning, agent orchestration, and making unstructured enterprise data actually usable at scale.
Subscribe for more conversations with the people building the future of AI and emerging technology.
Stay Updated:
Craig Smith on X: https://x.com/craigss
Eye on A.I. on X: https://x.com/EyeOn_AI
(00:00) Why IBM Skips Distillation and Trains Small Models Directly
(04:50) Did We Even Need Giant AI Models in the First Place?
(08:12) How Data Quality Became the New Competitive Moat
(11:54) Why Parameter Count Is the Wrong Way to Measure a Model
(15:36) Reinforcement Learning Without Losing Broad Capabilities
(22:05) Inference Time Scaling: Getting Big Model Results From Small Models
(28:12) Generative Computing: Treating AI as a Programming Element
(36:40) Why IBM Open Sources and How Small Models Make It Sustainable
(41:25) The Path to Continuous Learning Without Rewriting Weights
(51:00) IBM's Full Roadmap: Models, Data, and Agents