Ai Dipatch: Ai Training Make Superhuman Unreliable
No se pudo agregar al carrito
Solo puedes tener X títulos en el carrito para realizar el pago.
Add to Cart failed.
Por favor prueba de nuevo más tarde
Error al Agregar a Lista de Deseos.
Por favor prueba de nuevo más tarde
Error al eliminar de la lista de deseos.
Por favor prueba de nuevo más tarde
Error al añadir a tu biblioteca
Por favor intenta de nuevo
Error al seguir el podcast
Intenta nuevamente
Error al dejar de seguir el podcast
Intenta nuevamente
-
Narrado por:
-
De:
AI is passing the bar exam, acing medical licensing tests, and crushing coding challenges. So why does research show these same systems fail more than 90% of the time on tasks lasting over four hours?
The answer lies in how AI gets trained—and the limitations that process bakes in from the start.
In this episode of Surviving AI, we go deep on the training problem: the gap between benchmark performance and real-world reliability that creates both risks and opportunities for your career.
What you'll learn:
- How Reinforcement Learning from Human Feedback (RLHF) optimizes AI to sound right rather than be right—and why this creates "sycophantic" systems that tell you what you want to hear
- The "4-Hour Rule": why AI succeeds on quick tasks but struggles with complex, sustained work—and what that means for job vulnerability
- Four predictable failure modes you can learn to spot: temporal blindness, distribution shift, benchmark theater, and inherited bias
- Why 98% of companies feel urgency to deploy AI while only 13% are actually ready—and what happens in that gap
- The three emerging roles that become MORE valuable as AI capabilities grow: the Validator, the Translator, and the Accountability Layer
- Specific questions to ask when AI enters your workplace for hiring, strategy, or workforce decisions
Key research discussed:
- "Open Problems and Fundamental Limitations of RLHF" (Casper et al.)
- METR's research on AI task completion by duration
- Cisco's 2024 AI Readiness Index
- Stanford's Foundation Model Transparency Index
- Deloitte's findings on executive decisions based on AI hallucinations
The bottom line: The gap between what AI benchmarks measure and what work actually requires is your competitive advantage. This episode shows you exactly where to find it.
Todavía no hay opiniones