The Limits of Classical Computing
A Complete History, Present Crisis, and Future Beyond the Binary Paradigm
No se pudo agregar al carrito
Solo puedes tener X títulos en el carrito para realizar el pago.
Add to Cart failed.
Por favor prueba de nuevo más tarde
Error al Agregar a Lista de Deseos.
Por favor prueba de nuevo más tarde
Error al eliminar de la lista de deseos.
Por favor prueba de nuevo más tarde
Error al añadir a tu biblioteca
Por favor intenta de nuevo
Error al seguir el podcast
Intenta nuevamente
Error al dejar de seguir el podcast
Intenta nuevamente
Prueba gratis de 30 días de Audible Standard
Selecciona 1 audiolibro al mes de nuestra colección completa de más de 1 millón de títulos.
Es tuyo mientras seas miembro.
Obtén acceso ilimitado a los podcasts con mayor demanda.
Plan Standard se renueva automáticamente por $8.99 al mes después de 30 días. Cancela en cualquier momento.
Compra ahora por $12.00
-
Narrado por:
-
Virtual Voice
-
De:
-
Richard Murch
Este título utiliza narración de voz virtual
Voz Virtual es una narración generada por computadora para audiolibros..
The deepest constraint is the transistor limit. For decades, Moore's Law predicted that the number of transistors on a chip would double roughly every two years, and it held remarkably well. But transistors are now measured in just a handful of atoms across. At that scale, quantum effects like electron tunneling cause bits to leak and flip unpredictably.
You can't build a smaller classical switch because the laws of physics simply stop cooperating. Heat is the companion problem: packing billions of tiny switches that toggle billions of times per second generates enormous thermal energy in a tiny space. Data centers already consume electricity on the scale of small nations, and conventional chips are approaching the point where cooling them requires more engineering than running them.
.
Most critically, some problems are computationally hard in a way that classical hardware cannot escape. Simulating the quantum behavior of molecules, optimizing across enormous solution spaces, or breaking modern encryption requires resources that scale exponentially with the size of the problem. Double the number of variables and the computation time can square or cube. A classical computer handed a problem with 300 interacting quantum particles would need more operations to simulate it than there are atoms in the observable universe. These aren't engineering problems — they're mathematical ones.
What comes next is a portfolio of approaches rather than a single successor. Quantum computing leverages superposition and entanglement to process certain calculations in fundamentally different ways, with algorithms that could slash exponential problems down to polynomial ones — particularly for chemistry simulation, cryptography, and optimization. Neuromorphic computing takes inspiration from biological brains, building chips that process information with spiking signals closer to how neurons fire, consuming far less power for AI-style inference. Analog computing is seeing a quiet renaissance for specific tasks, trading digital precision for the physical efficiency of computing with continuous voltages rather than discrete bits. And at the architecture level, processing-in-memory chips are beginning to blur the line between storage and computation, attacking the von Neumann bottleneck directly.
The honest picture is that classical computers will remain dominant for most tasks for a long time — quantum machines today are fragile, error-prone, and operate at temperatures colder than outer space.
The future is likely a hybrid one: classical processors handling general-purpose logic, while specialized co-processors — quantum, neuromorphic, or analog — handle the specific problem types where classical silicon hits its wall.
Todavía no hay opiniones