Locust Performance Testing with AI and Observability with Lars Holmberg
No se pudo agregar al carrito
Solo puedes tener X títulos en el carrito para realizar el pago.
Add to Cart failed.
Por favor prueba de nuevo más tarde
Error al Agregar a Lista de Deseos.
Por favor prueba de nuevo más tarde
Error al eliminar de la lista de deseos.
Por favor prueba de nuevo más tarde
Error al añadir a tu biblioteca
Por favor intenta de nuevo
Error al seguir el podcast
Intenta nuevamente
Error al dejar de seguir el podcast
Intenta nuevamente
-
Narrado por:
-
De:
Performance testing often fails for one simple reason: teams can't see where the slowdown actually happens.
In this episode, we explore Locust load testing and why Python-based performance testing is becoming the go-to choice for modern DevOps, QA, and SRE teams. You'll learn how Locust enables highly realistic user behavior, massive concurrency, and distributed load testing — without the overhead of traditional enterprise tools.
We also dive into:
Why Python works so well for AI-assisted load testing
- How Locust fits naturally into CI/CD and GitHub Actions
- The real difference between load testing vs performance testing
- How observability and end-to-end tracing eliminate guesswork
- Common performance testing mistakes even experienced teams make
Whether you're a software tester, automation engineer, or QA leader looking to shift-left performance testing, this conversation will help you design smarter tests and catch scalability issues before your users do.
Todavía no hay opiniones