Gimlet Labs just raised an $80 million Series A for tech that lets AI run across NVIDIA, AMD, Intel, ARM, Cerebras and ...
Gimlet Labs raises $80M in Series A funding to tackle the AI inference bottleneck with a new multi-silicon cloud platform.
The edge inference conversation has been dominated by latency. Read any survey paper, attend any infrastructure conference, ...
KubeCon Europe 2026 made AI inference its central focus with major CNCF donations including llm-d, Nvidia's GPU DRA driver and a growing AI conformance program.
“The rapid release cycle in the AI industry has accelerated to the point where barely a day goes past without a new LLM being announced. But the same cannot be said for the underlying data,” notes ...
At the center of this gap are five systemic dysfunctions that reinforce one another: communication bottlenecks, memory ...
As AI workloads shift from centralized training to distributed inference, the network faces new demands around latency requirements, data sovereignty boundaries, model preferences, and power ...
Paper: "Robust Nonparametric Bias-Corrected Inference in the Regression Discontinuity Design", (joint work with Sebastian Calonico and Rocio Titiunik).