ScaleFlux, FarmGPU, and Lightbits Labs have announced a collaborative architecture that addresses one of the most persistent challenges in AI inference: memory and I/O constraints imposed by long-context workloads. The joint solution, featuring Lightbits’ LightInferra software running on ScaleFlux high-performance NVMe storage within FarmGPU’s managed inference environment, intelligently persists and streams KV-cache data to eliminate GPU stalls, reduce latency, and significantly improve infrastructure efficiency. The companies will publicly debut the implementation at NVIDIA GTC in San Jose, March 16–19, 2026, at ScaleFlux booth 7006.
As AI models adopt longer context windows to handle complex conversations, documents, and reasoning tasks, KV-cache memory demands grow exponentially—often exceeding GPU capacity and causing stalls during recomputation. Traditional approaches rely on limited GPU memory or inefficient recomputation, leading to unpredictable latency, wasted compute, and poor scalability.
The collaborative solution transforms KV-cache from a reactive, GPU-bound cache into an intelligent, streamed data layer:
“We’re transforming inference memory from a reactive cache into an intelligent, streamed data layer,” said Arthur Rassmuson, Director of AI Architecture at Lightbits Labs. “By prefetching only the data that matters and delivering it to GPUs over high-speed RDMA before it's needed, we eliminate the stalls that traditionally limit long-context performance. The result is lower Time-to-First-Token (TTFT), more stable throughput under real-world load, and significantly higher effective GPU utilization. For enterprises, that means serving larger models and longer conversations at lower infrastructure cost—and for end users, it means faster, smoother, more responsive AI experiences.”
“Fast networked storage from Lightbits unlocks a lot of new use cases for long context inference,” said Jonmichael Hands, Chief Executive Officer at FarmGPU. “By pairing our managed service with Lightbit’s high-performance storage running on ScaleFlux NVMe, we are able to lower time to first token and increase utilization on GPUs, drastically lowering the TCO for inference.”
“As members of the NVIDIA Magnum IO GPU Direct Network, we see this as an opportunity to collaborate openly with the ecosystem,” said Keith McKay, Senior Director of Solutions Architecture and Technical Partnerships at ScaleFlux. “What we’re showing at GTC is an early look at how smarter data placement and persistent attention state management could help inference systems stay responsive as context windows grow. This is very much a collaboration we want to shape alongside real operators.”
This early-stage collaboration invites design partners and operators running large-scale or long-context inference to provide feedback and shape future development. Attendees at NVIDIA GTC are encouraged to visit ScaleFlux booth 7006 for live demonstrations and discussions with engineers from all three companies.
About ScaleFlux
ScaleFlux advances Flash Storage and CXL Memory with breakthrough performance, efficiency, security, and scalability for AI/ML workloads and demanding applications in data centers, enterprise and edge infrastructure.
About FarmGPU
FarmGPU is redefining the future of GPU-powered cloud computing by offering cost-effective, scalable, and high-performance GPU resources tailored specifically for AI developers, innovative startups, and enterprises worldwide.
About Lightbits Labs
Lightbits Labs® (Lightbits) invented the NVMe over TCP storage protocol and embedded it natively in their software-defined block storage to deliver ultra-low latency and exceptional throughput while leveraging commodity infrastructure—essential for reducing the cost and complexity of data infrastructure at scale. Built from the ground up for high performance, scalability, resiliency, and cost efficiency, Lightbits software delivers the best price-performance value for real-time analytics, transactional, and AI workloads.