A faster, more efficient way to run AI that works across every type of hardware.
Why we invested
Modern AI relies on a fragile and inefficient foundation. Today's systems depend on thousands of hand-tuned, hardware-specific "kernels" that must be rewritten for every accelerator and every model, locking organizations into particular vendors and leaving enormous amounts of compute power unused. This drives up costs, slows innovation, and forces engineering teams to spend months rewriting code instead of building better AI. Lemurian replaces this kernel-bound world with Tachyon, a new compiler-runtime stack that automatically generates high-performance, hardware-agnostic code. The same code can run on NVIDIA, AMD, Intel, or future accelerators and scale from a single chip to thousands—dramatically increasing utilization and lowering AI serving costs by 60–80%.
Lemurian is uniquely positioned to reshape the AI infrastructure landscape. Their team includes pioneers in compilers, GPU architecture, and large-scale systems—people who built the foundations of Java's HotSpot compiler, contributed to CUDA, and led cutting-edge systems research at places like MIT. This depth has already earned active proof-of-concepts with major cloud providers and strategic alignment from leading hardware vendors. As organizations demand more flexibility, lower costs, and an escape from hardware lock-in, Lemurian's approach offers a path to far more efficient and scalable AI. We believe Tachyon is the kind of platform shift that will define the next era of AI infrastructure.