Announcing our $5.3M Seed Round
We're excited to announce that Luminal has raised a $5.3M seed round to bring speed-of-light inference to everyone. Our round was led by Felicis Ventures, with incredible angels like Paul Graham, Guillermo Rauch, and many more.
The Software Problem
As increasingly powerful models begin to accelerate various parts of the global economy, demand for compute continues to skyrocket. Every week a new article breaks about some multi-billion dollar datacenter buildout or compute partnership. To meet these demands, the semiconductor industry has shifted to an accelerated pace of development, releasing chips capable of higher and higher FLOPs / $ and FLOPs / watt.
Meanwhile, the software that runs on those chips continues to lag far behind, leading to huge swatchs of these chips running dark and unutilized. The best chips in the world are only as good as their software, as seen on Nvidia's Hopper generation only reaching software maturity a full 2 years after release. The problem is only getting worse: as chip complexity increases, speed-of-light (peak) performance is increasingly out of reach for developers.
A Compiled Cloud
Luminal is building a future where reaching full hardware utilization (and positive unit economics) is as simple as running luminal.deploy(). AI companies should get back to worrying about their customers and product, not niche CUDA instructions and complex inference infrastructure.
We're building a tightly integrated high-performance compiler and inference cloud to overcome this "software bottleneck". We believe large-scale kernel search holds the key to enabling speed-of-light performance on a wide variety of accelerators, from GPUs to ASICs. And we believe the best way to deliver this capability is in a tightly integrated, high-performance inference cloud.
An Open Source Future
From the start, Luminal has been an open source project, with incredible community backing and adoption. For us to truely fulfill our mission of speed-of-light inference for all, building the core of our compiler in the open lets us build with the community and lets developers build and run on their own hardware.
Given the sheer complexity involved in solving accelerated computing, no single company can do it alone. If you're an AI engineer excited about deleting 90% of the complexity in AI, come build with us!
Looking Forward
We're working with companies running custom models to drive down latency and increase throughput in our deployments. If you want your models running faster and cheaper, sign up here and we'll reach out.