Fluid Inference is an applied research lab building the future of edge intelligence. We're bridging the gap between advanced AI models and the hardware they run on.
"Fluid Inference's work with Intel has unlocked a new tier of performance and privacy for customers looking to deploy transformer models on local devices. Developers and companies interested in building on-device, NPU-powered AI products can now follow the same path to bring AI native applications to production."— Intel
Our current focus is making it easy for developers to access state-of-the-art voice and language models on-device. No proprietary models, gated SDKs, or restricting licenses. The system is already very fragmented, we don't need to make it worse. You can find all of our native SDKs on our GitHub and our models on Hugging Face.
To get in touch, fill out the form below or email [email protected].