Reliable, self-improving enterprise AI
Help us reshape how enterprise companies build with LLMs. You'll collaborate with founders, other engineers, and partners to create the management layer for enterprise AI stacks. Expect to tackle the hardest challenges surrounding LLMs in production while working at the cutting edge of opensource models and accelerated compute. We’re at an inflection point: demand is exploding, new customers are pushing us to scale faster than ever, and we’re looking for ambitious engineers to help us meet it head-on.
We move fast and solve hard problems at every level of the stack. You'll be working on high-performance distributed systems that power our LLM proxy and fine-tuning infrastructure, ensuring models run reliably and efficiently. Our backend is built with Python (Quart) and Go, running on Kubernetes across AWS and GCP, with Terraform handling infrastructure as code. You'll optimize real-time LLM autocorrections, request routing, and backend latency, ensuring responses are fast and accurate. You’ll also have the chance to improve our fine-tuning pipeline, balancing speed, cost, and accuracy to make training models as efficient as possible. On the frontend side, you’ll contribute to our Portal (React/TypeScript), where customers configure guardrails, fine-tune models, and test their stack. If you love working across the stack, making systems faster and more reliable, and solving real production AI challenges, you'll thrive here.
Maitai ensures LLMs never fail by optimizing reliability, speed, and resilience. We act as an intelligent proxy, applying real-time autocorrections, rerouting requests, and fine-tuning models to maximize performance. We're growing fast, well-capitalized, and seizing a massive opportunity to redefine how enterprise companies build with AI. Our technology enables businesses to create AI models that are 10x faster and more accurate than closed-source alternatives, while also providing more reliable inference through online guardrails that catch and fix model failures in real time—unlocking entirely new product possibilities. Top YC startups and public companies rely on us to manage their model stack.
Let’s Chat (Video Call)
Hop on a quick 15-minute call and let’s just chat about what you’ve done and what you want to do.
Show Us Something
Build a simple app using the Maitai client to show you know how to work around LLMs, prove your competency in various parts of the stack, and show off your ability to make good product decisions.
Meet Up (In Person)
Join us for coffee or lunch. We want to make sure that there’s good fit - you want to work with us and we want to work with you.
We don’t drag out the process - we’ll make a decision quickly.
Maitai manages the LLM stack for enterprise companies, enabling the fastest and most reliable inference. The future of enterprise AI revolves around mosaics of small, domain-specific models powering powerful, responsive agents, and Maitai is well positioned to capture the market. If you're looking at getting in early with a company redefining how large companies build with AI, then let's talk.