Run serverless GPUs on your own cloud
As a key member of our team, you will be responsible for designing our managed solution, which includes:
We are looking for someone who:
Tensorfuse makes it easy to deploy and auto-scale AI models on your own infra using the CLI. It’s like using Modal/Replicate/Together with your cloud credits.
Tensorfuse automatically scales in response to the amount of traffic your app receives. Fast cold boots with our optimized container system. Describe container images and hardware specifications in simple Python. No YAML.
Behind the scenes, we manage K8s + Ray clusters, enabling you to scale without LLMOps overhead.