At Scale AI, we’re not just building AI tools—we’re pioneering the next era of enterprise AI. As businesses race to harness the power of Generative AI, Scale is at the forefront, delivering cutting-edge solutions that transform workflows, automate complex processes, and drive unparalleled efficiency for the largest enterprises. Our Scale Generative AI Platform (SGP) provides foundational services and APIs, enabling businesses to seamlessly integrate AI into their operations at production scale.
We’re looking for a Backend Engineer to help bring large-scale GenAI systems to production. In this role, you’ll build the core infrastructure that powers AI products for some of the world’s largest enterprises—designing scalable APIs, distributed data systems, and robust deployment pipelines that enable production-grade reliability and performance.
This is a rare opportunity to be at the center of the GenAI revolution, solving hard backend and infrastructure challenges that make AI truly work at enterprise scale. If you're excited about shaping how AI systems are deployed and scaled in the real world, we want to hear from you.
At Scale, we don’t just follow AI advancements — we lead them. Backed by deep expertise in data, infrastructure, and model deployment, we are uniquely positioned to solve the hardest problems in AI adoption. Join us in shaping the future of enterprise AI, where your work will directly impact how businesses operate, innovate, and grow in the age of GenAI.
You Will:
- Design, build, and scale backend systems that power enterprise GenAI products, focusing on reliability, performance, and deployment across both Scale’s and customers’ infrastructure.
- Develop core services and APIs that integrate AI models and enterprise data sources securely and efficiently, enabling production-scale AI adoption.
- Architect scalable distributed systems for data processing, inference, and orchestration of large-scale GenAI workloads.
- Optimize backend performance for latency, throughput, and cost—ensuring AI applications can operate at enterprise scale across hybrid and multi-cloud environments.
- Manage and evolve cloud infrastructure (AWS, Azure, or GCP), driving automation, observability, and security for large-scale AI deployments.
- Collaborate with ML and product teams to bring cutting-edge GenAI models into production through efficient APIs, model serving systems, and evaluation frameworks.
- Continuously improve reliability and scalability, applying strong engineering practices to make AI systems robust, maintainable, and enterprise-ready.
Ideally, You Have:
- 4+ years of experience developing large-scale backend or infrastructure systems, with a strong emphasis on distributed services, reliability, and scalability.
- Proficiency in Python or TypeScript, with experience designing high-performance APIs and backend architectures using frameworks such as FastAPI, Flask, Express, or NestJS.
- Deep familiarity with cloud infrastructure (AWS and Azure preferred), including container orchestration (Kubernetes, Docker) and Infrastructure-as-Code tools like Terraform.
- Experience managing data systems such as relational and NoSQL databases (PostgreSQL, DynamoDB, etc.) and building pipelines for data-intensive applications.
- Hands-on experience with GenAI applications, model integration, or AI agent systems—understanding how to deploy, evaluate, and scale AI workloads in production.
- Strong understanding of observability, CI/CD, and security best practices for running services in enterprise or multi-tenant environments.
- Ability to balance rapid iteration with production-grade quality, shipping reliable backend systems in fast-paced environments.
Collaborative mindset, working closely with ML, infra, and product teams to bring complex GenAI systems into production at enterprise scale.