This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
At Scale AI, we’re not just building AI tools—we’re pioneering the next era of enterprise AI. As businesses race to harness the power of Generative AI, Scale is at the forefront, delivering cutting-edge solutions that transform workflows, automate complex processes, and drive unparalleled efficiency for the largest enterprises. Our Scale Generative AI Platform (SGP) provides foundational services and APIs, enabling businesses to seamlessly integrate AI into their operations at production scale. We’re looking for a Backend Engineer to help bring large-scale GenAI systems to production. In this role, you’ll build the core infrastructure that powers AI products for some of the world’s largest enterprises—designing scalable APIs, distributed data systems, and robust deployment pipelines that enable production-grade reliability and performance. This is a rare opportunity to be at the center of the GenAI revolution, solving hard backend and infrastructure challenges that make AI truly work at enterprise scale. If you're excited about shaping how AI systems are deployed and scaled in the real world, we want to hear from you.
Job Responsibility:
Design, build, and scale backend systems that power enterprise GenAI products, focusing on reliability, performance, and deployment across both Scale’s and customers’ infrastructure
Develop core services and APIs that integrate AI models and enterprise data sources securely and efficiently, enabling production-scale AI adoption
Architect scalable distributed systems for data processing, inference, and orchestration of large-scale GenAI workloads
Optimize backend performance for latency, throughput, and cost—ensuring AI applications can operate at enterprise scale across hybrid and multi-cloud environments
Manage and evolve cloud infrastructure (AWS, Azure, or GCP), driving automation, observability, and security for large-scale AI deployments
Collaborate with ML and product teams to bring cutting-edge GenAI models into production through efficient APIs, model serving systems, and evaluation frameworks
Continuously improve reliability and scalability, applying strong engineering practices to make AI systems robust, maintainable, and enterprise-ready
Requirements:
4+ years of experience developing large-scale backend or infrastructure systems, with a strong emphasis on distributed services, reliability, and scalability
Proficiency in Python or TypeScript, with experience designing high-performance APIs and backend architectures using frameworks such as FastAPI, Flask, Express, or NestJS
Deep familiarity with cloud infrastructure (AWS and Azure preferred), including container orchestration (Kubernetes, Docker) and Infrastructure-as-Code tools like Terraform
Experience managing data systems such as relational and NoSQL databases (PostgreSQL, DynamoDB, etc.) and building pipelines for data-intensive applications
Hands-on experience with GenAI applications, model integration, or AI agent systems—understanding how to deploy, evaluate, and scale AI workloads in production
Strong understanding of observability, CI/CD, and security best practices for running services in enterprise or multi-tenant environments
Ability to balance rapid iteration with production-grade quality, shipping reliable backend systems in fast-paced environments
Collaborative mindset, working closely with ML, infra, and product teams to bring complex GenAI systems into production at enterprise scale