This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
As Microsoft continues to push the boundaries of AI, we are on the lookout for passionate leaders to help us tackle the most interesting and challenging AI questions of our time. Our vision is bold and broad, to build systems that have true artificial intelligence across agents, applications, services, and infrastructure. It's also inclusive: we aim to make AI accessible to all, consumers, businesses, developers, so that everyone can realize its benefits. We're looking for a Data Infrastructure Manager to lead a team of talented engineers building and scaling the data infrastructure that powers Microsoft's consumer AI. This role sits at the intersection of technical leadership and people management. You'll set the technical direction for large-scale data and ML pipelines, AI agentic workflows, and intelligent systems while growing a high-performing team of ICs. If you've architected big data platforms from the ground up and are now ready to multiply your impact through others, including on some of the most exciting AI infrastructure challenges in the industry, we want to hear from you.
Job Responsibility:
Team Leadership & People Development
Hire, mentor, and develop a team of Data Infrastructure Engineers, fostering a culture of technical excellence, ownership, and continuous growth
Conduct regular 1:1s, set clear goals, and provide actionable feedback to support each engineer's career development
Build and sustain an inclusive, collaborative team environment aligned with Microsoft's values of Respect, Integrity, Accountability, and Inclusion
Technical Strategy & Architecture
Define and drive the technical vision for a scalable, reliable, and observable Big Data Infrastructure serving mission-critical AI applications, including agentic and intelligent systems
Lead technical design reviews, establish engineering standards, and ensure a clean, secure, and well-documented codebase
Partner with engineers to architect data solutions across storage, compute, and analytics layers, including the pipelines and orchestration frameworks that underpin AI agent workflows, balancing long-term scalability with near-term delivery
Platform & Operations
Champion DevOps and SRE best practices across the team, including automated deployments, service monitoring, and incident response
Guide the team in building a self-service big data platform that empowers data engineers, researchers, and partner teams
Oversee robust CI/CD pipelines and infrastructure-as-code practices using tools like Bicep, Terraform, and ARM
Lead capacity planning and drive proactive resolution of bottlenecks in data pipelines and infrastructure
Cross-Functional Collaboration
Act as a key technical partner to Data Engineers, Data Scientists, AI Researchers, ML Engineers, and Developers to deliver secure, seamless big data workflows
Collaborate with Security teams to uphold strong infrastructure security practices (IAM, OAuth, Kerberos)
Represent the team in planning and prioritization discussions, translating organizational goals into actionable engineering roadmaps
Requirements:
Bachelor's Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 6+ years experience in business analytics, data science, software development, data modeling or data engineering work
OR Master's Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 4+ years experience in business analytics, data science, software development, or data engineering work
OR equivalent experience
Deep technical expertise in big data and distributed systems
A track record of leading and developing engineering talent
A passion for automation, observability, and operational excellence
The ability to translate complex technical strategy into clear, executable plans
Empathy, collaboration, and a growth mindset
Nice to have:
Master's Degree in Computer Science or related technical field AND 10+ years of technical engineering experience OR Bachelor's Degree AND 14+ years, OR equivalent experience
5+ years in Big Data Infrastructure, DevOps, SRE, or Platform Engineering
5+ years of hands-on experience with distributed systems from bare-metal to cloud-native environments
5+ years overseeing or contributing to containerized application deployments using Kubernetes and Helm/Kustomize
Solid scripting and automation fluency in Python, Bash, or PowerShell
Proven track record managing CI/CD pipelines, release automation, and production incident response
Hands-on expertise with modern data platforms like Databricks, including deep familiarity with relational and NoSQL databases, key-value stores, Spark compute engines, distributed file systems (e.g., HDFS, ADLS Gen2), and messaging systems (e.g., Event Hub, Kafka, RabbitMQ)
Proven experience with cloud-native infrastructure across Azure, AWS, or GCP
Strong collaboration history with Data Engineers, Data Scientists, ML Engineers, Networking, and Security teams
Experience with agentic workflow infrastructure, including orchestration frameworks (e.g., Semantic Kernel, AutoGen), retrieval pipelines, and the data infrastructure patterns that support multi-agent systems at scale
Familiarity with modern web stacks: TypeScript, Node.js, React, and optionally PHP