CrawlJobs Logo

MLOps Engineer

https://www.roberthalf.com Logo

Robert Half

Location Icon

Location:
United States , Houston

Category Icon

Job Type Icon

Contract Type:
Not provided

Salary Icon

Salary:

Not provided

Job Description:

We are seeking an experienced MLOps Engineer to design, deploy, monitor, and maintain machine learning solutions in production across AWS, Microsoft Azure, and Snowflake environments. This role will collaborate closely with data scientists, platform engineers, and cloud teams to operationalize ML models, automate pipelines, and build reliable, secure, and scalable ML/data platforms.

Job Responsibility:

  • Design and implement end-to-end ML pipelines for ingestion, feature engineering, training, validation, deployment, and monitoring
  • Deploy and manage ML models in production across AWS, Azure, and Snowflake ecosystems
  • Build batch and real-time inference pipelines using cloud-native and platform-native services
  • Automate model packaging, testing, releases, and rollback using CI/CD best practices
  • Integrate ML workflows with AWS SageMaker, AWS Lambda, Azure Machine Learning, Azure Data Factory, and Snowflake
  • Build and maintain orchestration workflows using Airflow, Azure Data Factory, or similar tools
  • Implement experiment tracking, model registries, and model governance processes
  • Monitor model accuracy, drift, latency, throughput, pipeline performance, and infrastructure usage
  • Establish advanced deployment strategies (canary, shadow, blue-green, rollback)
  • Collaborate with cross-functional teams to transition models from research to production
  • Ensure security, compliance, traceability, and access control for ML systems and data
  • Optimize platform reliability, performance, and cost across AWS, Azure, and Snowflake

Requirements:

  • Master’s degree or higher (PhD preferred) in Computer Science, Engineering, or related discipline
  • 5+ years of relevant experience in MLOps, ML engineering, platform engineering, or DevOps
  • Strong hands-on experience with AWS, Microsoft Azure, and Snowflake
  • Proficiency in Python and SQL
  • Proven experience deploying and managing ML models in production environments
  • Experience with AWS SageMaker and Azure Machine Learning
  • Experience building and integrating data pipelines with Snowflake
  • Strong understanding of CI/CD pipelines, infrastructure automation, and model versioning
  • Experience with containers and orchestration tools such as Docker and Kubernetes
  • Experience with Airflow, Azure Data Factory, or similar workflow orchestrators
  • Familiarity with model monitoring, logging, alerting, and observability
  • Strong understanding of data engineering concepts, distributed systems, and APIs
  • Excellent troubleshooting skills and proven cross-team collaboration abilities

Nice to have:

  • Experience with Snowflake Cortex AI, Snowpark, or ML workloads within Snowflake
  • Experience with AWS Bedrock, Azure OpenAI, or production deployment of LLM-based systems
  • Experience building real-time inference pipelines, serverless architectures, and event-driven systems
  • Familiarity with feature stores, vector databases, and RAG-based architectures
  • Experience with IaC tools such as Terraform, CloudFormation, or Azure Bicep
  • Understanding of compliance, governance, and security requirements in regulated industries
  • Experience with A/B testing, shadow deployments, canary releases, and model rollback strategies
What we offer:
  • medical
  • vision
  • dental
  • life and disability insurance
  • 401(k) plan

Additional Information:

Job Posted:
March 26, 2026

Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for MLOps Engineer

MLOps Engineer

As an MLOps Engineer, you will help our clients with automating and managing the...
Location
Location
Belgium , Brussels
Salary
Salary:
Not provided
https://www.soprasteria.com Logo
Sopra Steria
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • At least 3 years of experience as an MLOps Engineer, Machine Learning Engineer or in a similar position
  • Experience managing the machine learning lifecycle in production, and related MLOps frameworks
  • Master CI/CD pipelines and Infrastructure as Code
  • In depth knowledge of containers and container orchestration
  • Solid understanding of Linux and in depth knowledge of at least one cloud provider
  • Programming experience with Python
  • Experience with monitoring and logging technologies
  • Good grasp of big data technologies such as Apache Spark, ElasticSearch, Kafka
  • Able to coach others and give technical advice and direction
  • Understanding of networking, firewalls, and IT infrastructure
Job Responsibility
Job Responsibility
  • Design and implement solutions to automate the experimentation and release cycle of machine learning, from traditional ML to LLMs
  • Identify, design, and implement process improvements in the machine learning lifecycle: automating manual processes, optimizing data delivery, re-design infrastructure for greater scalability, etc
  • Collaborate with Data Scientists, Machine Learning Engineers, Data Engineers, and other IT roles to integrate all parts of the solution
  • Implement CI/CD pipelines, IaC, monitoring for models and infrastructure
What we offer
What we offer
  • A variety of perks, such as mobility options (including a company car), insurance coverage, meal vouchers, eco-cheques, and more
  • Continuous learning opportunities through the Sopra Steria Academy to support your career development
  • The opportunity to connect with fellow Sopra Steria colleagues at various team events
  • Fulltime
Read More
Arrow Right

Staff MLOps Engineer

At Inworld, we’re building the AI framework behind the next generation of real-t...
Location
Location
Canada , Vancouver
Salary
Salary:
190000.00 - 240000.00 CAD / Year
inworld.ai Logo
Inworld AI
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 7+ years of software engineering experience
  • 5+ years of infrastructure-as-code
  • Proficiency in managing Kubernetes clusters and applications, including creating Helm charts/Kustomize manifests for new applications
  • Experience in creating and maintaining CI/CD pipelines for both applications and infrastructure deployments (using tools like Terraform/Terragrunt, ArgoCD, GitHub Actions, Ansible, etc.)
  • Deep knowledge of at least one major cloud provider (Google Cloud Platform, Microsoft Azure, Oracle Cloud)
  • Proficient in at least one backend programming/scripting languages such as Golang, Python, and Bash
  • Knowledge of SLURM or similar job schedulers for distributed training
  • Experience with data pipeline and workflow management tools
  • Desire to work at a fast-growing Series A startup, comfortable with uncertainty, owning and scaling new products, and embracing an experimental and iterative development process
Job Responsibility
Job Responsibility
  • Build and scale MLOps systems to streamline the end-to-end ML model lifecycle on the Inworld AI platform, from training to deployment
  • Design and implement robust model training, evaluation, and release pipelines
  • Collaborate cross-functionally with ML and backend teams to design, deploy, and maintain scalable secure infrastructure for Inworld’s AI Engine and Studio
  • Facilitate a "you build it, you run it" culture by providing the necessary tools and processes for monitoring the reliability, availability, and performance of services
  • Manage CI/CD pipelines to ensure smooth and efficient code integration and deployment
  • Identify and implement opportunities to enhance engineering speed and efficiency
  • Provide technical leadership in ML engineering best practices, raise the technical bar, and mentor junior engineers in MLOps principles
What we offer
What we offer
  • equity
  • benefits
  • Fulltime
Read More
Arrow Right

Staff MLOps Engineer

At Inworld, we’re building the AI framework behind the next generation of real-t...
Location
Location
United States , Mountain View
Salary
Salary:
180000.00 - 280000.00 USD / Year
inworld.ai Logo
Inworld AI
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 7+ years of software engineering experience, with 5+ years of infrastructure-as-code
  • Proficiency in managing Kubernetes clusters and applications, including creating Helm charts/Kustomize manifests for new applications
  • Experience in creating and maintaining CI/CD pipelines for both applications and infrastructure deployments (using tools like Terraform/Terragrunt, ArgoCD, GitHub Actions, Ansible, etc.)
  • Deep knowledge of at least one major cloud provider (Google Cloud Platform, Microsoft Azure, Oracle Cloud)
  • Proficient in at least one backend programming/scripting languages such as Golang, Python, and Bash
  • Knowledge of SLURM or similar job schedulers for distributed training
  • Experience with data pipeline and workflow management tools
  • Desire to work at a fast-growing Series A startup, comfortable with uncertainty, owning and scaling new products, and embracing an experimental and iterative development process
  • In-office location: Mountain View, CA, United States. You must be available for hybrid work
Job Responsibility
Job Responsibility
  • Build and scale MLOps systems to streamline the end-to-end ML model lifecycle on the Inworld AI platform, from training to deployment
  • Design and implement robust model training, evaluation, and release pipelines
  • Collaborate cross-functionally with ML and backend teams to design, deploy, and maintain scalable secure infrastructure for Inworld’s AI Engine and Studio
  • Facilitate a "you build it, you run it" culture by providing the necessary tools and processes for monitoring the reliability, availability, and performance of services
  • Manage CI/CD pipelines to ensure smooth and efficient code integration and deployment
  • Identify and implement opportunities to enhance engineering speed and efficiency
  • Provide technical leadership in ML engineering best practices, raise the technical bar, and mentor junior engineers in MLOps principles
What we offer
What we offer
  • equity and benefits
  • Fulltime
Read More
Arrow Right

MLOps Engineer

WFH flexibility! Up to 4 days/week! Global Environment! Competitive salary!
Location
Location
Japan , Tokyo
Salary
Salary:
8000000.00 - 10000000.00 JPY / Year
https://www.randstad.com Logo
Randstad
Expiration Date
June 10, 2026
Flip Icon
Requirements
Requirements
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field
  • 5+ years of experience in MLOps, DevOps, or a related field
  • Proven experience in deploying and managing machine learning models in production environments
  • Proficiency in programming languages such as Python
  • Strong knowledge of machine learning frameworks and libraries
  • Experience with cloud platforms and services, ideally with Azure ML, but others such as AWS SageMaker, Google AI Platform, or Databricks etc are also acceptable
  • Hands-on experience with MLOps tools and practices, including CI/CD pipelines, containerization (Docker), and orchestration (Kubernetes)
  • Proficiency in SQL and working with large datasets
  • Familiarity with version control systems (e.g., Git) and collaborative development workflows
  • Strong problem-solving skills and attention to detail
Job Responsibility
Job Responsibility
  • Deploying and managing machine learning models in production environments
  • Working with MLOps tools, CI/CD pipelines, containerization, and orchestration
  • Working with cloud platforms and large datasets
What we offer
What we offer
  • Health insurance
  • Employee pension insurance
  • Employment insurance
  • Saturday
  • Sunday
  • Holiday
  • Fulltime
Read More
Arrow Right

Senior MLOps Engineer

A fast-growing, VC-backed AI company is seeking a Senior MLOps Engineer to help ...
Location
Location
United States
Salary
Salary:
Not provided
weareorbis.com Logo
Orbis Consultants
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 4+ years of experience in MLOps with hands-on ownership of production ML systems
  • Degree in Computer Science or a related field (Bachelor’s or Master’s)
  • Strong expertise in Python and ML libraries (TensorFlow, PyTorch, or similar)
  • Experience with ML lifecycle management tools
  • Practical experience deploying and monitoring deep learning models
  • Proficiency with cloud platforms (AWS, Azure, or GCP)
Job Responsibility
Job Responsibility
  • Design, build, and maintain infrastructure for deploying, monitoring, and updating ML models in production
  • Automate end-to-end pipelines covering ingestion, preprocessing, training, validation, and deployment
  • Implement monitoring for model performance, drift, latency, and accuracy
  • Ensure ML systems are secure, scalable, and cost-efficient
  • Partner with ML engineers, data scientists, and product teams to move models seamlessly from research to production
  • Apply software engineering best practices—including CI/CD, testing, and version control—to ML workflows
  • Continuously improve documentation, tooling, and infrastructure
What we offer
What we offer
  • great equity and the chance to join a rocketship
  • Fulltime
Read More
Arrow Right

Data Engineer Senior

We are looking for a highly skilled professional to lead the industrialisation o...
Location
Location
Portugal , Lisbon
Salary
Salary:
Not provided
https://www.inetum.com Logo
Inetum
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Minimum of 5 years’ experience in MLOps, data engineering, or DevOps with a focus on ML/DL/LLM/AI agents in production environments
  • Strong proficiency in Python
  • Hands-on experience with CI/CD tools such as GitLab, Docker, Kubernetes, Jenkins
  • Solid understanding of ML, DL, and LLM models
  • Experience with ML lifecycle tools such as MLflow or DVC
  • Good understanding of model lifecycle, data traceability, and governance frameworks
  • Experience with on-premise and hybrid infrastructures
  • Excellent communication skills and ability to collaborate with remote teams
  • Proactive mindset, technical rigour, and engineering mentality
  • Willingness to learn, document, and standardise best practices
Job Responsibility
Job Responsibility
  • Analyse, monitor, and optimise ML models, tracking their performance
  • Design and implement CI/CD pipelines for ML models and data flows
  • Containerise and deploy models via APIs, batch processes, and streaming
  • Manage model versioning and traceability
  • Ensure continuous improvement and adaptation of AI use cases and ML models
  • Set up monitoring and alerting for model performance
  • Establish incident response protocols in collaboration with IT
  • Maintain dashboards and automated reports on model health
  • Implement validation frameworks for data and models (e.g., Great Expectations, unit tests, stress tests), in collaboration with Group Governance
  • Contribute to documentation and apply technical best practices
What we offer
What we offer
  • Work in a constantly evolving environment
  • Contribute to digital impact
  • Opportunity for growth and development
  • Fulltime
Read More
Arrow Right

Machine Learning Platform / Backend Engineer

We are seeking a Machine Learning Platform/Backend Engineer to design, build, an...
Location
Location
Serbia; Romania , Belgrade; Timișoara
Salary
Salary:
Not provided
everseen.ai Logo
Everseen
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 4-5+ years of work experience in either ML infrastructure, MLOps, or Platform Engineering
  • Bachelors degree or equivalent focusing on the computer science field is preferred
  • Excellent communication and collaboration skills
  • Expert knowledge of Python
  • Experience with CI/CD tools (e.g., GitLab, Jenkins)
  • Hands-on experience with Kubernetes, Docker, and cloud services
  • Understanding of ML training pipelines, data lifecycle, and model serving concepts
  • Familiarity with workflow orchestration tools (e.g., Airflow, Kubeflow, Ray, Vertex AI, Azure ML)
  • A demonstrated understanding of the ML lifecycle, model versioning, and monitoring
  • Experience with ML frameworks (e.g., TensorFlow, PyTorch)
Job Responsibility
Job Responsibility
  • Design, build, and maintain scalable infrastructure that empowers data scientists and machine learning engineers
  • Own the design and implementation of the internal ML platform, enabling end-to-end workflow orchestration, resource management, and automation using cloud-native technologies (GCP/Azure)
  • Design and manage Kubernetes-based infrastructure for multi-tenant GPU and CPU workloads with strong isolation, quota control, and monitoring
  • Integrate and extend orchestration tools (Airflow, Kubeflow, Ray, Vertex AI, Azure ML or custom schedulers) to automate data processing, training, and deployment pipelines
  • Develop shared services for model behavior/performance tracking, data/datasets versioning, and artifact management (MLflow, DVC, or custom registries)
  • Build out documentation in relation to architecture, policies and operations runbooks
  • Share skills, knowledge, and expertise with members of the data engineering team
  • Foster a culture of collaboration and continuous learning by organizing training sessions, workshops, and knowledge-sharing sessions
  • Collaborate and drive progress with cross-functional teams to design and develop new features and functionalities
  • Ensure that the developed solutions meet project objectives and enhance user experience
  • Fulltime
Read More
Arrow Right

Senior/Architect Data Engineer

We are seeking a highly skilled and experienced Senior/Architect Data Engineer t...
Location
Location
Poland , Warsaw; Poznań; Lublin; Katowice; Rzeszów
Salary
Salary:
Not provided
https://www.inetum.com Logo
Inetum
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Proven experience architecting solutions on the Databricks Lakehouse using Unity Catalog, Delta Lake, MLflow, Model Serving, Feature Store, AutoML, and Databricks Workflows
  • Expertise in real-time/low latency model serving architectures with auto-scaling, confidence-based routing, and A/B testing
  • Strong knowledge of cloud security and governance on Azure or AWS, including Azure AD/AWS IAM, encryption, audit trails, and compliance frameworks
  • Hands-on MLOps skills across experiment tracking, model registry/versioning, drift monitoring, automated retraining, and production rollout strategies
  • Proficiency in Python and Databricks native tooling, with practical integration of REST APIs/SDKs and Databricks SQL in analytics products
  • Familiarity with React dashboards and human-in-the-loop operational workflows for ML and data quality validation
  • Demonstrated ability to optimize performance, reliability, and cost for large-scale analytics/ML platforms with strong observability
  • Experience leading multi-phase implementations with clear success metrics, risk management, documentation, and training/change management
  • Domain knowledge in telemetry, time series, or industrial data (aerospace a plus) and prior work with agentic patterns on Mosaic AI
  • Databricks certifications and experience in enterprise deployments of the platform are preferred
Job Responsibility
Job Responsibility
  • Lead the design and implementation of a Databricks-centric multi-agent processing engine
  • Design governed data ingestion, storage, and real-time processing workflows using Delta Lake, Structured Streaming, and Databricks Workflows
  • Own the model lifecycle with MLflow, including experiment tracking, registry/versioning, A/B testing, drift monitoring, and automated retraining pipelines
  • Architect low latency model serving endpoints with auto-scaling and confidence-based routing for sub-second agent decisioning
  • Establish robust data governance practices with Unity Catalog, including access control, audit trails, data quality, and compliance
  • Drive performance and cost optimization strategies, including auto-scaling, spot usage, and observability dashboards
  • Define production release strategies (blue-green), monitoring and alerting mechanisms, operational runbooks, and Service Level Objectives (SLOs)
  • Partner with engineering, MLOps, and product teams to deliver human-in-the-loop workflows and dashboards
  • Lead change management, training, and knowledge transfer while managing a parallel shadow processing path
  • Plan and coordinate phased delivery, success metrics, and risk mitigation
What we offer
What we offer
  • Flexible working hours
  • Hybrid work model
  • Cafeteria system
  • Generous referral bonuses (up to PLN6,000)
  • Additional revenue sharing opportunities
  • Ongoing guidance from dedicated Team Manager
  • Tailored technical mentoring from assigned technical leader
  • Dedicated team-building budget for online and on-site team events
  • Opportunities to participate in charitable initiatives and local sports programs
  • Supportive and inclusive work culture
  • Fulltime
Read More
Arrow Right