This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
As a Data Engineer, your duties will be focused on three core areas: Data Platform Infrastructure & DevOps, Data Engineering & Pipeline Development, and Data Modeling & Collaboration.
Job Responsibility:
Data Platform Infrastructure & DevOps: Administer, optimize, and scale our Databricks Lakehouse environment
Develop, maintain, and enhance our data platform infrastructure and security configurations using Terraform
Manage and enforce Unity Catalog for data governance, access control, and metadata management
Implement and manage CI/CD pipelines for data pipelines, dbt projects, and infrastructure deployments using GitHub Actions
Automate operational tasks, monitoring, and alerting for the data platform
Implement and enforce DevSecOps principles
Data Engineering & Pipeline Development: Design and implement data ingestion patterns into Databricks using Delta Lake
Develop, optimize, and troubleshoot complex Spark jobs (PySpark/Scala) for data processing and transformation within Databricks
Manage and extend data ingestion pipelines using Airbyte (or similar modern tools like Fivetran, Stitch)
Orchestrate and automate data pipelines and dbt models using Databricks Workflows
Data Modeling & Collaboration: Collaborate with Analytics Engineers to translate business requirements into efficient and scalable data models using dbt
Implement dbt best practices for modularity, testing, documentation, and version control
Partner effectively with Analytics Engineers, Data Scientists, and business stakeholders
Provide technical guidance and mentorship to junior team members
Requirements:
Bachelor's degree in Computer Science, Data Engineering, or a related technical field required
5+ years of progressive experience as a Data Engineer, with a strong focus on cloud-based data platforms
Deep Databricks Expertise: Extensive experience with Spark (PySpark/Scala), Delta Lake, Unity Catalog, Databricks SQL, and platform administration
Data Modeling: Proven experience with dbt for data modeling, transformation, and testing
Infrastructure as Code (IaC): Strong proficiency with Terraform for defining, provisioning, and managing cloud infrastructure and Databricks resources as code
DevOps & CI/CD: Expertise in Git and GitHub Actions for version control and implementing robust CI/CD pipelines
Programming: Proficiency in SQL and at least one programming language (Python strongly preferred, Scala is a plus)
Data Architecture: Solid understanding of data warehousing, data lake, and lakehouse architectures
Work Requirements: US Citizen, GC Holders or Authorized to Work in the U.S