This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
We are seeking an experienced Lead Databricks Developer with deep expertise in Spark, PySpark, and SQL to design, build, and optimize large-scale data processing solutions. The ideal candidate will have strong hands-on experience in cloud-based data environments (preferably AWS) and will be capable of leading end-to-end data engineering initiatives. This role involves collaborating with cross‑functional teams, ensuring high data quality, and building robust pipelines to support business analytics and data-driven decision-making.
Job Responsibility:
Design, develop, and maintain scalable data processing applications using Python, PySpark, and Apache Spark
Work in cloud-based environments—AWS preferred, with flexibility to work on Azure Databricks
Collaborate closely with data engineers, data scientists, analysts, and business stakeholders to understand requirements and deliver effective solutions
Ensure data integrity, reliability, and performance across data pipelines and workflows
Write clean, well-structured, and maintainable code following industry best practices
Perform exploratory data analysis and implement data validation frameworks
Monitor and troubleshoot performance bottlenecks in Spark jobs and data workflows
Implement and manage CI/CD pipelines for automated build, test, and deployment of data pipelines
Optimize data pipelines for cost, performance, and scalability in cloud environments
Requirements:
Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field
8–12 years of experience in data engineering with a strong focus on Databricks and Spark/PySpark
Strong proficiency in SQL and experience working with relational databases
Advanced hands‑on experience in Python for data engineering
Solid understanding of cloud platforms (AWS preferred
Azure Databricks acceptable)
Experience with Git or other version control systems
Strong analytical thinking, debugging, and problem‑solving skills
Excellent communication and cross‑team collaboration abilities
Nice to have:
Experience with Delta Lake, Unity Catalog, or advanced Databricks features
Exposure to data warehousing concepts and dimensional modeling
Familiarity with Airflow, Glue, or other workflow orchestration tools
Knowledge of data security, governance, and compliance best practices
Understanding of DevOps, containerization (Docker), or Kubernetes
Experience integrating Databricks with BI tools (Power BI, Tableau, etc.)
Knowledge of machine learning workflows or MLOps within Databricks is a plus