This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
The Data Engineer role requires expertise in Python, Kafka, and Snowflake, focusing on data pipeline development and optimization. Candidates should have a Bachelor’s or Master’s degree in a quantitative field and 5-7 years of experience in data engineering. Strong SQL skills and familiarity with CI/CD practices are essential. Join our innovative team to enhance data processes.
Job Responsibility:
data pipeline development and optimization
data engineering
Requirements:
Bachelor’s or Master’s degree in Computer Science, Applied Mathematics, Engineering, or a related quantitative field
5–7 years of professional hands-on coding experience in collaborative, team-based environments
strong troubleshooting skills in SQL and scripting
Proficiency in Python or Java
deep familiarity with SDLC, CI/CD best practices, and Kubernetes deployment
expertise in temporal data modeling (e.g., SCD Type 2)
schema management with a focus on schema evolution (Iceberg Apache)
performance optimization through data partitioning and clustering
architectural theory involving normalization/denormalization and natural vs. surrogate keys