This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
We are looking for a skilled Cloud Data Engineer with strong experience in Azure-based data platforms, particularly Azure Data Factory, Databricks and Python. The ideal candidate will design, build, and optimize scalable data pipelines and data solutions to support analytics and business intelligence needs. Experience with Teradata ETL development is a strong plus.
Job Responsibility:
Designing, developing, and maintaining robust data pipelines using Azure Data Factory (ADF)
Build and optimize data processing workflows using Azure Databricks (PySpark/Scala)
Develop scalable data architectures in Azure (Data Lake, Synapse, etc.)
Develop and maintain data solutions using Python for data transformation and automation
Perform data ingestion, transformation, and integration from various sources (structured/unstructured)
Ensure data quality, integrity, and governance across pipelines
Optimize performance and cost of data processing solutions
Collaborate with data analysts, data scientists, and business stakeholders to deliver data solutions
Implement monitoring, logging, and alerting data pipelines
Support CI/CD processes and deployment automation for data workflows
Monitoring and responding to the issues or failures in the Cloud Data Engineering processes for data processing and loading
Data loading and processing occur as either API, Streaming or Batch processing
Be a part of a cross-North America Cloud Data Engineering Delivery team
Requirements:
Must have 8 – 10 years’ experience in Cloud Data Engineering Development & support roles preferably in Microsoft Azure Data Warehouse
Must have Cloud Data Engineering Support experience working with large datasets and complex data environments, processes, and associated solutions
Knowledge and experience with data warehouse star schema and ETL concepts
Proficiency in SQL and data modeling concepts
8-10 years’ experience with Databricks (workflows and jobs orchestrations), Py Spark, Python, Data factory, Delta Optimization techniques, SQL, transformations, Databricks Cluster
Knowledge of Streaming Support with Event Hub, Cosmos DB, Scala, Azure SQL DB and Azure App Insights, key vaults, Azure Functions and logic apps, Azure AAD + key vaults, spark SQL, pandas etc.
Knowledge of automated and scheduled batch job tools such as IBM IWS and Control-M
Knowledge of data governance and unity catalog
Knowledge of performance tuning and optimization techniques
Familiarity with version control tools (e.g., Git)
Knowledge Service now, incident management, ticket handling, change management, on call roaster management, etc.
Must have good oral and written communication skills to effectively communicate with various IT teams and business representatives
Must be able to both collaborate in a team-oriented environment and work independently with direction
Nice to have:
Experience in ETL development using Teradata
Knowledge of Azure Synapse Analytics
Experience with streaming technologies (e.g., Kafka, Event Hub)
Exposure to DevOps practices and tools (Azure DevOps, CI/CD pipelines)
Understanding of data governance and security best practices