This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
We are looking for an experienced Data Engineer to join our team in Jacksonville, Florida. In this role, you will take the lead in designing and building a cutting-edge Azure lakehouse platform that enables business leaders to access analytics through natural language queries. This position combines hands-on technical expertise with leadership responsibilities, offering an opportunity to mentor a team of skilled engineers while driving innovation.
Job Responsibility:
Architect and develop a robust Azure lakehouse platform, utilizing Azure Data Lake Gen2, Delta Lake, and PySpark to create efficient data pipelines
Implement a semantic layer and metric store to ensure consistent data translation and definitions across the organization
Design and maintain real-time and batch data pipelines, incorporating medallion architecture, schema evolution, and data contracts
Build retrieval systems for large language models (LLMs) using Azure OpenAI and vectorized Delta tables to support chat-based analytics
Ensure data quality, lineage, and observability through tools like Great Expectations and Unity Catalog, while optimizing costs through partitioning and compaction
Develop automated systems for anomaly detection and alerting using Azure ML pipelines and Event Grid
Collaborate with product and operations teams to translate complex business questions into actionable data models and queries
Lead and mentor a team of data and Python engineers, establishing best practices in CI/CD, code reviews, and documentation
Ensure compliance with security, privacy, and governance standards by designing and implementing robust data handling protocols
Requirements:
Minimum of 7-10 years of experience in data engineering
At least 2-4 years leading technical teams
Expertise in Azure Data Lake Gen2, Delta Lake, Unity Catalog, PySpark, and Python
Proven ability to design and implement Lakehouse architectures, including medallion frameworks and schema evolution
Proficiency in Databricks workflows, query optimization, and monitoring tools
Experience with semantic/metrics layers and advanced analytics for business applications
Familiarity with Azure OpenAI integrations, vector search, and structured data grounding
Strong skills in CI/CD development using GitHub Actions or Azure DevOps
Strong skills in infrastructure automation through Bicep or Terraform
Excellent communication skills
Ability to mentor engineers and translate business needs into technical solutions
What we offer:
Medical, vision, dental, and life and disability insurance