This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
We are looking for a talented Senior Data Engineer to build and scale the data infrastructure that powers Torus's mission. As a core member of our data team, you'll build and maintain our modern data stack (dbt, Redshift, Airflow, Fivetran, Metabase, and Streamlit), designing and implementing scalable data pipelines that ingest, transform, and serve data from our complex ecosystem of IoT devices, grid systems, and business applications.
Job Responsibility:
Design, build, and maintain scalable batch and streaming data pipelines that handle high-volume IoT telemetry and business data
Develop robust ELT workflows to ingest, transform, and load data from diverse sources including APIs, databases, IoT devices, and third-party systems
Build and optimize our data warehouse using Redshift, implementing dimensional models that support analytics and machine learning use cases
Implement real-time data processing systems that enable immediate insights and rapid response to system events
Develop incremental SQL patterns in dbt for efficient data transformation
Build tools, processes, and pipelines to enforce, check, and manage data quality at scale
Develop monitoring and alerting systems to ensure pipeline reliability and data freshness
Create data validation frameworks and automated testing for data pipelines
Establish best practices for data governance, documentation, and lineage tracking
Build frameworks that enable data scientists to deploy models to production efficiently
Develop self-service analytics capabilities and data access patterns for non-technical stakeholders
Create and enhance analytics tools to facilitate intuitive data consumption
Own the full software development lifecycle for data services, focusing on automation, testing, monitoring, and documentation
Develop and maintain infrastructure using AWS CDK and Terraform
Build and maintain CI/CD pipelines for data operations
Manage cloud infrastructure on AWS (ECS, Redshift, Lambda, S3)
Support ad hoc data requests and maintain core pipeline operations
Requirements:
Typically requires a bachelor's degree in Computer Science, Engineering, Information Technology, Data Science, or a related technical field and 5+ years of experience building scalable data pipelines
Strong experience building batch and streaming data pipelines using distributed processing frameworks
3+ years of experience designing and implementing ELT pipelines for data extraction, transformation, and loading from diverse sources
Expert proficiency in Python with strong software engineering fundamentals
Advanced SQL skills and experience with relational databases and data warehousing
Hands-on experience with data warehouse modeling, including dimensional modeling and schema design
Experience with cloud platforms (AWS preferred) and infrastructure-as-code tools
Practical experience owning production data systems with DevOps fundamentals
Experience with containerization (Docker) and orchestration concepts
Passion for data quality, monitoring, and building reliable systems
Applicants must already have the legal authorization to work in the US without requiring any employer sponsorship
Proficiency in reading, writing, and speaking English required
Nice to have:
Experience with modern data stack tools: dbt, Redshift, Airflow, Fivetran, Streamlit, or similar