CrawlJobs Logo

Middle Data Engineer

leverx.com Logo

LeverX

Location Icon

Location:

Category Icon

Job Type Icon

Contract Type:
Not provided

Salary Icon

Salary:

Not provided

Job Description:

At LeverX, we have had the privilege of delivering over 950 projects. With 20+ years in the market, our team of 2,200 is strong, reliable, and always evolving: learning, growing, and striving for excellence. We are looking for a Middle Data Engineer to join us. Let’s see if we are a good fit for each other!

Job Responsibility:

  • Design, build, and maintain batch/streaming pipelines (ELT/ETL) from diverse sources into DWH/Lakehouse
  • Model data for analytics (star/snowflake, slowly changing dimensions, semantic/metrics layers)
  • Write production-grade SQL and Python
  • optimize queries, file layouts, and partitioning
  • Implement orchestration, monitoring, testing, and CI/CD for data workflows
  • Ensure data quality (validation, reconciliation, observability) and document lineage
  • Collaborate with BI/analytics to deliver trusted, performant datasets and dashboards

Requirements:

  • 3–5 years of experience in data engineering
  • Strong SQL and solid Python for data processing
  • Hands-on experience with at least one cloud and a modern warehouse/lakehouse: Snowflake, Redshift, Databricks, or Apache Spark/Iceberg/Delta
  • Experience delivering on Data Warehouse or Lakehouse projects: star/snowflake modeling, ELT/ETL concepts
  • Familiarity with orchestration (Airflow, Prefect, or similar) and containerization fundamentals (Docker)
  • Understanding of data modeling, performance tuning, cost-aware architecture, and security/RBAC
  • English B1+

Nice to have:

  • Vendor certifications: Snowflake, Databricks, or AWS
  • BI exposure: Tableau, Metabase, Looker/Looker Studio, etc.
  • dbt (models, tests, macros, exposures) for ELT and documentation
  • Git (branching strategies, PR reviews, basic CI) for collaborative delivery
  • Experience with streaming (Kafka/Kinesis), data contracts/metric layers, and data observability tools
What we offer:
  • Projects in different domains: Healthcare, manufacturing, e-commerce, fintech, etc.
  • Projects for every taste: Startup products, enterprise solutions, research & development projects, and projects at the crossroads of SAP and the latest web technologies
  • Global clients based in Europe and the US, including Fortune 500 companies
  • Employment security: We hire for our team, not just a specific project. If your project ends, we will find you a new one
  • Healthy work atmosphere: On average, our employees stay in the company for 4+ years
  • Market-based compensation and regular performance reviews
  • Internal expert communities and courses
  • Perks to support your growth and well-being

Additional Information:

Job Posted:
December 09, 2025

Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for Middle Data Engineer

Senior Software Engineer - Trade Processing Middle Office Platform

As an experienced Staff / Senior Software Engineer, you’ll shape our flagship Mi...
Location
Location
United States , New York
Salary
Salary:
170000.00 - 240000.00 USD / Year
clearstreet.io Logo
Clear Street
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor's Degree in Computer Science or Engineering
  • 10+ years of strong proficiency in Java / Spring Boot, Spring, RDBMS, Service Oriented Architecture (SOA), microservice based server side application development
  • Strong experience with distributed systems, event-driven architecture, and tools like Kafka
  • Practical knowledge of relational databases (e.g., Postgres) and schema design
  • You have contributed to systems that deliver solutions to complex business problems that handle massive amounts of data
  • You prioritize end user experience and it shows in your API designs, functionality, and performance
  • You have a strong command over design patterns, data structures, and algorithms
  • You have strong problem-solving skills with a keen eye for performance optimization
  • You can clearly explain the nuances of system design and paradigms to engineers and stakeholders
  • Strong understanding of multi-threading, concurrency, and performance tuning
Job Responsibility
Job Responsibility
  • Architect and build highly available, horizontally scalable mission critical applications in a modern technology stack
  • Design, build, and optimize core components responsible for processing a high volume of trade data in a low latency environment
  • Solve complex performance and scalability challenges, ensuring our systems handle large-scale financial data efficiently
  • Collaborate with product managers, and other engineers to translate financial methodologies into robust software solutions
  • Lead by example in system design discussions, architectural trade-offs, and best practices
  • Mentor team members, contributing to a strong culture of engineering excellence
What we offer
What we offer
  • Competitive compensation, benefits, and perks
  • Company equity
  • 401k matching
  • Gender neutral parental leave
  • Full medical, dental and vision insurance
  • Lunch stipends
  • Fully stocked kitchens
  • Happy hours
  • Fulltime
Read More
Arrow Right

Middle Palantir Foundry Developer

At LeverX, we have had the privilege of delivering 1,500+ projects. With 20+ yea...
Location
Location
Uzbekistan, Georgia
Salary
Salary:
Not provided
leverx.com Logo
LeverX
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 3+ years in data/analytics engineering or software development
  • Hands-on experience with Palantir Foundry (pipelines and/or applications)
  • Proficiency in Python and SQL
  • Confidence with Git
  • Ability to translate business requirements into working solutions
  • English B1+
Job Responsibility
Job Responsibility
  • Build and maintain data pipelines and transformations in Foundry
  • Implement application logic, views, and access controls
  • Validate data and ensure basic documentation and support
  • Work with stakeholders to clarify requirements and iterate on features
What we offer
What we offer
  • Impactful use-case delivery on real data
  • Possibility to progress into a Team Lead role: mentoring, design facilitation, and coordination
  • Projects in different domains: Healthcare, manufacturing, e-commerce, fintech, etc.
  • Projects for every taste: Startup products, enterprise solutions, research & development projects, and projects at the crossroads of SAP and the latest web technologies
  • Global clients based in Europe and the US, including Fortune 500 companies
  • Employment security: We hire for our team, not just for a specific project. If your project ends, we will find you a new one
  • Healthy work atmosphere: On average, our employees stay in the company for 4+ years
Read More
Arrow Right

Middle Data Engineer

We are seeking a skilled Data Engineer with 3-5 years of experience to join our ...
Location
Location
Ukraine
Salary
Salary:
Not provided
n-ix.com Logo
N-iX
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 2+ years of professional experience in data engineering or backend software engineering with a data focus
  • Strong proficiency in Python for data manipulation and scripting
  • Expert-level SQL skills for complex querying and performance tuning
  • Hands-on production experience with modern cloud data platforms, specifically Snowflake and Databricks
  • Proven experience using dbt in a production environment for transformation layers
  • Experience building and managing complex DAGs in Apache Airflow
  • Cloud platform experience: AWS
  • Working knowledge of Terraform for deploying and managing cloud resources
Job Responsibility
Job Responsibility
  • Design, develop, and maintain reliable ETL/ELT pipelines using Python and SQL to ingest data from various sources into our data lake/warehouse
  • Orchestrate complex data workflows and dependencies using Apache Airflow, ensuring timely data delivery and robust failure handling
  • Champion the use of dbt (data build tool) for developing, testing, and documenting data transformation logic within the warehouse
  • Develop clean, highly optimized SQL models for reporting and analytics (data modeling concepts like Star Schema or Data Vault are a plus)
  • Work hands-on with both Snowflake and Databricks, optimizing compute resources, managing access controls, and ensuring high performance for end-users
  • Utilize Terraform to provision and manage cloud infrastructure (e.g., S3 buckets, IAM roles, Snowflake warehouses) in an Infrastructure-as-Code paradigm
  • Implement data quality checks and monitoring within pipelines to ensure the accuracy and integrity of our data
  • Troubleshoot pipeline failures, identify performance bottlenecks, and implement long-term fixes
What we offer
What we offer
  • Flexible working format - remote, office-based or flexible
  • A competitive salary and good compensation package
  • Personalized career growth
  • Professional development tools (mentorship program, tech talks and trainings, centers of excellence, and more)
  • Active tech communities with regular knowledge sharing
  • Education reimbursement
  • Memorable anniversary presents
  • Corporate events and team buildings
  • Other location-specific benefits
Read More
Arrow Right

Middle Data Engineer

We are seeking a motivated Data Engineer to join our team. In this role, you wil...
Location
Location
Salary
Salary:
Not provided
n-ix.com Logo
N-iX
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 2+ years of experience in batch and streaming ETL using Spark, Python, Scala, Snowflake, or Databricks for Data Engineering or Machine Learning workloads
  • 2+ years orchestrating and implementing pipelines with workflow tools like Databricks Workflows, Apache Airflow, or Luigi
  • 2+ years of experience prepping structured and unstructured data for data science models
  • 2+ years of experience with containerization and orchestration technologies (Docker, Kubernetes discussable) and experience with shell scripting in Bash/Unix shell is preferable
  • Proficiency in Oracle & SQL and data manipulation techniques
  • Experience using machine learning in data pipelines to discover, classify, and clean data
  • Implemented CI/CD with automated testing in Jenkins, Github Actions, or Gitlab CI/CD
  • Familiarity with AWS Services not limited to Lambda, S3, and DynamoDB
  • Demonstrated experience implementing data management life cycle, using data quality functions like standardization, transformation, rationalization, linking, and matching
Job Responsibility
Job Responsibility
  • Developing and maintaining robust data pipelines that drive our business intelligence and analytics
What we offer
What we offer
  • Flexible working format - remote, office-based or flexible
  • A competitive salary and good compensation package
  • Personalized career growth
  • Professional development tools (mentorship program, tech talks and trainings, centers of excellence, and more)
  • Active tech communities with regular knowledge sharing
  • Education reimbursement
  • Memorable anniversary presents
  • Corporate events and team buildings
  • Other location-specific benefits
Read More
Arrow Right

Middle QA Big Data

The objective of this project is to enhance the QA processes through the impleme...
Location
Location
Ukraine
Salary
Salary:
Not provided
n-ix.com Logo
N-iX
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 3+ years of QA experience with a strong focus on Big Data testing, particularly with hands-on experience in Data Lake environments on any cloud platform (preferably Azure)
  • Experience with Azure
  • Hands-on experience in Azure Data Factory, Azure Synapse Analytics, or similar services
  • Proficiency in SQL, capable of writing and optimizing both simple and complex queries for data validation and testing purposes
  • Experienced in PySpark, with experience in data manipulation and transformation, and a demonstrated ability to write and execute test scripts for data processing and validation (ability to understand the code and convert the logic to SQL)
  • Hands-on experience with Functional & System Integration Testing in big data environments, ensuring seamless data flow and accuracy across multiple systems
  • Knowledge and ability to design and execute test cases in a behavior-driven development environment
  • Fluency in Agile methodologies, with active participation in Scrum ceremonies and a strong understanding of Agile principles
  • Familiarity with tools like Jira, including experience with X-Ray for defect management and test case management
  • Proven experience working on high-traffic and large-scale software products, ensuring data quality, reliability, and performance under demanding conditions
Job Responsibility
Job Responsibility
  • Design and execute data validation tests to ensure completeness in Azure Data Lake Storage (ADLS), Azure Synapse, and Databricks
  • Verify data ingestion, transformation, and loading (ETL/ELT) processes in Azure Data Factory (ADF)
  • Validate data schema, constraints, and format consistency across different storage layers
  • Conduct performance testing on data pipelines
  • Optimize query performance by working with data engineers
  • Identify, log, and track defects in JIRA
  • Collaborate with Data Engineers and Business Analysts to resolve data inconsistencies
  • Generate detailed test reports, dashboards, and documentation for stakeholders
What we offer
What we offer
  • Flexible working format - remote, office-based or flexible
  • A competitive salary and good compensation package
  • Personalized career growth
  • Professional development tools (mentorship program, tech talks and trainings, centers of excellence, and more)
  • Active tech communities with regular knowledge sharing
  • Education reimbursement
  • Memorable anniversary presents
  • Corporate events and team buildings
  • Other location-specific benefits
Read More
Arrow Right

Rust Engineer - Platform

As a Platform Backend Engineer (Rust) at Keyrock, you will drive the development...
Location
Location
Salary
Salary:
Not provided
keyrock.com Logo
Keyrock
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field, or equivalent experience
  • Proven experience in building and maintaining data-intensive, large-scale, high-performance trading data platforms
  • Strong expertise in Rust (or C++), Python, and TypeScript for system development and automation in the financial services industry
  • Good understanding of data engineering principles, including data modeling, ETL pipelines, and stream processing
  • Experience with financial services data workflows, including trading, middle office, and back office operations
  • Extensive experience in cloud-native architectures, with proficiency in AWS
  • Proficient in GitOps tools and methodologies for infrastructure automation and deployment
  • Strong background in DevSecFinOps, ensuring compliance, security, and cost efficiency across the development lifecycle
  • Hands-on experience with CI/CD pipelines, infrastructure as code (IaC), and monitoring tools
Job Responsibility
Job Responsibility
  • Rust Development: Design, build, and maintain high-performance backend services and APIs using Rust, ensuring low latency and high availability for critical trading data platforms
  • Strong systems engineering fundamentals: Concurrency, memory management, networking, serialization, and observability Solid understanding of performance tuning and profiling in real-world systems
  • System Integration: Create seamless integrations between live trading operations (exchanges/DeFi) and backoffice systems, automating workflows to improve operational efficiency
  • Cloud-Native Deployment: Deploy and manage services in a cloud-native environment, leveraging AWS, Kubernetes, and Terraform to scale infrastructure infrastructure-as-code
  • DevOps & Observability: Maintain GitOps-driven workflows, ensuring robust CI/CD pipelines and implementing deep system observability (logging, metrics, tracing) for rapid incident response
  • Database Optimization: Optimize data storage and retrieval strategies (SQL/NoSQL), balancing query performance, cost efficiency, and data integrity in a high-volume financial environment
  • Security & Compliance: Engineer solutions with a "Security-First" mindset, ensuring strict adherence to compliance standards and secure handling of sensitive financial data
  • Cross-Functional Collaboration: Partner with Product Managers, Risk teams, and other engineers to translate complex business requirements into reliable technical specifications and features
  • Technical Excellence: Actively participate in code reviews, contribute to architectural discussions, and mentor fellow engineers to foster a culture of high code quality and innovation
  • Continuous Improvement: Stay updated on emerging trends in the Rust ecosystem, cloud infrastructure, and blockchain technologies to continuously refine the platform’s capabilities
What we offer
What we offer
  • A competitive salary package
  • Autonomy in your time management thanks to flexible working hours and the opportunity to work remotely
  • The freedom to create your own entrepreneurial experience by being part of a team of people in search of excellence
  • Fulltime
Read More
Arrow Right

Middle Data Engineer

Vention is a global engineering partner to tech leaders and fast-growing startup...
Location
Location
Georgia , Batumi
Salary
Salary:
Not provided
ventionteams.com Logo
Vention
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Hands-on experience with Apache Hop, including designing, orchestrating, and monitoring ETL/ELT pipelines, managing transformations and workflows, and integrating with relational and cloud data sources
  • Strong experience with Microsoft SQL Server and SSRS, including writing and optimizing complex T-SQL queries and stored procedures, designing relational schemas, and developing, deploying, and maintaining SSRS reports (tabular and paginated) for business stakeholders
  • Technical requirements: Python, SQL Spark, PySpark, Pandas AWS: Glue, S3, DMS, Lambda, Athena, RDS Kubernetes, Helm Terraform, Terraform Cloud
Job Responsibility
Job Responsibility
  • Design and implement automated pipelines to collect data from diverse sources (APIs, RDBMS, Cloud) into a centralized Data Lake/Warehouse
  • Develop logic to transform and map heterogeneous data into a unified, consistent style and schema
  • Build, monitor, and optimize end-to-end data workflows using Apache Hop, AWS Glue, and Lambda
  • Ensure cost-effective data processing by optimizing cloud resource consumption and minimizing LLM token usage
  • Maintain and optimize complex T-SQL queries, schemas, and SSRS reports within MS SQL Server environments
What we offer
What we offer
  • EDU corporate community (300+ members): tech communities, interest clubs, events, a small R&D lab, a knowledge base, and a dedicated AI track
  • Licenses for AI tools: GitHub Copilot, Cursor, and others
  • 24 working days of vacation per year
  • Expanded medical insurance
  • Corporate getaways & team building activities
  • Fitpass sport program
  • Support for significant life events
  • Access to discounts across a variety of stores, restaurants & cafes through a corporate discount program
  • Referral bonuses for bringing in new talent
  • Fulltime
Read More
Arrow Right

Middle Data Engineer

Vention is a global engineering partner to tech leaders and fast-growing startup...
Location
Location
Georgia , Tbilisi
Salary
Salary:
Not provided
ventionteams.com Logo
Vention
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Hands-on experience with Apache Hop, including designing, orchestrating, and monitoring ETL/ELT pipelines, managing transformations and workflows, and integrating with relational and cloud data sources
  • Strong experience with Microsoft SQL Server and SSRS, including writing and optimizing complex T-SQL queries and stored procedures, designing relational schemas, and developing, deploying, and maintaining SSRS reports (tabular and paginated) for business stakeholders
  • Technical requirements: Python, SQL Spark, PySpark, Pandas AWS: Glue, S3, DMS, Lambda, Athena, RDS Kubernetes, Helm Terraform, Terraform Cloud
Job Responsibility
Job Responsibility
  • Design and implement automated pipelines to collect data from diverse sources (APIs, RDBMS, Cloud) into a centralized Data Lake/Warehouse
  • Develop logic to transform and map heterogeneous data into a unified, consistent style and schema
  • Build, monitor, and optimize end-to-end data workflows using Apache Hop, AWS Glue, and Lambda
  • Ensure cost-effective data processing by optimizing cloud resource consumption and minimizing LLM token usage
  • Maintain and optimize complex T-SQL queries, schemas, and SSRS reports within MS SQL Server environments
What we offer
What we offer
  • EDU corporate community (300+ members): tech communities, interest clubs, events, a small R&D lab, a knowledge base, and a dedicated AI track
  • Licenses for AI tools: GitHub Copilot, Cursor, and others
  • 24 working days of vacation per year
  • Expanded medical insurance
  • Corporate getaways & team building activities
  • Fitpass sport program
  • Support for significant life events
  • Access to discounts across a variety of stores, restaurants & cafes through a corporate discount program
  • Referral bonuses for bringing in new talent
  • Fulltime
Read More
Arrow Right