CrawlJobs Logo

Cloud Big-data Engineer

phasorsoft.com Logo

PhasorSoft Group

Location Icon

Location:
United States , Starkville

Category Icon

Job Type Icon

Contract Type:
Not provided

Salary Icon

Salary:

45.00 USD / Hour

Job Description:

An expert with 4-5 years of experience in Hadoop ecosystem and cloud- (AWS ecosystem/Azure), relational data stores, Data Integration techniques, XML, Python, Spark, and ETL techniques.

Requirements:

  • 4-5 years of experience in Hadoop ecosystem and cloud (AWS ecosystem/Azure)
  • Experience working with in-memory computing using R, Python, Spark, PySpark, Kafka, and Scala
  • Experience in parsing and shredding XML and JSON, shell scripting, and SQL
  • Experience working with Hadoop ecosystem - HDFS, Hive
  • Experience working with AWS ecosystem - S3, EMR, EC2, Lambda Cloud Formation, Cloud Watch, SNS/SQS
  • Experience with Azure – Azure Data Factory (ADF)
  • Experience working with SQL and No SQL databases
  • Experience designing and developing data sourcing routines utilizing typical data quality functions involving standardization, transformation, rationalization, linking, and matching
  • Work Authorization: H1, GC, US Citizen

Additional Information:

Job Posted:
December 11, 2025

Work Type:
On-site work
Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for Cloud Big-data Engineer

Data Engineer

Are you a Data Engineer based in Austin, Texas, who is inspired by working with ...
Location
Location
United States , Austin
Salary
Salary:
100000.00 - 135000.00 USD / Year
beezwax.net Logo
Beezwax Datatools
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 4-6 years of hands-on data modeling and data engineering experience
  • Strong expertise in dimensional modeling and data warehousing
  • Database design and development experience with relational or MPP databases such as Postgres/ Oracle/ Teradata/ Vertica
  • Experience in design and development of custom ETL pipelines using SQL and scripting languages (Python/ Shell/ Golang)
  • Proficiency in advanced SQL and performance tuning
  • Hands on experience with Big-Data platforms like Spark, Dremio, Hadoop, MapReduce, Hive etc
  • Experience with Java, Scala and Python
  • Experience with cloud computing platforms like AWS, Google Cloud
  • Experience working with APIs
  • Ability to learn and adapt to new tools and technologies
What we offer
What we offer
  • Competitive compensation
  • Retirement plan with employer matching
  • Excellent healthcare package with vision and dental
  • Support for productivity and continued learning in the forms of hardware, software, learning materials, training, and conferences
  • Fulltime
Read More
Arrow Right

Senior Data Engineer

Senior Data Engineer – Dublin (Hybrid) Contract Role | 3 Days Onsite. We are see...
Location
Location
Ireland , Dublin
Salary
Salary:
Not provided
solasit.ie Logo
Solas IT Recruitment
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 7+ years of experience as a Data Engineer working with distributed data systems
  • 4+ years of deep Snowflake experience, including performance tuning, SQL optimization, and data modelling
  • Strong hands-on experience with the Hadoop ecosystem: HDFS, Hive, Impala, Spark (PySpark preferred)
  • Oozie, Airflow, or similar orchestration tools
  • Proven expertise with PySpark, Spark SQL, and large-scale data processing patterns
  • Experience with Databricks and Delta Lake (or equivalent big-data platforms)
  • Strong programming background in Python, Scala, or Java
  • Experience with cloud services (AWS preferred): S3, Glue, EMR, Redshift, Lambda, Athena, etc.
Job Responsibility
Job Responsibility
  • Build, enhance, and maintain large-scale ETL/ELT pipelines using Hadoop ecosystem tools including HDFS, Hive, Impala, and Oozie/Airflow
  • Develop distributed data processing solutions with PySpark, Spark SQL, Scala, or Python to support complex data transformations
  • Implement scalable and secure data ingestion frameworks to support both batch and streaming workloads
  • Work hands-on with Snowflake to design performant data models, optimize queries, and establish solid data governance practices
  • Collaborate on the migration and modernization of current big-data workloads to cloud-native platforms and Databricks
  • Tune Hadoop, Spark, and Snowflake systems for performance, storage efficiency, and reliability
  • Apply best practices in data modelling, partitioning strategies, and job orchestration for large datasets
  • Integrate metadata management, lineage tracking, and governance standards across the platform
  • Build automated validation frameworks to ensure accuracy, completeness, and reliability of data pipelines
  • Develop unit, integration, and end-to-end testing for ETL workflows using Python, Spark, and dbt testing where applicable
Read More
Arrow Right

Applications Development Senior Group Manager

This role will be part of the Risk Data team and is a senior management level po...
Location
Location
United Kingdom , London
Salary
Salary:
Not provided
https://www.citi.com/ Logo
Citi
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Strong academic record, ideally with a Bachelor’s or Master’s degree in Computer Science or engineering or related technical discipline
  • Proven experience in enterprise application development with full stack technologies
  • Strong Architect and hands on technical experience in implementing large volume real time complex solutions in Big Data Platforms & Public Cloud platforms
  • Experience in Data architecture, strong Software development fundamentals, data structures, design patterns, object-oriented principles
  • Experience in design and delivery of multi-tiered applications and high performance server side components
  • Skills on system performance tuning, high performance, low latency, multithreading and experience with Java server side programming
  • Preferred experience in Handling high volumes of data and working with In-memory databases and Caching solutions
  • Experience of building and leading teams, ideally with a global resource profile and demonstrated ability to deliver large projects efficiently and on time
  • Significant experience in large Financial Services Technology services companies is expected for this position
  • Hands-on development, architecture and leadership experience in real-time data engineering platforms implementation
Job Responsibility
Job Responsibility
  • Lead the efforts in Institutional Data Platform (ICG) that span multiple businesses, products and functions
  • Delivery of Price Risk related Data initiatives and Capital reporting (GSIB) related deliverables
  • Establish strong relationships with the global business stakeholders and ensure transparency of project deliveries
  • Actively identify and manage risks and issues, working with disparate teams to create mitigation plans and follow-through to resolution
  • Adhere to all key Project Management (PMQC) & Engineering Excellence standards
  • Ensure timely communications to Senior Technology Management and Business Partners in Front Office, Middle Office & other Operations functions
  • Drive the design and development of system architecture, work with end-users of the systems, and enhance the quality of deliverables
  • Ensure staff follows Citi documented policy and procedures as well as maintain desktop procedures and supporting documentation for filings on a current basis and in comprehensive manner
  • Ensure change is managed with appropriate controls, documentation, and approvals including implementation of new and revised regulatory reporting requirements
  • Manage and maintain all disaster recovery plans, oversee appropriate testing, and provide permit-to-operate for new applications
What we offer
What we offer
  • 27 days annual leave (plus bank holidays)
  • A discretional annual performance related bonus
  • Private Medical Care & Life Insurance
  • Employee Assistance Program
  • Pension Plan
  • Paid Parental Leave
  • Special discounts for employees, family, and friends
  • Access to an array of learning and development resources
  • Fulltime
Read More
Arrow Right

Principal Software Engineer

Principal Software Engineer role at Hewlett Packard Enterprise to design, develo...
Location
Location
United States , San Jose
Salary
Salary:
148000.00 - 340500.00 USD / Year
https://www.hpe.com/ Logo
Hewlett Packard Enterprise
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor or Masters degree in Computer science, Computer Engineering or a related field
  • 10+ years of experience in software engineering with a focus on Python, Go or Java
  • Strong understanding of RESTful API design and development
  • 2+ years of Experience working with large scale distributed systems based on either cloud technologies or Kubernetes
  • 2+ years of experience on event-driven technologies like Kafka and Apache Storm/Flink
  • 2+ years of experience in Big-data technologies like Apache spark/Databricks
  • Proficient in working with Redis and databases like Cassandra/Datastax
  • Must hold U.S. citizenship
Job Responsibility
Job Responsibility
  • Design, develop, and test software related to the cloud-based network configuration and reporting system
  • Solve complex problems and designing subsystems for Mist platform
  • Develop software for highly scalable and fault-tolerant cloud-scale distributed applications
  • Develop microservices using Python, and/or Go (golang)
  • Develop event-driven systems using Python and Java
  • Develop software for AIDE's real-time data pipeline and batch processing
  • Develop ETL pipelines aiding in training and inference of various ML models using big-data frameworks like Apache Spark
  • Build metrics, monitoring and structured logging into the product
  • Write unit, integration and functional tests
  • Participate in collaborative, DevOps style, lean practices
What we offer
What we offer
  • Health & Wellbeing benefits
  • Personal & Professional Development programs
  • Unconditional Inclusion environment
  • Comprehensive benefits suite supporting physical, financial and emotional wellbeing
  • Fulltime
Read More
Arrow Right
New

Data Engineer 3

Our analysts transform data into meaningful insights that drive strategic decisi...
Location
Location
India , Chennai
Salary
Salary:
Not provided
comcastadvertising.com Logo
Comcast Advertising
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor's degree in engineering, Computer Science, Information Systems, or related field with 5-8 years of relevant experience
  • 2+ years of hands-on experience in design, build, test, orchestrate, deploy large scale Scala Spark ETL Jobs
  • 2+ years of experience with AWS technologies such as S3, Athena, Lambdas, EMR or EMR Serverless or similar
  • 1+ years of Experience in Data Modeling, Data architecture, Data Quality, Metadata, ETL or Data Warehouse methodologies and technologies
  • Experience in Open Table Format such as Iceberg or similar, Datawarehouse such as Snowflake or similar, MySQL or Postgres or similar is a plus
  • Strong problem-solving skills
  • General understanding of SDLC and Agile Methodologies
  • General understanding of Test-Driven Development and CI/CD
  • General understanding of cloud data engineering design patterns, data storage principles, and related use cases
Job Responsibility
Job Responsibility
  • Responsible for designing, building and overseeing the deployment and operation of technology architecture, solutions and software to capture, manage, store and utilize structured and unstructured data from internal and external sources
  • Establishes and builds processes and structures based on business and technical requirements to channel data from multiple inputs, route appropriately and store using any combination of distributed (cloud) structures, local databases, and other applicable storage forms as required
  • Develops technical tools and programming that leverage artificial intelligence, machine learning and big-data techniques to cleanse, organize and transform data and to maintain, defend and update data structures and integrity on an automated basis
  • Creates and establishes design standards and assurance processes for software, systems and applications development to ensure compatibility and operability of data connections, flows and storage requirements
  • Reviews internal and external business and product requirements for data operations and activity and suggests changes and upgrades to systems and storage to accommodate ongoing needs
  • Work with data modelers/analysts to understand the business problems they are trying to solve then create or augment data assets to feed their analysis
  • Contributor to the overall Data Product roadmap by working closely with our business partners to understand their challenges and develop analytical tools to help drive business decisions
  • Must understand the basics of relational data modeling and be able to clearly articulate the reasons to use non-relational systems in our architecture
  • Interface with other technology teams to leverage cloud-native data engineering principles to deliver high quality data faster
  • Foster a DevOps culture and participate in development sprints, retrospectives, releases, and deployments
What we offer
What we offer
  • Paid Time off
  • Physical Wellbeing benefits
  • Financial Wellbeing benefits
  • Emotional Wellbeing benefits
  • Life Events + Family Support Benefits
  • Fulltime
Read More
Arrow Right

Senior C++ Developer

VarSome.com is the world’s leading website for professional human genetics. VarS...
Location
Location
Greece , Athens
Salary
Salary:
Not provided
saphetor.com Logo
Saphetor
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • At least 4 years of commercial C++ development experience, using templates, STL containers, smart pointers, memory management and multithreading
  • BSc and/or MSc degree in Computer Science, Engineering or Mathematics
  • Hands-on experience with relational databases, optimizing queries and table schemas for optimal performance
  • Good experience with distributed cloud-based computing and big-data systems
  • Familiarity with agile methodologies and the ability to adapt to a fast-paced development environment, actively taking in code reviews, scrums, technical discussions etc
  • Very good understanding of fundamental application design principles and object-oriented design, in order to build & maintain a large high-quality code base
  • Excellent written & oral communication skills in English
Job Responsibility
Job Responsibility
  • Develop our core C++ applications and library for importing and analyzing genetic data
  • Write reusable, testable, and efficient code, including unit & regression tests
  • Take complete ownership of projects (ranging from a few days to a month) to deliver a working end-to-end implementation, including unit tests & testing
  • Optimize & architect our platform for maximum speed, high availability and scalability
  • Maintain & improve our internal high-performance clinical annotation tools and the custom databases built, optimized for genetics
  • Contribute to the documentation of software architecture, design and implementation details
What we offer
What we offer
  • A position in a fascinating healthcare growth domain, at the cutting edge of technology and research
  • A competitive compensation package combined with additional benefits
  • Endless learning opportunities, while transferring new technologies from academics to clinical practice all over the world
  • Fulltime
Read More
Arrow Right

Data Engineer 3

Responsible for designing, building and overseeing the deployment and operation ...
Location
Location
India , Chennai
Salary
Salary:
Not provided
comcastcorporation.com Logo
Comcast
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor's Degree
  • 5-7 Years Relevant Work Experience
  • In-depth experience, knowledge and skills in own discipline
  • Experience in designing, building and overseeing deployment and operation of technology architecture, solutions and software for data
  • Experience in developing data structures and pipelines
  • Experience with data acquisition, archive recovery, and database implementation
  • Experience with on-prem platforms like Kubernetes and Teradata
  • Experience with Cloud platforms like Databricks, AWS S3, Redshift
  • Understanding of data lineage and transformation rules
  • Understanding of data sensitivity and customer data privacy rules and regulations
Job Responsibility
Job Responsibility
  • Design, build and oversee deployment and operation of technology architecture, solutions and software to capture, manage, store and utilize structured and unstructured data
  • Establish and build processes and structures to channel data from multiple inputs
  • Develop technical tools and programming that leverage AI, machine learning and big-data techniques to cleanse, organize and transform data
  • Create and establish design standards and assurance processes for software, systems and applications development
  • Review internal and external business and product requirements for data operations
  • Work with data modelers/analysts to understand business problems and create or augment data assets
  • Develop data structures and pipelines to organize, collect, standardize and transform data
  • Ensure data quality during ingest, processing and final load
  • Create standard ingestion frameworks for structured and unstructured data
  • Create standard methods for end users to consume data (database views, extracts, APIs)
What we offer
What we offer
  • Paid Time off
  • Physical Wellbeing benefits
  • Financial Wellbeing benefits
  • Emotional Wellbeing benefits
  • Life Events + Family Support benefits
  • Fulltime
Read More
Arrow Right

Principal Software Engineer

Principal Software Engineer role at Hewlett Packard Enterprise to design, develo...
Location
Location
United States , San Jose
Salary
Salary:
148000.00 - 340500.00 USD / Year
https://www.hpe.com/ Logo
Hewlett Packard Enterprise
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor or Masters degree in Computer science, Computer Engineering or a related field
  • 10+ years of experience in software engineering with a focus on Python, Go or Java
  • Strong understanding of RESTful API design and development
  • 2+ years of Experience working with large scale distributed systems based on either cloud technologies or Kubernetes
  • 2+ years of experience on event-driven technologies like Kafka and Apache Storm/Flink
  • 2+ years of experience in Big-data technologies like Apache spark/Databricks
  • Proficient in working with Redis and databases like Cassandra/Datastax
  • Excellent problem-solving and analytical skills
  • Strong communication and collaboration skills
Job Responsibility
Job Responsibility
  • Design, develop, and test software related to the cloud-based network configuration and reporting system
  • Solve complex problems and design subsystems for the Mist platform
  • Develop software for highly scalable and fault-tolerant cloud-scale distributed applications
  • Develop microservices using Python, and/or Go (golang)
  • Develop event-driven systems using Python and Java
  • Develop software for AIDE's real-time data pipeline and batch processing
  • Develop ETL pipelines aiding in training and inference of various ML models using big-data frameworks like Apache Spark
  • Build metrics, monitoring and structured logging into the product
  • Write unit, integration and functional tests
  • Participate in collaborative, DevOps style, lean practices
What we offer
What we offer
  • Health & Wellbeing benefits
  • Personal & Professional Development programs
  • Unconditional Inclusion environment
  • Comprehensive suite of benefits supporting physical, financial and emotional wellbeing
  • Fulltime
Read More
Arrow Right