This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
Do you love solving real-world data problems with the latest and best techniques? And having fun while solving them in a team! Then come and join our high-energy team of passionate data people. Jash Data Sciences is the right place for you. We are a cutting-edge Data Sciences and Data Engineering startup based in Pune, India. We believe in continuous learning and evolving together. And we let the data speak!
Job Responsibility:
Discovering trends in the data sets and developing algorithms to transform raw data for further analytics
Create Data Pipelines to bring in data from various sources, with different formats, transform it, and finally load it to the target database
Implement ETL/ ELT processes in the cloud using tools like AirFlow, Glue, Stitch, Cloud Data Fusion, and DataFlow
Design and implement Data Lake, Data Warehouse, and Data Marts in AWS, GCP, or Azure using Redshift, BigQuery, PostgreSQL, etc
Creating efficient SQL queries and understanding query execution plans for tuning queries on engines like PostgreSQL
Performance tuning of OLAP/ OLTP databases by creating indices, tables, and views
Write Python scripts for the orchestration of data pipelines
Have thoughtful discussions with customers to understand their data engineering requirements
Break complex requirements into smaller tasks for execution
Requirements:
Strong Python coding skills with basic knowledge of algorithms/data structures and their application
Strong understanding of Data Engineering concepts including ETL, ELT, Data Lake, Data Warehousing, and Data Pipelines
Experience designing and implementing Data Lakes, Data Warehouses, and Data Marts that support terabytes of scale data
A track record of implementing Data Pipelines on public cloud environments (AWS/GCP/Azure) is highly desirable
A clear understanding of Database concepts like indexing, query performance optimization, views, and various types of schemas
Hands-on SQL programming experience with knowledge of windowing functions, subqueries, and various types of joins
Experience working with Big Data technologies like PySpark/ Hadoop
A good team player with the ability to communicate with clarity
Show us your git repo/ blog
1-2 years of experience working on Data Engineering projects for Data Engineer I
2-5 years of experience working on Data Engineering projects for Data Engineer II
1-5 years of Hands-on Python programming experience
Bachelors/Masters' degree in Computer Science is good to have
Courses or Certifications in the area of Data Engineering will be given a higher preference
Candidates who have demonstrated a drive for learning and keeping up to date with technology by continuing to do various courses/self-learning will be given high preference
Nice to have:
Bachelors/Masters' degree in Computer Science is good to have
Courses or Certifications in the area of Data Engineering will be given a higher preference
Candidates who have demonstrated a drive for learning and keeping up to date with technology by continuing to do various courses/self-learning will be given high preference