CrawlJobs Logo

Python PySpark Engineer

spyro-soft.com Logo

Spyrosoft

Location Icon

Location:
Poland , Wroclaw

Category Icon

Job Type Icon

Contract Type:
Not provided

Salary Icon

Salary:

100.00 - 160.00 PLN / Hour

Job Responsibility:

You will play a key role in migrating Building ETL/ELT processes in the Client’s Palantir Foundry infrastructure under the Data Sphere Program, establishing Foundry as the primary Data Lake platform for the Healthcare Commercial

Requirements:

  • Excellent knowledge of PySpark / Python
  • Great knowledge of ETL/ELT processes
  • Experience with working with data lake systems (preferably Palantir Foundry) for data ingestions
  • Practice with creating documentation on the Confluence platform
  • Ability to use ticketing systems such as JIRA and/or Azure DevOps
  • Familiarity with Snowflake infrastructure as an advance
  • Ability to work in an agile BI team (DevOps) and to share skills and experience
  • Fluency in English

Nice to have:

Data Lake -Palantir Foundry

Additional Information:

Job Posted:
January 29, 2026

Employment Type:
Fulltime
Work Type:
Remote work
Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for Python PySpark Engineer

Pyspark Data Engineer

The Data Analytics Intmd Analyst is a developing professional role. Deals with m...
Location
Location
India , Chennai
Salary
Salary:
Not provided
https://www.citi.com/ Logo
Citi
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 4-8 years relevant experience in Data Analytics and Big Data
  • SQL, Python, Pyspark, with Spark components
  • Minimum 4 years of experience as a python developer with expertise in automation testing to design, develop, and automate robust software solutions and testing frameworks like Pytest, Behave
  • 2-4 years of experience as Big Data Engineer to develop, optimize, and manage large-scale data processing systems and analytics platforms
  • 4 years of experience in distributed data processing & near real-time data analytics using PySpark
  • Strong understanding of PySpark execution plans, partitioning & optimization techniques
Job Responsibility
Job Responsibility
  • Integrates in-depth data analysis knowledge with a solid understanding of industry standards and practices
  • Demonstrates a Good understanding of how data analytics teams and area integrate with others in accomplishing objectives
  • Applies project management skills
  • Applies analytical thinking and knowledge of data analysis tools and methodologies
  • Analyzes factual information to make accurate judgments and recommendations focused on local operations and broader impacts
  • Applies professional judgment when interpreting data and results breaking down information in a systematic and communicable manner
  • Employs developed communication and diplomacy skills to exchange potentially complex/sensitive information
  • Demonstrates attention to quality and timeliness of service to ensure the effectiveness of the team and group
  • Provides informal guidance or on-the-job-training to new team members
  • Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency
  • Fulltime
Read More
Arrow Right

Gen AI/Data Engineer

Job Title: Gen AI/Data Engineer. Duration: 12 Months. Location: Addison TX. Pay ...
Location
Location
United States , Addison
Salary
Salary:
65.00 - 68.26 USD / Hour
apexsystems.com Logo
Apex Systems
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Expert Python development skills
  • Python, PySpark, Scala - heavy data engineer skills
  • RAG LLM, (Chat GPT – Copilot) –need a someone who has experience developing these technologies. NOT a User
  • Hands-on experience in application development in one or more areas MongoDB, Redis, Angular/React Frameworks, Containerization.
  • Experience with AI/ML/GenAI Lifecycle Management and Development and its Ecosystem. Hands on experience building frameworks using MLFlow, KubeFlow, Fine – Tuning techniques, Inference Frameworks
  • Building API based application leveraging FAST API services, JWT Integration, API Gateway
  • Working in large sized teams that collaboratively develop on a shared multi-repo codebase using IDEs (e.g. VS Code rather than Jupyter Notebooks), Continuous Integration (CI), Continuous Deployment (CD) and Continuous Testing
  • Hands-on DevOps experience with one or more of the following enterprise development tools: Version Control (GIT/Bitbucket), Build Orchestration (Jenkins), Code Quality (SonarQube and pytest Unit Testing), Artifact Management (Artifactory) and Deployment (Ansible)
Job Responsibility
Job Responsibility
  • Codes solutions and unit test to deliver a requirement/story per the defined acceptance criteria and compliance requirements.
  • Utilizes multiple architectural components (across data, application, business) in design and development of client requirements.
  • Performs Continuous Integration and Continuous Development (CI-CD) activities.
  • Contributes to story refinement and definition of requirements.
  • Participates in estimating work necessary to realize a story/requirement through the delivery lifecycle.
What we offer
What we offer
  • medical, dental, vision, life, disability, and other insurance plans
  • ESPP (employee stock purchase program)
  • 401K program with a company match after 12 months of tenure
  • HSA (Health Savings Account on the HDHP plan)
  • SupportLinc Employee Assistance Program (EAP) with up to 8 free counseling sessions
  • corporate discount savings program
  • on-demand training program
  • access to certification prep and a library of technical and leadership courses/books/seminars after 6+ months of tenure
  • certification discounts
  • dedicated customer service team
  • Fulltime
Read More
Arrow Right

Lead Data Engineer

We are seeking an experienced Senior Data Engineer to lead the development of a ...
Location
Location
India , Kochi; Trivandrum
Salary
Salary:
Not provided
experionglobal.com Logo
Experion Technologies
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 5+ years experience in data engineering with analytical platform development focus
  • Proficiency in Python and/or PySpark
  • Strong SQL skills for ETL processes and large-scale data manipulation
  • Extensive AWS experience (Glue, Lambda, Step Functions, S3)
  • Familiarity with big data systems (AWS EMR, Apache Spark, Apache Iceberg)
  • Database experience with DynamoDB, Aurora, Postgres, or Redshift
  • Proven experience designing and implementing RESTful APIs
  • Hands-on CI/CD pipeline experience (preferably GitLab)
  • Agile development methodology experience
  • Strong problem-solving abilities and attention to detail
Job Responsibility
Job Responsibility
  • Architect, develop, and maintain end-to-end data ingestion framework for extracting, transforming, and loading data from diverse sources
  • Use AWS services (Glue, Lambda, EMR, ECS, EC2, Step Functions) to build scalable, resilient automated data pipelines
  • Develop and implement automated data quality checks, validation routines, and error-handling mechanisms
  • Establish comprehensive monitoring, logging, and alerting systems for data quality issues
  • Architect and develop secure, high-performance APIs for data services integration
  • Create thorough API documentation and establish standards for security, versioning, and performance
  • Work with business stakeholders, data scientists, and operations teams to understand requirements
  • Participate in sprint planning, code reviews, and agile ceremonies
  • Contribute to CI/CD pipeline development using GitLab
Read More
Arrow Right

Senior Data Engineer

At Rearc, we're committed to empowering engineers to build awesome products and ...
Location
Location
India , Bangalore
Salary
Salary:
Not provided
rearc.io Logo
Rearc
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 8+ years of experience in data engineering, showcasing expertise in diverse architectures, technology stacks, and use cases
  • Strong expertise in designing and implementing data warehouse and data lake architectures, particularly in AWS environments
  • Extensive experience with Python for data engineering tasks, including familiarity with libraries and frameworks commonly used in Python-based data engineering workflows
  • Proven experience with data pipeline orchestration using platforms such as Airflow, Databricks, DBT or AWS Glue
  • Hands-on experience with data analysis tools and libraries like Pyspark, NumPy, Pandas, or Dask
  • Proficiency with Spark and Databricks is highly desirable
  • Experience with SQL and NoSQL databases, including PostgreSQL, Amazon Redshift, Delta Lake, Iceberg and DynamoDB
  • In-depth knowledge of data architecture principles and best practices, especially in cloud environments
  • Proven experience with AWS services, including expertise in using AWS CLI, SDK, and Infrastructure as Code (IaC) tools such as Terraform, CloudFormation, or AWS CDK
  • Exceptional communication skills, capable of clearly articulating complex technical concepts to both technical and non-technical stakeholders
Job Responsibility
Job Responsibility
  • Strategic Data Engineering Leadership: Provide strategic vision and technical leadership in data engineering, guiding the development and execution of advanced data strategies that align with business objectives
  • Architect Data Solutions: Design and architect complex data pipelines and scalable architectures, leveraging advanced tools and frameworks (e.g., Apache Kafka, Kubernetes) to ensure optimal performance and reliability
  • Drive Innovation: Lead the exploration and adoption of new technologies and methodologies in data engineering, driving innovation and continuous improvement across data processes
  • Technical Expertise: Apply deep expertise in ETL processes, data modelling, and data warehousing to optimize data workflows and ensure data integrity and quality
  • Collaboration and Mentorship: Collaborate closely with cross-functional teams to understand requirements and deliver impactful data solutions—mentor and coach junior team members, fostering their growth and development in data engineering practices
  • Thought Leadership: Contribute to thought leadership in the data engineering domain through technical articles, conference presentations, and participation in industry forums
Read More
Arrow Right

Data Engineering Architect

Data engineering involves the development of solutions for the collection, trans...
Location
Location
India
Salary
Salary:
Not provided
lingarogroup.com Logo
Lingaro
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 10+ years’ experience in the Data & Analytics area
  • 4+ years’ experience into Data Engineering Architecture
  • Proficiency in Python, PySpark, SQL
  • Strong expertise in Azure cloud services such as: ADF, databricks, pyspark, Logic app
  • Strong understanding of data engineering concepts, including data modeling, ETL processes, data pipelines, and data governance
  • Expertise in designing and implementing scalable and efficient data processing frameworks
  • In-depth knowledge of various data technologies and tools, such as relational databases, NoSQL databases, data lakes, data warehouses, and big data frameworks (e.g., Hadoop, Spark)
  • Experience in selecting and integrating appropriate technologies to meet business requirements and long-term data strategy
  • Ability to work closely with stakeholders to understand business needs and translate them into data engineering solutions
  • Strong analytical and problem-solving skills, with the ability to identify and address complex data engineering challenges
Job Responsibility
Job Responsibility
  • Collaborate with stakeholders to understand business requirements and translate them into data engineering solutions
  • Design and oversee the overall data architecture and infrastructure, ensuring scalability, performance, security, maintainability, and adherence to industry best practices
  • Define data models and data schemas to meet business needs, considering factors such as data volume, velocity, variety, and veracity
  • Select and integrate appropriate data technologies and tools, such as databases, data lakes, data warehouses, and big data frameworks, to support data processing and analysis
  • Create scalable and efficient data processing frameworks, including ETL (Extract, Transform, Load) processes, data pipelines, and data integration solutions
  • Ensure that data engineering solutions align with the organization's long-term data strategy and goals
  • Evaluate and recommend data governance strategies and practices, including data privacy, security, and compliance measures
  • Collaborate with data scientists, analysts, and other stakeholders to define data requirements and enable effective data analysis and reporting
  • Provide technical guidance and expertise to data engineering teams, promoting best practices and ensuring high-quality deliverables
  • Support to team throughout the implementation process, answering questions and addressing issues as they arise
What we offer
What we offer
  • Stable employment
  • “Office as an option” model
  • Flexibility regarding working hours and your preferred form of contract
  • Comprehensive online onboarding program with a “Buddy” from day 1
  • Cooperation with top-tier engineers and experts
  • Unlimited access to the Udemy learning platform from day 1
  • Certificate training programs
  • Upskilling support
  • Internal Gallup Certified Strengths Coach to support your growth
  • Grow as we grow as a company
Read More
Arrow Right

Data Engineer

The Data Engineer is accountable for developing high quality data products to su...
Location
Location
India , Pune
Salary
Salary:
Not provided
https://www.citi.com/ Logo
Citi
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 5-8 years of relevant experience
  • Experience with 'big data' platforms such as Hadoop, Hive or Snowflake for data storage and processing
  • Good exposure to data modeling techniques
  • Design, optimization and maintenance of data models and data structures
  • Proficient in one or more programming languages commonly used in data engineering such as Python, PySpark
  • Understanding of Data Warehousing concepts
  • Demonstrated problem-solving and decision-making skills
  • Ability to work under pressure and manage deadlines or unexpected changes in expectations or requirements
  • Bachelor's degree/University degree or equivalent experience
Job Responsibility
Job Responsibility
  • Developing high quality data products to support the Bank's regulatory requirements and data driven decision making
  • Serving as an example to other team members
  • Working closely with customers
  • Removing or escalating roadblocks
  • Contributing to business outcomes on an agile team
What we offer
What we offer
  • Resources to meet unique needs
  • Empowerment to make healthy decisions
  • Support for managing financial well-being
  • Help planning for future
  • Fulltime
Read More
Arrow Right

Data Engineer

The Data Engineer is accountable for developing high quality data products to su...
Location
Location
India , Pune
Salary
Salary:
Not provided
https://www.citi.com/ Logo
Citi
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 5-8 years of relevant experience
  • Experience with 'big data' platforms such as Hadoop, Hive or Snowflake for data storage and processing
  • Good exposure to data modeling techniques
  • Design, optimization and maintenance of data models and data structures
  • Proficient in one or more programming languages commonly used in data engineering such as Python, PySpark
  • Understanding of Data Warehousing concepts
  • Demonstrated problem-solving and decision-making skills
  • Ability to work under pressure and manage deadlines or unexpected changes in expectations or requirements
  • Bachelor's degree/University degree or equivalent experience
Job Responsibility
Job Responsibility
  • Developing high quality data products to support the Bank's regulatory requirements and data driven decision making
  • Serving as an example to other team members
  • Working closely with customers
  • Removing or escalating roadblocks
  • Contributing to business outcomes on an agile team
What we offer
What we offer
  • Resources to meet unique needs
  • Empowerment to make healthy decisions
  • Support for managing financial well-being
  • Help planning for future
  • Fulltime
Read More
Arrow Right

Associate Data Engineer

This role offers a unique opportunity to work at the intersection of a pioneerin...
Location
Location
India , Mumbai
Salary
Salary:
Not provided
blenheimchalcot.com Logo
Blenheim Chalcot
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 1–2 years of experience in data engineering or related fields
  • Strong foundational knowledge of data engineering principles
  • Proficiency in Python and SQL
  • Familiarity with Databricks (preferred)
  • Experience with PySpark and Google Cloud Platform (GCP)
  • Excellent problem-solving and communication skills
  • Ability to work independently and take ownership of tasks
Job Responsibility
Job Responsibility
  • Support the development, optimization, and maintenance of scalable data pipelines
  • Collaborate with cross-functional teams to ensure data integrity and accessibility
  • Assist in data ingestion, transformation, and integration from various sources
  • Contribute to documentation and best practices for data engineering workflows
  • Participate in code reviews and continuous improvement initiatives
What we offer
What we offer
  • Be part of the UK's Leading Digital Venture Builder
  • Opportunity to learn from and collaborate with diverse talent across BC
  • Exposure to GenAI-enabled ventures and cutting-edge technologies
  • A fun and open, cricket-obsessed atmosphere – we own the Rajasthan Royals IPL team
  • 24 days of annual leave & 10 public holiday days
  • Private Medical for you and your immediate family & Life Insurance for yourself
Read More
Arrow Right