This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
The Data Engineer is responsible for building Data Engineering Solutions using next generation data techniques. The individual will be working directly with product owners, customers and technologists to deliver data products/solutions in a collaborative and agile environment.
Job Responsibility:
Responsible for design and development of big data solutions
Partner with domain experts, product managers, analyst, and data scientists to PySpark and Python
Work with data scientist to build Client pipelines using heterogeneous sources and provide engineering services for data science applications
Ensure automation through CI/CD across platforms both in cloud and on-premises
Define needs around maintainability, testability, performance, security, quality and usability for data platform
Drive implementation, consistent patterns, reusable components, and coding standards for data engineering processes
Convert Talend based pipelines into languages like PySpark, Python to execute on Hadoop and non-Hadoop ecosystems
Tune Big data applications on Hadoop and non-Hadoop platforms for optimal performance
Evaluate new IT developments and evolving business requirements and recommend appropriate systems alternatives and/or enhancements to current systems by analyzing business processes, systems and industry standards
Applies in-depth understanding of how data analytics collectively integrate within the sub-function as well as coordinates and contributes to the objectives of the entire function
Produces detailed analysis of issues where the best course of action is not evident from the information available, but actions must be recommended/taken
Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency
Requirements:
4 -8 years of total IT experience
4+ years of relevant experience with Pyspark and Python
Experience on designing and developing Data Pipelines for Data Ingestion or Transformation using Python
Experience with Spark programming (pyspark or Python)
Hands-on experience with Python/Pyspark and basic libraries for machine learning is required
Exposure to containerization and related technologies (e.g. Docker, Kubernetes)
Exposure to aspects of DevOps (source control, continuous integration, deployments, etc.)
Can-do attitude on solving complex business problems, good interpersonal and teamwork skills
Possess team management experience and have led a team of data engineers and analysts
Experience in Oracle performance tuning, SQL, Autosys and basic Unix scripting
Welcome to CrawlJobs.com – Your Global Job Discovery Platform
At CrawlJobs.com, we simplify finding your next career opportunity by bringing job listings directly to you from all corners of the web. Using cutting-edge AI and web-crawling technologies, we gather and curate job offers from various sources across the globe, ensuring you have access to the most up-to-date job listings in one place.
We use cookies to enhance your experience, analyze traffic, and serve personalized content. By clicking “Accept”, you agree to the use of cookies.