This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
The Big data engineer is a developing professional role in the field of Data engineering. The data engineer is accountable for developing high quality data products to support the Bank's regulatory and in-business data requirements and data driven decision making. The team is seeking a skilled Big Data Engineer to design, build, and maintain robust, scalable data pipelines using Apache Spark.
Job Responsibility:
Developing and supporting scalable, extensible, and highly available data solutions
Deliver on critical business priorities while ensuring alignment with the wider architectural vision
Identify and help address potential risks in the data supply chain
Follow and contribute to technical standards
Design and develop analytical data models
Considers the business implications of the application of technology to the current business environment
identifies and communicates risks and impacts
Employs developed communication and diplomacy skills to exchange potentially complex/sensitive information
Demonstrates attention to quality and timeliness of service to ensure the effectiveness of the team and group
Provides informal guidance or on-the-job-training to new team members
Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency
Requirements:
2+ years of hands-on experience in building data pipelines with Apache Spark
Strong experience in Big Data platforms including Hadoop, Hive, or Snowflake for data storage and processing
Strong proficiency in either Python or Java programming
Expertise in data modeling techniques, including the design and structuring of data models
Comprehensive understanding of Data Warehousing principles, alongside experience with RDBMS (Oracle, MSSQL, MySQL) and NoSQL databases (MongoDB, DynamoDB)
Familiarity with DevOps concepts, specifically CI/CD platforms and version control
Bachelor’s/University degree or equivalent experience
Nice to have:
Exposure to data quality management, controls, validation, and enrichment
Understanding of containerization technologies like Docker and Kubernetes
Experience with various event, file, and table formats, such as Parquet, ORC, and Iceberg
Basic knowledge of job schedulers (e.g., Autosys) and entitlement management
What we offer:
medical, dental & vision coverage
401(k)
life, accident, and disability insurance
wellness programs
paid time off packages, including planned time off (vacation), unplanned time off (sick leave), and paid holidays