CrawlJobs Logo

Big Data Engineer

https://www.citi.com/ Logo

Citi

Location Icon

Location:
India, Pune

Category Icon
Category:
IT - Software Development

Job Type Icon

Contract Type:
Employment contract

Salary Icon

Salary:

Not provided

Job Description:

The Applications Development Programmer Analyst is responsible for participation in the establishment and implementation of new or revised application systems and programs in coordination with the Technology team. Additionally, the Big Data Engineer role focuses on collecting, storing, processing, and analyzing large datasets, while maintaining, implementing, and monitoring optimal solutions.

Job Responsibility:

  • Utilize knowledge of applications development procedures and concepts, and basic knowledge of other technical areas to identify and define necessary system enhancements
  • identify and analyze issues, make recommendations, and implement solutions
  • utilize knowledge of business processes, system processes, and industry standards to solve complex issues
  • analyze information and make evaluative judgements to recommend solutions and improvements
  • conduct testing and debugging, utilize script tools, and write basic code for design specifications
  • assess applicability of similar experiences and evaluate options under circumstances not covered by procedures
  • develop working knowledge of Citi’s information systems, procedures, standards, client server application development, network operations, database administration, systems administration, data center operations, and PC-based applications
  • selecting and integrating any Big Data tools and frameworks required to provide requested capabilities
  • implementing data wrangling, scraping, cleaning using both Java or Python.

Requirements:

  • 2+ years of relevant experience
  • experience in programming/debugging used in business applications
  • working knowledge of industry practice and standards
  • comprehensive knowledge of specific business area for application development
  • working knowledge of program languages
  • consistently demonstrates clear and concise written and verbal communication
  • proficient understanding of distributed computing principles
  • proficient in Java or Python and some part of machine learning
  • proficiency with Hadoop v2, MapReduce, HDFS, Pyspark, Spark
  • experience with building stream-processing systems, using solutions such as Storm or Spark-Streaming
  • good knowledge of Big Data querying tools, such as Pig, Hive, and Impala
  • experience with integration of data from multiple data sources
  • experience with NoSQL databases, such as HBase, Cassandra, MongoDB
  • knowledge of various ETL techniques and frameworks, such as Flume
  • experience with various messaging systems, such as Kafka or RabbitMQ
  • experience with Big Data ML toolkits, such as Mahout, SparkML, or H2O
  • good understanding of Lambda Architecture, along with its advantages and drawbacks
  • experience with Cloudera, MapR, or Hortonworks.
What we offer:

equal consideration without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, veteran status, or any other protected characteristic.

Additional Information:

Job Posted:
July 10, 2025

Employment Type:
Fulltime
Work Type:
On-site work
Job Link Share:
Welcome to CrawlJobs.com
Your Global Job Discovery Platform
At CrawlJobs.com, we simplify finding your next career opportunity by bringing job listings directly to you from all corners of the web. Using cutting-edge AI and web-crawling technologies, we gather and curate job offers from various sources across the globe, ensuring you have access to the most up-to-date job listings in one place.