CrawlJobs Logo

Big Data Junior Scientist

https://www.citi.com/ Logo

Citi

Location Icon

Location:
Poland , Warsaw

Category Icon

Job Type Icon

Contract Type:
Employment contract

Salary Icon

Salary:

90440.00 - 142960.00 PLN / Year

Job Description:

We are seeking a highly motivated and analytically driven The Big Data Junior Scientist to join our team. In this pivotal role, you will be responsible for creating insightful reports and conducting in-depth analysis derived from a unique financial monitoring system, an integral component of Citi's Anti-Money Laundering program. You will collaborate within a team of experienced statisticians, engaging with global colleagues to execute complex analyses and discover intricate patterns of money flows, leveraging advanced AI tools. This presents an exceptional opportunity to operate within a Big Data environment, contribute to the creation and implementation of new projects at Citi, and continuously enhance your English communication skills in daily interactions. This position embraces a hybrid work model, offering a blend of in-office collaboration and remote flexibility.

Job Responsibility:

  • Conduct data-driven analyses and reports
  • apply qualitative and quantitative data analysis methods, prepare statistical and non-statistical data exploration
  • Apply statistical methods to organize, analyze, and interpret data related to AML monitoring scenarios and thresholds
  • Build and implement models and algorithms
  • Detect data anomalies and identify source of data problems
  • Use visualization techniques to display data and the results of analysis in clear presentations
  • Design statistically valid samples, design and conduct statistical experiments
  • Work with existing AI models and contribute to the exploration and implementation of new AI-driven solutions for data analysis and pattern recognition
  • Ensure Data Quality and Reliability
  • Design and implement various Data Quality methods

Requirements:

  • Master or Bachelor degree in technical science/mathematics/statistics/quantitative methods (degree in economics/banking may be considered if strong analytical skills proved) (4-5th year students may be considered as well)
  • Some experience in data mining methods and their application in business practice
  • Programming and statistical/data analytical skills (SQL, Python)
  • Initiative, creativity and attention to details
  • Ability to discuss reasoning, critical assessment of facts and opinions
  • Demonstrated ability to communicate effectively, both orally and in writing
  • Self-motivated with high desire of self-development and learning

Nice to have:

Big Data experience

What we offer:
  • Employer paid Defined Contribution Pension Plan contribution of 6% of employee’s pensionable earnings (PPE Program)
  • Employer paid Private Medical Care Package for employees and Private Medical Care Packages for certain family members available at preferential rates
  • Employer paid Life Insurance Program for employees and Life Insurance for certain family members available at preferential rates
  • Employee Assistance Program financed by Employer
  • Paid Parental Leave Program (maternity and paternity leave
  • statutory and 2 weeks additional paid paternity leave)
  • Sport Card for employees subsidized via Social Benefits Fund and Sport Cards for certain family members available at preferential rates
  • Additional benefits from Company’s Social Benefit Fund, in particular: Holidays Allowance, support for sport and cultural activities, team building events
  • Additional day off for volunteering
  • Cafeteria/ flex benefit
  • Opportunity to receive an annual discretionary incentive award
  • Special offers and discounts for employees

Additional Information:

Job Posted:
May 13, 2026

Employment Type:
Fulltime
Work Type:
Hybrid work
Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for Big Data Junior Scientist

Junior Data Scientist - AI

Hewlett Packard Enterprise (HPE) is seeking a Junior Data Scientist - AI to desi...
Location
Location
United States , San Juan
Salary
Salary:
Not provided
https://www.hpe.com/ Logo
Hewlett Packard Enterprise
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor's degree in computer science, engineering, information systems, or closely related quantitative discipline
  • typically, 2-4 years’ experience
  • strong background in statistical and machine learning techniques such as anomaly detection, clustering and ranking of events, time series analysis, event stream mining, hypothesis testing, causal inference, deep learning
  • great at communicating concepts and results
  • strong data visualization skills
  • expert Python coder (PySpark, Scikit-learn)
  • experience with software engineering best practice
  • relevant industry experience in data science, machine learning
  • experience with online learning algorithms, reinforcement learning, semi-supervised learning, or mixed time-series/event streams
  • experience with Agentic AI
Job Responsibility
Job Responsibility
  • Works with domain experts to identify and formalize machine learning problems for wireless and wired network diagnostics, root causing, problem remediation, and optimization
  • discover new problem signatures in customer networks
  • design, implement, and validate machine learning algorithms on big data
  • guide and oversee deployment of implemented machine learning solutions and monitor their operation
  • use Agentic AI to solve networking problems
  • analyses the feature specifications and determines the required coding, testing, and integration activities
  • designs and develops moderate to complex cloud application modules per feature specifications adhering to security policies
  • identifies debugs and creates solutions for issues with code and integration into application architecture
  • develops and executes comprehensive test plans for features adhering to performance, scale, usability, and security requirements
  • deploy cloud-based systems and applications code using continuous integration/deployment (CI/CD) pipelines to automate cloud applications' management, scaling, and deployment
What we offer
What we offer
  • comprehensive suite of benefits supporting physical, financial and emotional wellbeing
  • specific programs catered to helping reach career goals
  • unconditional inclusion.
  • Fulltime
Read More
Arrow Right

Senior Big Data Engineer

Location
Location
United States , Flowood
Salary
Salary:
Not provided
phasorsoft.com Logo
PhasorSoft Group
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Proficiency in Python programming for data manipulation and analysis
  • Experience with PySpark for processing large-scale data
  • Strong understanding and practical experience with big data technologies such as Hadoop, Spark, Kafka, etc.
  • Knowledge of designing and implementing ETL processes for data integration
  • Ability to work with large datasets, perform data cleansing, transformations, and aggregations
  • Familiarity with machine learning concepts and experience implementing ML models
  • Understanding of data governance principles and experience implementing data security measures
  • Ability to create clear and concise documentation for data pipelines and processes
  • Strong teamwork and collaboration skills to work with cross-functional teams
  • Analytical and problem-solving skills to optimize data workflows and processes
Job Responsibility
Job Responsibility
  • Design and develop scalable data pipelines and solutions using Python and PySpark
  • Utilize big data technologies such as Hadoop, Spark, Kafka, or similar tools for processing and analyzing large datasets
  • Develop and maintain ETL processes to extract, transform, and load data into data lakes or warehouses
  • Collaborate with data engineers and scientists to implement machine learning models and algorithms
  • Optimize and tune data processing workflows for performance and efficiency
  • Implement data governance and security measures to ensure data integrity and privacy
  • Create and maintain documentation for data pipelines, workflows, and processes
  • Provide technical leadership and mentorship to junior team members
  • Fulltime
Read More
Arrow Right

Data Scientist

We are seeking an innovative and analytical Data Scientist to unlock valuable in...
Location
Location
Netherlands , Nieuwegein
Salary
Salary:
3000.00 - 6000.00 EUR / Month
https://www.soprasteria.com Logo
Sopra Steria
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • HBO or WO degree in Computer Science, Data Science, Statistics, or a related field
  • Minimum of 2.5 years of experience in data science or a related role
  • Proficiency in programming languages such as Python and R
  • Strong knowledge of machine learning algorithms, statistical modelling, and data mining techniques
  • Experience with big data technologies like Hadoop and Spark
  • Familiarity with cloud platforms (AWS, Azure) and containerisation technologies (Docker, Kubernetes)
  • Expertise in developing and deploying machine learning models to production environments
  • Hands-on experience with natural language processing, computer vision, and deep learning techniques
  • Proficiency in data visualisation tools such as Tableau or Power BI
  • Strong analytical and problem-solving skills with attention to detail
Job Responsibility
Job Responsibility
  • Analyse large datasets using advanced statistical and machine learning techniques to extract meaningful insights
  • Develop and implement predictive models and algorithms to solve complex business problems
  • Utilise programming languages such as Python and R to manipulate data and create efficient data pipelines
  • Collaborate with cross-functional teams to identify opportunities for data-driven improvements and innovations
  • Create compelling data visualisations using tools like Tableau and Power BI to communicate findings to stakeholders
  • Design and implement machine learning solutions, including natural language processing and computer vision applications
  • Contribute to the development of MLOps practices and the implementation of explainable AI (XAI) techniques
  • Optimise and deploy models to production environments using cloud platforms such as AWS and Azure
  • Stay current with the latest advancements in data science, machine learning, and artificial intelligence
  • Provide technical guidance and mentorship to junior team members
What we offer
What we offer
  • Permanent contract
  • Gross monthly salary between €3,000 and €6,000 (based on 40 hours per week)
  • 8% holiday allowance
  • Generous mobility budget including options such as an electric lease car with an NS Business Card, a lease bike, or alternative transportation
  • 8% profit sharing on target (or a fixed OTB amount)
  • 27 paid vacation days
  • Flex benefits budget of €1,800 per year plus an additional percentage of your salary
  • Home office setup with a laptop, phone, and a monthly internet allowance
  • Hybrid working: from home or at the office
  • Development opportunities through training, knowledge-sharing sessions, and inspiring (networking) events
  • Fulltime
Read More
Arrow Right

Data Lake SME

We are looking for an experienced Data Lake / ETL Engineer with 7+ years of expe...
Location
Location
India , Bangalore
Salary
Salary:
Not provided
https://www.hpe.com/ Logo
Hewlett Packard Enterprise
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 7+ years of experience in data engineering, ETL/ELT development, or data lake management
  • Strong expertise in ETL tools (Informatica, Talend, dbt, SSIS, or similar)
  • Hands-on experience with big data ecosystems: Hadoop, Spark, Hive, Presto, Delta Lake, or Iceberg
  • Proficiency with SQL, Python, or Scala for data processing and transformation
  • Experience with cloud data platforms (AWS Glue, Redshift, Azure Synapse, GCP BigQuery)
  • Familiarity with workflow orchestration tools (Airflow, Temporal, Oozie)
Job Responsibility
Job Responsibility
  • Design and implement data ingestion pipelines for structured, semi-structured, and unstructured data
  • Develop and manage ETL/ELT processes for large-scale data processing
  • Optimize storage and retrieval strategies across on-prem and cloud-based data lakes
  • Integrate data from multiple sources (databases, APIs, streaming platforms)
  • Implement real-time and batch processing using Apache Spark, Kafka, or Flink
  • Support metadata management, data lineage, and cataloging
  • Tune queries and pipelines for high performance and cost efficiency
  • Implement partitioning, indexing, and caching strategies for large datasets
  • Automate routine ETL/ELT workflows for reliability and speed
  • Ensure compliance with data governance, privacy, and regulatory standards (GDPR, HIPAA, etc.)
What we offer
What we offer
  • Health & Wellbeing
  • Personal & Professional Development
  • Unconditional Inclusion
  • Fulltime
Read More
Arrow Right

Vice President - Bigdata Engineer - AI & NLP

The Applications Development Technology Lead Analyst is a senior-level position ...
Location
Location
India , Chennai
Salary
Salary:
Not provided
https://www.citi.com/ Logo
Citi
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 13+ years of relevant experience in Apps Development or systems analysis role
  • Extensive experience in system analysis and programming of software applications
  • Experience in managing and implementing successful projects
  • Expert in coding Python in building Machine Learning and developing LLM-based applications in a professional environment
  • SQL skills able to perform data interrogations
  • Proficiency in enterprise-level application development using Java 8, Scala, Oracle (or comparable database), and Messaging infrastructure like Solace, Kafka, Tibco EMS
  • Develop LLM solutions for querying structured data with natural language, including RAG architectures on enterprise knowledge bases
  • Build, scale, and optimize data science workloads, applying best MLOps practices for production
  • Lead the design and development of LLM-based tools to increase data accessibility, focusing on text-to-SQL platforms
  • Train and fine-tune LLM models to accurately interpret natural language queries and generate SQL queries
Job Responsibility
Job Responsibility
  • Partner with multiple management teams to ensure appropriate integration of functions to meet goals
  • Identify and define necessary system enhancements to deploy new products and process improvements
  • Resolve variety of high impact problems/projects through in-depth evaluation of complex business processes, system processes, and industry standards
  • Provide expertise in area and advanced knowledge of applications programming
  • Ensure application design adheres to the overall architecture blueprint
  • Utilize advanced knowledge of system flow and develop standards for coding, testing, debugging, and implementation
  • Develop comprehensive knowledge of how areas of business, such as architecture and infrastructure, integrate to accomplish business goals
  • Provide in-depth analysis with interpretive thinking to define issues and develop innovative solutions
  • Serve as advisor or coach to mid-level developers and analysts, allocating work as necessary
  • Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets
What we offer
What we offer
  • Global Benefits
  • Best-in-class benefits to be well, live well and save well
  • Fulltime
Read More
Arrow Right

Senior Bigdata Developer

The Data Analytics Senior Analyst is a seasoned professional role. Applies in-de...
Location
Location
India , Pune
Salary
Salary:
Not provided
https://www.citi.com/ Logo
Citi
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Database Development & Architecture:Design, develop, and maintain complex database solutions across MongoDB, Oracle, and other SQL databases. Create optimal data models, schemas, and stored procedures to support high-throughput applications.
  • Data Pipeline Construction:Build and manage resilient, scalable ETL/ELT pipelines using Python to process and integrate large volumes of data from diverse source systems into our core data platforms.
  • Big Data Engineering:Engineer and implement solutions within our Big Data ecosystem (e.g., Hadoop, Spark, Hive, Kafka) to handle large-scale data processing, batch analytics, and real-time data streams.
  • Python Development:Write high-quality, production-ready Python code for data manipulation, API development, and automation. Utilize a range of libraries and frameworks relevant to data engineering (e.g., Pandas, PySpark, SQLAlchemy, PyMongo).
  • Performance Optimization:Proactively monitor, troubleshoot, and optimize the performance of our databases and data pipelines. Focus on query tuning, indexing strategies, and resource management to ensure low-latency data access.
  • Data Quality and Integrity:Implement data quality checks, validation rules, and monitoring frameworks within the data pipelines to ensure the accuracy, consistency, and reliability of our KYC data.
  • Collaboration:Work closely with application developers, data scientists, and data analysts to understand their data requirements and provide robust, well-documented data solutions and services.
  • Technical Leadership:Provide subject matter expertise on database and data engineering best practices. Mentor junior engineers and contribute to a culture of technical excellence.
  • Education:Bachelor’s/University degree or equivalent experience
Job Responsibility
Job Responsibility
  • Applies in-depth disciplinary knowledge, contributing to the development of new techniques and the improvement of processes and work-flows.
  • Coordinates and contribute to the objectives of data science initiatives and overall business through leveraging in-depth understanding of how areas collectively integrate within the sub-function.
  • Assumes informal/formal leadership role through coaching and training of new recruits.
  • Significantly influences decisions, work, and performance of all teams through advice, counsel and/or facilitating services to others in the business.
  • Conducts strategic data analysis, identifies insights and implications and make strategic recommendations, develops data displays that clearly communicate complex analysis.
  • Mines and analyzes data from various banking platforms to drive optimization and improve data quality.
  • Delivers analytics initiatives to address business problems with the ability to identify data required, assess time & effort required and establish a project plan.
  • Consults with business clients to identify system functional specifications. Applies comprehensive understanding of how multiple areas collectively integrate to contribute towards achieving business goals.
  • Consults with users and clients to solve complex system issues/problems through in-depth evaluation of business processes, systems and industry standards
  • recommends solutions.
  • Fulltime
Read More
Arrow Right

Data Scientist

At Vodafone, we’re not just shaping the future of connectivity for our customers...
Location
Location
Italy , Milano
Salary
Salary:
Not provided
vodafone.com Logo
Vodafone
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Degree in Computer Science, Maths, Engineering or equivalent
  • Junior Profile with Experience in similar position (Max 2 years)
  • SQL, Python (Pandas, Tensorflow, Scikit-learn, and main other ML libs), Pyspark and SW development capabilities
  • Machine learning algorithms knowledge (NLP, Neural Networks., Random Forest, SVM, Anomaly Detection especially on time series, Gradient Boost and all other main ML models both supervised and unsupervised)
  • Knowledge of deployment best practices and DevOps pipeline
  • Excellent analytics and mathematics skills
  • Professional English (spoken and written)
  • Experience in Machine Learning SW development and data analysis
  • Experience in designing and implementing use cases over big data architectures involving massive data volume, also under real-time constraints
  • Knowledge of pros and cons of existing data storage technologies (relational DB, Big Data Frameworks, no-SQL DB on cloud and on prem)
Job Responsibility
Job Responsibility
  • Identification of data science / big data / analytics use cases for Network Operations and architectural High Level Design
  • Choice and implementation of the best machine learning algorithm suited to the use case
  • Industrialization of the use cases on Cloudera, Openshift/Kubernetes or on AWS/Google cloud environments, with the support of data engineers
  • Technical leadership in analysis and data management domains
  • Data-driven evaluation of vendor product adoption
Read More
Arrow Right

Fullstack Big Data Developer Application Development Technical Lead Analyst Vice President

Discover your future at Citi. Working at Citi is far more than just a job. A car...
Location
Location
Canada , Mississauga
Salary
Salary:
120800.00 - 170800.00 USD / Year
https://www.citi.com/ Logo
Citi
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 6+ years of Application development experience
  • 6+ years of experience in full stack development, with a focus on Bigdata and Python/Scala
  • 6+ years experience with big data technologies such as Python, Pyspark, Hadoop, Kafka, etc.
  • Experience with Core Java/J2EE Application with complete command over OOPs and Design Patterns
  • Commendable in Data Structures and Algorithms
  • Worked on Core Application Development of complex size encompassing all areas of Java/J2EE
  • Thorough knowledge and hands on experience in following technologies Hadoop, Map Reduce Framework, Spark, YARN, Sqoop, Pig , Hue, Unix, Java, Sqoop, Impala, Cassandra on Mesos
  • Should have implemented or part complex project execution in Big Data Spark eco system, where processing volumes of data thorough understanding of distributed processing and integrated applications
  • Exposure to ETL and BI tools
  • Work in an agile environment following through the best practices of agile Scrum
Job Responsibility
Job Responsibility
  • Partner with multiple management teams to ensure appropriate integration of functions to meet goals as well as identify and define necessary system enhancements to deploy new products and process improvements
  • Resolve variety of high impact problems/projects through in-depth evaluation of complex business processes, system processes, and industry standards
  • Provide expertise in area and advanced knowledge of applications programming and ensure application design adheres to the overall architecture blueprint
  • Utilize advanced knowledge of system flow and develop standards for coding, testing, debugging, and implementation
  • Develop comprehensive knowledge of how areas of business, such as architecture and infrastructure, integrate to accomplish business goals
  • Provide in-depth analysis with interpretive thinking to define issues and develop innovative solutions
  • Serve as advisor or coach to mid-level developers and analysts, allocating work as necessary
  • Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency
  • Design, develop, and maintain scalable and robust architecture for the project using Java/Python/Scala and other full stack technologies
  • Manage big data technologies such as python, pyspark to ensure seamless data integration, storage, and analysis
  • Fulltime
Read More
Arrow Right