Portfolio Company Careers

Discover opportunities across our network of values-driven companies!
Sovereign’s Capital
Sovereign’s Capital

Senior Data Engineer (Digibank)India



Data Science
Posted on Saturday, September 2, 2023

Life at Grab

At Grab, every Grabber is guided by The Grab Way, which spells out our mission, how we believe we can achieve it, and our operating principles - the 4Hs: Heart, Hunger, Honour and Humility. These principles guide and help us make decisions as we work to create economic empowerment for the people of Southeast Asia.

Get to know the Team

We are living in dynamic times. Technology is reshaping how we live, and we want to use it to redefine how financial services are offered. Singtel is Asia’s leading communications group connecting millions of consumers and enterprises to essential digital services. This is why we are coming together to unlock big dreams, and financial inclusion for people in our region is just one of them. We want to build a digital bank with the right foundation - using data, technology and trust to solve problems and serve customers. If you have what it takes to help build this new Digibank with us.

Get to know the Role:

As the Data Engineer in the Data Technology team, you will be working on all aspects of Data, from Platform and Infra build out to pipeline engineering and writing tooling/services for augmenting and fronting the core platform. You will be responsible for building and maintaining the state-of-the-art data Life Cycle management platform, including acquisition, storage, processing and consumption channels. The team works closely with Data scientists, Product Managers, Finance, Legal, Compliance and business stakeholders across the SEA in understanding and tailoring the offerings to their needs. As a member of the Data Tech team, you will be an early adopter and contributor to various open source big data technologies and you are encouraged to think out of the box and have fun exploring the latest patterns and designs in the fields of Software and Data Engineering.

The day-to-day activities: ​Build and manage the data asset using some of the most scalable and resilient open source big data technologies like Airflow, Spark, Snowflake, Kafka, Kubernetes, ElasticSearch, Superset and more on cloud infrastructure.

  • Design and deliver the next-gen data lifecycle management suite of tools/frameworks, including ingestion and consumption on the top of the data lake to support real-time, API-based and serverless use-cases, along with batch (mini/micro) as relevant

  • Build and expose metadata catalog for the Data Lake for easy exploration, profiling as well as lineage requirements

  • Enable Data Science teams to test and productionize various ML models, including propensity, risk and fraud models to better understand, serve and protect our customers. Lead technical discussions across the organization through collaboration, including running RFC and architecture review sessions, tech talks on new technologies as well as retrospectives

  • Apply core software engineering and design concepts in creating operational as well as strategic technical roadmaps for business problems that are vague/not fully understood. Obsess security by ensuring all the components, from a platform, frameworks to the applications are fully secure and are compliant by the group’s infosec policies.

The must haves:

  • At least 5+ years of relevant experience in developing scalable, secured, distributed, fault tolerant, resilient & mission-critical Big Data platforms.

  • Able to maintain and monitor the ecosystem with 99.99% availability

  • Candidates will be aligned appropriately within the organization depending on experience and depth of knowledge.

  • Must have good fundamental hands-on knowledge of Linux and building a big data stack on top of AWS using Kubernetes.

  • Proficiency in at least one of the programming languages Python, Scala or Java.

  • Strong understanding of big data and related technologies like Spark, Airflow, Kafka etc.

  • Experience with NoSQL databases – KV, Document and Graph

  • Able to drive devops best practices like CI/CD, containerization, blue-green deployments, 12-factor apps, secrets management etc in the Data ecosystem.

  • Good understanding of Machine Learning models and efficiently supporting them is a plus.

Our Commitment

We are committed to building diverse teams and creating an inclusive workplace that enables all Grabbers to perform at their best, regardless of nationality, ethnicity, religion, age, gender identity or sexual orientation and other attributes that make each Grabber unique.

Forward together

Follow us and keep updated!

Equal opportunity

Grab is an equal opportunity employer. We owe our success to the talents of our globally-diverse team and the varying perspectives they add to our thriving community.

Recruitment agencies

Grab does not accept unsolicited resumes sent by recruiting agencies. Please do not forward resumes to our job postings, Grab employees or other parts of the business. Grab will not be liable to pay any fees to agencies for candidates hired as a result of unrequested resumes.