MoMo is the market leader in mobile payments in Vietnam. We strive to make life better for every Vietnamese with innovation in technology. As many parts of the business grow, we are looking for an experienced Data Engineer to join the team. At MoMo BigData AI department, we focus on Smart, Efficienct, Execellent execution and data quality which providing high quality marterial for:
We are striving to build the new hydrid data platform on multiple cloud vendors helping us to better control the budget and technologies behind.
What you will do
With the AI- first company mission, design and build the self- serve data platform to serve people in MoMo and partner needs. They will be allocated resources based on their needs to:
Ingesting multiple data sources, from batch to stream, from pull to push mechanism.
Developing and deploying the resilient data pipeline in the data lake, data warehouse, and streaming data.
Delivering the derived data downstream with high quality, such as BI solution (PowerBI, Google Data Studio,…), Marketing Platform, Promotion Platform, and Experiment Platform,… in multiple ways (API, Dataset, Streaming Data,…)
Building Machine Learning models, Business Intelligent dashboards.
Monitoring the data quality of data pipelines in the data platform.
Monitoring and optimizing resource usage.
Design and build the Data Management Systems enabling the Data Governance team and data consumers:
Managing the data life cycle in the big data platform.
Having MoMo data ecosystem self- exploration ability.
Proving the single source of truths to multiple downstream with high quality.
Managing the attribution infrastructure cost from various big projects, teams, and departments.
Design and build the Data Loss Prevention (DLP) solution to protect our data and achieve data visibility in large organizations.
Collaborate with Machine Learning Engineers, Data Scientists, Business Analysts, Data Analysts, Product Owners, and Product Operators to strive for greater functionality in our data systems.
What you will need
Language: Java/Kotlin, SQL, Python, Scala, BashScripts;
Love of data. You run “SELECT COUNT(SHEEP) FROM BACKYARD” during your sleep!
Data Source: App Events, Oracle, MySQL, MSSQL, Kafka, Pubsub, REST API;
Data Warehouse: BigQuery, Trino, Clickhouse.
Experience with cloud platforms such as Google Cloud Platform or Amazon Web Service is a plus.
Data Pipeline Orchestration: Airflow;
Infrastructure: Google Cloud Platform, Amazon Web Service, Kubernetes, Docker;
BI: Datastudio, PowerBI;
Strong programming skills;
Problem- solving skills, Teamwork spirit;