Required Competencies
· Solid understanding of data pipeline orchestration (e.g., cron, Airflow, Power Automate) and containerization (e.g., Docker)
· Strong programming skills in Python, with experience building modular and reusable components for data workflows
· Familiar with version control, infrastructure documentation, and maintaining data dictionaries or runbooks
· Fluent in CI/CD and automation frameworks, particularly using GitLab pipelines, Power Automate, and scripting on Linux/Windows environments
· Demonstrated ability to collaborate with cross- functional technical teams, supporting data infrastructure integration and maintenance
· Proficient in SQL for data manipulation and performance tuning across relational databases (PostgreSQL, MySQL)
· Hands- on experience with AWS services such as Lambda, Glue, S3, RDS, and Athena, with a focus on cost and performance optimization
· Knowledge of data governance, privacy, and quality control, including validation, monitoring, and alerting mechanisms
Qualification and Experience Required
· Familiarity with data quality, security, and governance best practices
· Proven track record in building and maintaining production- grade ETL/ELT pipelines using Python and SQL
· Minimum 3 years of hands- on experience in a data engineering or backend data infrastructure role
· Professional certification in data engineering, cloud architecture, or AWS (e.g., AWS Certified Data Analytics or Solutions Architect) is an advantage
· Bachelor’s degree in Computer Science, Information Technology, Engineering, or a related technical field
· Experience in a multi- business or regional group environment is a plus
· Experience working with DevOps practices, CI/CD pipelines (e.g., GitLab), and automation tools for deployment and workflow orchestration
· Demonstrated experience managing cloud- based infrastructure, with a strong understanding of cost optimization, monitoring, and scalability (preferably on AWS)