The role will be responsible to bridge the gap between data science and IT operations, enabling seamless model lifecycle management and scalable ML infrastructure. Apart from that, the role is also responsible for designing, developing and optimizing data pipelines, ensuring reliable and efficient data workflows across the organization for structure and unstructured data related to Boost's use cases and mainly to support Boost's to incorporate unified data modelling across AI, Machine learning, and Analytics projects. This role will collaborate with cross-functional teams, including data scientists, analysts, software engineers and DevOps to optimize the production and deployment of machine learning solutions, and support and enhance data-driven decision-making and analytics.
SCOPE & AUTHORITY
QUALIFICATIONS
Bachelor's Degree in Computer Science, Data Engineering, Machine Learning or a related field with a minimum of 3+ years of experience in MLOps / Data Engineer - designing, developing and maintaining huge Data Warehouse and analytics projects.
Strong problem-solving skills, collaboration skills, adaptability to evolving technology, commit to process improvement, attention to detail and the ability to communicate technical concepts effectively to non-technical stakeholders.
Strong knowledge on cloud platforms for data solutions (AWS / Azure / GCP / Databricks / Snowflake). Strong knowledge in ETL / ELT tools (e.g., Apache Airflow, AWS Glue, Databricks Jobs / Pipelines). Proficiency in data modeling and schema design.
Proficiency in SQL and data warehouse (e.g., Redshift, BigQuery, Databricks, Snowflake or similar).
Familiarity with ML / MLOps frameworks (e.g., mlflow, TensorFlow, PyTorch, scikit-learn).
Familiarity with data governance frameworks and best practices.
Familiarity with data lake architectures.
Familiarity with big data processing frameworks (e.g., Apache Spark, Hadoop).
Familiarity with Infrastructure as Code - IaC (e.g., Terraform, Terragrunt, Serverless framework).
Experience with CI / CD tools (e.g., Jenkins, GitLab CI / CD).
Data Engineer • Kuala Lumpur, Kuala Lumpur, Malaysia