We are looking for a Data Engineer to help build and manage robust data pipelines and platforms that power advanced analytics and AI solutions.
Key Responsibilities
- Develop and maintain batch and streaming data pipelines using GCP services , such as BigQuery, Dataflow, Dataproc, Composer, Dataform, Cloud Functions .
- Load, transform, and optimize data in BigQuery for analytics and reporting.
- Integrate data from multiple sources including APIs, databases, and files.
- Assist in data migration from legacy systems such as Oracle and MicroStrategy.
- Ensure data quality, governance, and security compliance.
- Collaborate with analysts and business teams to support reporting needs.
Requirements
3–5 years’ experience in data engineering or ETL development.Hands-on experience with GCP Data Stack (BigQuery, Dataflow, Composer, Dataproc).Solid SQL and Python skills.Familiarity with Azure Data Stack is a plus.Understanding of data modelling concepts and performance optimization.Willingness to learn and work on large-scale migration projects.CloudMile is an equal opportunities employer and welcomes applications from all qualified candidates.
#J-18808-Ljbffr