Job Description
DATA ENGINEER (Hybrid, Jakarta)Office- > Equity Tower, Jendral Sudirman Kav. 52 53 SCBD, Kebayoran Baru Jakarta Selatan, 12190About the RoleOur client, a leading regional cloud and AI transformation partner, helps enterprises modernize their data infrastructure and implement advanced analytics and AI solutions across Asia. With a strong presence in Southeast Asia and over 1,200 enterprise customers, our client specializes in data modernization, cloud migration, and AI-driven business intelligence.We are seeking a Data Engineer experienced in designing and implementing end-to-end data pipelines on cloud platforms, particularly GCP or Azure. This role will work on a large-scale retail data modernization project involving data integration, transformation, and optimization across multiple sources and environments.Location: Jakarta (Hybrid working arrangement)Key ResponsibilitiesDesign, build, and maintain batch and streaming data pipelines using GCP (BigQuery, Dataflow, Dataproc, Composer, Dataform, Cloud Functions).Perform ETL/ELT operations to load and optimize data in BigQuery for analytics and reporting.Integrate data from APIs, databases, and file-based systems across multiple environments.Support data migration from legacy systems (Oracle, MicroStrategy, etc.).Ensure data governance, data quality, and compliance with organizational standards.Collaborate with business intelligence and analytics teams to support reporting and dashboard needs.Requirements3–5 years of experience in data engineering or ETL development.Proven experience with GCP Data Stack (BigQuery, Dataflow, Dataproc, Composer).Strong skills in SQL and Python for data transformation and automation.Familiarity with Azure Data Stack (Data Factory, Databricks, Synapse Analytics, Data Lake) is a plus.Understanding of data modeling, performance tuning, and orchestration tools (Airflow or dbt).Exposure to data migration or modernization projects preferred.Strong problem-solving mindset, collaborative, and proactive learner.Nice to HaveGCP Professional Data Engineer or Cloud Architect certification.Experience with CI/CD pipelines, CDC (Change Data Capture), or data governance frameworks.Skills Must-Have GCP BigQuery, Dataflow, Dataproc, Composer, Dataform, Python, SQL, ETL/ELT, Airflow/dbt, Data Modeling, Data Migration
Job Application Tips
- Tailor your resume to highlight relevant experience for this position
- Write a compelling cover letter that addresses the specific requirements
- Research the company culture and values before applying
- Prepare examples of your work that demonstrate your skills
- Follow up on your application after a reasonable time period