Job Description

Data Engineer / Streaming Engineer📍 Johannesburg, Gauteng, South Africa (Hybrid)🕒 Employment Type: Full-TimeAbout the RoleWe are looking for a highly skilled and motivated Data Engineer / Streaming Engineer to join our team. This role is central to building and supporting scalable, secure, and repeatable data pipelines that power intelligent decision-making. You’ll work in an agile environment, integrating diverse data sources and leveraging emerging technologies across big data and cloud platforms to deliver innovative solutions. A Software development is essential for this role! Key ResponsibilitiesArchitect and implement scalable data analytics frameworks.Translate complex functional and technical requirements into high-performing software solutions.Lead the development of batch and real-time data pipelines using transformational technologies.Create and support data feeds from on-premise systems to Google Cloud Platform.Build data marts using Talend or similar ETL tools.Manipulate and process data using Python and PySpark/Spark.Leverage AWS services such as EMR, EC2, and S3 for big data processing.Support production data feeds and troubleshoot issues on a break-fix basis.Contribute to the research, evaluation, and implementation of new tools and technologies.Ensure compliance with policies, standards, and disaster recovery protocols.Collaborate across teams to deliver robust data solutions aligned with business needs.Required Experience & SkillsBachelor’s degree in Computer Science, Computer Engineering, or equivalent.AWS Certification (preferred).5+ years of experience in data engineering and software engineering.2+ years of experience with big data technologies.2+ years of experience with ETL processes and tools (e.g., Talend).2+ years of hands on experience with GCP - Google Cloud Platform is ESSENTIALStrong programming skills in Python and scripting languages.Experience with PySpark or Spark for distributed data processing.Deep understanding of data modeling and various data structures.Proven ability to architect scalable distributed systems using open-source tools.Familiarity with Hadoop and AWS EMR.Knowledge of automated testing and deployment practices.Technologies You’ll Work WithBig Data Tools: Hadoop, Spark, PySparkETL Tools: TalendCloud Platforms: AWS (EMR, EC2, S3)Languages: Python, scripting languagesData Processing: Batch and streaming pipelinesWhy join us?Work on impactful, data-driven projects using the latest technologies.Be part of a collaborative and innovative team.Opportunities for continuous learning and professional growth.Competitive compensation and benefits.

Job Application Tips

  • Tailor your resume to highlight relevant experience for this position
  • Write a compelling cover letter that addresses the specific requirements
  • Research the company culture and values before applying
  • Prepare examples of your work that demonstrate your skills
  • Follow up on your application after a reasonable time period