Job Description
MLOps EngineerThe Role:We are looking for a skilled MLOps Engineer to join our team and play a key role in bridging the gap between our Data Science and Infrastructure teams. You will be responsible for supporting the Data Science team in MLOps-related tasks while also helping in DevOps initiatives, including CI/CD pipeline creation, provisioning of cloud resources using tools like Terraform, and Kubernetes orchestration. You will collaborate closely with data scientists and engineers to deploy data pipelines, train machine learning models, and manage their deployment within scalable cloud environments while ensuring high performance, security, and reliability throughout the ML lifecycle.The main responsibilities of the position include:Assist in designing, implementing, and maintaining scalable MLOps pipelines on AWS using services such as SageMaker, EC2, EKS, S3, Lambda and other relevant AWS toolsCoordinate with our platform team to troubleshoot Kubernetes clusters (EKS) to orchestrate the deployment of machine learning models and other microservicesDevelop and maintain CI/CD pipelines for model and application deployment, testing, and monitoringCollaborate closely with Data Science, and DevOps team to streamline the model development lifecycle, from experimentation to production deploymentImplement security best practices, including network security, data encryption, and role-based access controls within the AWS infrastructureMonitor, troubleshoot, and optimize data and ML pipelines to ensure high availability and performanceSet up and manage model monitoring systems for performance drift, ensuring continuous model improvementMain requirements:Bachelor’s degree in Computer Science, Engineering, or related field1+ years of hands-on experience in MLOps, DevOps, or related fieldsKnowledge and preferable working experience in AWS services for machine learning, such as SageMaker, EKS, S3, EC2, Lambda, and othersExposure to Kubernetes for container orchestrationExperience with DockerExposure to infrastructure-as-code tools such as Terraform or CloudFormationFamiliarity with CI/CD tools such as GitLab CIUnderstanding machine learning model lifecycleFamiliarity with monitoring and logging solutions like Prometheus, Grafana, CloudWatch and ELK StackUnderstanding of networking concepts and cloud security best practicesProficiency in Python and Bash, and comfortable working in Linux environmentsStrong problem-solving and communication skillsThe following will be considered an advantage:Experience working with serverless architectures and event-driven processing on AWSFamiliarity with advanced Kubernetes concepts such as HelmExperience with Data Engineering pipelines, ETL processes, or big data platformsExperience with ML frameworks like TensorFlow, PyTorch and KerasExperience with ML platforms like Kubeflow and/or SageMakerExperience with workflow engines like Argo Workflows and/or AirflowBenefit from:Attractive remuneration package plus performance related rewardPrivate health insuranceCorporate pension fundIntellectually stimulating work environmentContinuous personal development and international training opportunitiesThe Hiring Experience: What Awaits YouLet’s Connect – Intro Chat with Talent AcquisitionDeep Dive – First Interview with Your Future TeamFinal Connection – Final InterviewAll applications will be treated with strict confidentiality!
Job Application Tips
- Tailor your resume to highlight relevant experience for this position
- Write a compelling cover letter that addresses the specific requirements
- Research the company culture and values before applying
- Prepare examples of your work that demonstrate your skills
- Follow up on your application after a reasonable time period