We are a global digital business services organization with a mission to improve the way people and companies work. Forever. Our Consulting, Delivery and Learning teams design and deliver transformational digital products and experiences that add value to our clients’ businesses and to their customers’ lives. Every day, across the world, our teams are pioneering faster, better ways to bring our client’s most exciting ideas to life. We are looking for a seasoned Senior Data Engineer to help us shape Emergn’s exciting future and play an important role in our growth. We want you to:
Design, build, and maintain scalable data pipelines and workflows using Microsoft Fabric, including Data Factory, Lakehouse, and Synapse Pipelines. Use Apache Spark in Microsoft Fabric Notebooks for large-scale data processing, cleansing, and transformation tasks. Develop efficient SQL-based solutions for data modeling, data warehousing, and analytics layers. Leverage Python and PySpark to automate data flows, integrate sources, and apply advanced data logic. Collaborate with analysts, engineers, and stakeholders to deliver clean, trustworthy datasets to reporting and ML pipelines. Implement and monitor ETL/ELT processes for batch and near-real-time use cases across diverse data sources.
Optimize performance of queries and transformations in Spark and SQL environments. Support Lakehouse Architecture within Microsoft Fabric and ensure best practices in data lake management. Assist in establishing data quality, data lineage, and governance processes across the data stack. Act as a subject matter expert on data workflows within the Microsoft ecosystem, helping to guide best practices across teams. This job might be for you if you have: Bachelor’s or Master’s degree in Computer Science, Data Engineering, or a related field. Hands-on experience with Microsoft Fabric components including Spark Notebooks, Data Factory, and Synapse.
Strong understanding of Apache Spark (especially via PySpark) for distributed data processing. Proficiency in SQL for data manipulation and optimization. Solid Python skills for scripting, automation, and transformation logic. Experience with cloud-native data solutions—preferably on Microsoft Azure. Understanding of data warehouse design, dimensional modeling, and Lakehouse patterns. Familiarity with CI/CD and version control tools (e. g. , Git, Azure DevOps). Comfortable working in agile, iterative data development cycles. Excellent communication and stakeholder collaboration skills. Nice to Have: Familiarity with OneLake architecture and Delta Lake implementation in Fabric.
Knowledge of Power BI data modeling and how backend data impacts reports. Experience with streaming data ingestion (e. g. , Azure Event Hubs, Kafka, Fabric Real-Time Analytics). Exposure to notebook-based development workflows in Jupyter or Databricks. Awareness of data privacy, security best practices, and compliance (e. g. , GDPR, DLP tools). Previous experience with other Spark platforms like Databricks or HDInsight is a plus. What we offer:
Work within a dynamic international team of expertsExcellent opportunity for personal and professional developmentComfortable office and flexible working hoursAbility to work with modern technologiesHuge catalogue of educational programs, the possibility of training and certification at the expense of company20 working days of vacation per yearLife and disability insuranceHealth insuranceBig free parkingReady to make an impact?Join us today!
Customize your resume to highlight skills and experiences relevant to this specific position.
Learn about the company's mission, values, products, and recent news before your interview.
Ensure your LinkedIn profile is complete, professional, and matches your resume information.
Prepare thoughtful questions to ask about team dynamics, growth opportunities, and company culture.