Job Description

Role: Data Engineer (PySpark & Python Focus)Location: Kraków (Hybrid – 3 days/week onsite)Project Duration: Long TermClient: ConfidentialExperience:5–10 years of hands-on experience in data analytics as a Data Engineer.Strong understanding of ETL, Data Quality (DQ), Data Modelling (DM), reject & recycle concepts.Experience working in big data environments using Hadoop clusters or cloud platforms.Technical Skills:Extensive hands-on experience building data pipelines using Spark, especially PySpark.Minimum 3 years of practical coding experience in PySpark using RDDs, DataFrames, and Datasets (not Spark SQL).Proven experience developing multiple Spark applications processing large volumes of data.Strong skills in performance tuning and complex transformations (groups, window functions, etc.).Participation in PySpark hackathons is a strong plus.Ability to confidently write PySpark code during interviews.Strong proficiency in Python:Clean, efficient, reusable codeBug fixing & troubleshootingCreating scripts/tools for automationNote: This role requires advanced Python coding skills. Candidates will be required to demonstrate their Python skills in the interview.Familiarity with Databricks, Airflow, Azure is a plus (not mandatory).Core focus remains Spark, PySpark, and data engineering.Additional Skills:Strong problem-solving & analytical abilitiesAbility to translate business requirements into technical solutionsGood communication & teamwork skillsExperience with CI/CD pipelinesExperience working in Agile environments

Job Application Tips

  • Tailor your resume to highlight relevant experience for this position
  • Write a compelling cover letter that addresses the specific requirements
  • Research the company culture and values before applying
  • Prepare examples of your work that demonstrate your skills
  • Follow up on your application after a reasonable time period

You May Also Be Interested In