Data Platform Engineer

Remote / Prague /
Hybrid
Lokation: Prague, HO 3x
Language: English required, Czech

Level: Senior
Form of cooperation: Contraktor
Start date: asap
Allocation: Full-Time
Allocation length: Long term

Your impact as a Data Platform Engineer:

  • The GSDAP team consists of 7 Data (Platform) Engineers managing couple of Data Platforms specific for each support functions (HR, Facility management, IT, Non-Financial Risk). Within the team, you will play a key role in constructing and automating the data platform, as well as developing functional data components. 
  • To set the foundation of the materialization of data ambitions, you must be self-directed and comfortable with the complexity of supporting the data needs of multiple teams, systems and products. 

Technologies you will work with:

  • Our infrastructure runs on MS Azure cloud and we use Terraform as IaC. 
  • We use Azure Data Factory for pipelines and orchestration and Databricks for processing and transforming data.

Your role:

  • Maintain and further enhancing the Data Platforms 
  • Understand the needs of your various customers from different units and translate them to modular solutions which are usable by the vast majority
  • Enable support functions teams to run successful data and AI projects by providing technical guidance 
  • Brainstorm and design solutions to optimize our data platform solution architecture
  • Share knowledge in our data engineering community, pair program and collaborate on code together with other data engineers
  • Participating in code reviews and ensuring adherence to development standards
  • Staying up-to-date with the latest data platform technologies and industry trends

Your profile

  • You have 5+ years of relevant software, data and/or platform engineering experience, building platforms which are modular, testable, scalable and easily consumable.
  • Have 3+ years hands-on experience on one or more cloud services (Azure/AWS) like: ADF, Data-lake, Delta-lake, Databricks, Key Vaults, BigQuery, Cloud Dataflow, Datapipeline, , etc.
  • Experience with Infra as a Code (Terraform, BICEP)
  • Experience with Data as Code; version control, small and regular commits, unit tests, CI/CD, packaging, branching etc.
  • Demonstrated programming and debugging experience with Python/PySpark, SQL
  • Preferably experience with open source projects run with a “build once, deploy often” mindset & Experience or interest in Domain Driven Design.

You are:

  • Self-directed
  • Both pragmatic and methodical with a strong software engineering mindset, and metadata-driven data solution design
  • Actively helping the team to achieve flow and achieve the sprint goal.
  • Collaborative and proactive about working in our inner source community
  • Passionate about automation looking for automating our data products to the highest scale.