Data Platform Engineer
Remote / Prague /
Hybrid
Lokation: Prague, HO 3x
Language: English required, Czech
Level: Senior
Form of cooperation: Contraktor
Start date: asap
Allocation: Full-Time
Allocation length: Long term
Your impact as a Data Platform Engineer:
- The GSDAP team consists of 7 Data (Platform) Engineers managing couple of Data Platforms specific for each support functions (HR, Facility management, IT, Non-Financial Risk). Within the team, you will play a key role in constructing and automating the data platform, as well as developing functional data components.
- To set the foundation of the materialization of data ambitions, you must be self-directed and comfortable with the complexity of supporting the data needs of multiple teams, systems and products.
Technologies you will work with:
- Our infrastructure runs on MS Azure cloud and we use Terraform as IaC.
- We use Azure Data Factory for pipelines and orchestration and Databricks for processing and transforming data.
Your role:
- Maintain and further enhancing the Data Platforms
- Understand the needs of your various customers from different units and translate them to modular solutions which are usable by the vast majority
- Enable support functions teams to run successful data and AI projects by providing technical guidance
- Brainstorm and design solutions to optimize our data platform solution architecture
- Share knowledge in our data engineering community, pair program and collaborate on code together with other data engineers
- Participating in code reviews and ensuring adherence to development standards
- Staying up-to-date with the latest data platform technologies and industry trends
Your profile
- You have 5+ years of relevant software, data and/or platform engineering experience, building platforms which are modular, testable, scalable and easily consumable.
- Have 3+ years hands-on experience on one or more cloud services (Azure/AWS) like: ADF, Data-lake, Delta-lake, Databricks, Key Vaults, BigQuery, Cloud Dataflow, Datapipeline, , etc.
- Experience with Infra as a Code (Terraform, BICEP)
- Experience with Data as Code; version control, small and regular commits, unit tests, CI/CD, packaging, branching etc.
- Demonstrated programming and debugging experience with Python/PySpark, SQL
- Preferably experience with open source projects run with a “build once, deploy often” mindset & Experience or interest in Domain Driven Design.
You are:
- Self-directed
- Both pragmatic and methodical with a strong software engineering mindset, and metadata-driven data solution design
- Actively helping the team to achieve flow and achieve the sprint goal.
- Collaborative and proactive about working in our inner source community
- Passionate about automation looking for automating our data products to the highest scale.