Azure Data Engineer

December 5

Apply Now
Logo of Xebia Poland

Xebia Poland

Software Development • Agile Development • DevOps • Scrum • Mobile Applications

1001 - 5000 employees

Founded 2001

🤖 Artificial Intelligence

☁️ SaaS

Description

• responsible for at-scale infrastructure design, build and deployment with a focus on distributed systems • building and maintaining architecture patterns for data processing, workflow definitions, and system to system integrations using Big Data and Cloud technologies • evaluating and translating technical design to workable technical solutions/code and technical specifications at par with industry standards • driving creation of re-usable artifacts • establishing scalable, efficient, automated processes for data analysis, data model development, validation, and implementation • working closely with analysts/data scientists to understand impact to the downstream data models • writing efficient and well-organized software to ship products in an iterative, continual release environment • contributing and promoting good software engineering practices across the team • communicating clearly and effectively to technical and non-technical audiences • defining data retention policies • monitoring performance and advising any necessary infrastructure changes.

Requirements

• 2+ years’ experience with Azure (Data Factory, Databricks) • 3+ years’ experience with data engineering or backend/fullstack software development • solid SQL and Git skills • Python scripting proficiency • experience with data transformation tools - Databricks and Spark • experience in structuring and modelling data in both relational and non-relational forms • ability to elaborate and propose relational/non-relational approach • normalization / denormalization and data warehousing concepts (star, snowflake schemas) • good verbal and written communication skills in English • Work from the European Union region and a work permit are required • Nice to have: experience with CI/CD tooling (GitHub, Azure DevOps, Harness etc.) • data manipulation libraries (such as Pandas, NumPy, PySpark) • experience with Azure Event Hubs, Azure Blob Storage, Azure Synapse, Spark Streaming • experience with data modelling tools, preferably DBT • experience with Enterprise Data Warehouse solutions, preferably Snowflake • familiarity with ETL tools (such as Informatica, Talend, Datastage, Stitch, Fivetran etc) • experience in containerization and orchestration (Docker, Kubernetes etc) • cloud (Azure, AWS, GCP) certification.

Apply Now

Similar Jobs

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or lior@remoterocketship.com