December 5
AWS
Azure
Cloud
Distributed Systems
Docker
ETL
Google Cloud Platform
Informatica
Kubernetes
Numpy
Pandas
PySpark
Python
Spark
SQL
Go
• responsible for at-scale infrastructure design, build and deployment with a focus on distributed systems • building and maintaining architecture patterns for data processing, workflow definitions, and system to system integrations using Big Data and Cloud technologies • evaluating and translating technical design to workable technical solutions/code and technical specifications at par with industry standards • driving creation of re-usable artifacts • establishing scalable, efficient, automated processes for data analysis, data model development, validation, and implementation • working closely with analysts/data scientists to understand impact to the downstream data models • writing efficient and well-organized software to ship products in an iterative, continual release environment • contributing and promoting good software engineering practices across the team • communicating clearly and effectively to technical and non-technical audiences • defining data retention policies • monitoring performance and advising any necessary infrastructure changes.
• 2+ years’ experience with Azure (Data Factory, Databricks) • 3+ years’ experience with data engineering or backend/fullstack software development • solid SQL and Git skills • Python scripting proficiency • experience with data transformation tools - Databricks and Spark • experience in structuring and modelling data in both relational and non-relational forms • ability to elaborate and propose relational/non-relational approach • normalization / denormalization and data warehousing concepts (star, snowflake schemas) • good verbal and written communication skills in English • Work from the European Union region and a work permit are required • Nice to have: experience with CI/CD tooling (GitHub, Azure DevOps, Harness etc.) • data manipulation libraries (such as Pandas, NumPy, PySpark) • experience with Azure Event Hubs, Azure Blob Storage, Azure Synapse, Spark Streaming • experience with data modelling tools, preferably DBT • experience with Enterprise Data Warehouse solutions, preferably Snowflake • familiarity with ETL tools (such as Informatica, Talend, Datastage, Stitch, Fivetran etc) • experience in containerization and orchestration (Docker, Kubernetes etc) • cloud (Azure, AWS, GCP) certification.
Apply NowNovember 21
Join a technology company as Data Engineer, optimizing data management strategies in finance. Work with complex datasets and improve data infrastructure.
November 7
Enhance financial operations through data management and integration strategies.
October 24
Data Engineer at Tecknoworks, optimizing data pipelines for client solutions.
October 17
Data Engineer for building data pipelines at Sales Consulting.