09/06/2022 updated


100 % available
Big Data & Cloud Architect
Poland
Poland
B.Sc. Computer Science, Warsaw UniversityExperienced Big Data Architect working as a freelancer since 2015 for large European customers including Schaeffler, Nordea, BNP Paribas, E.On, Adidas, Essity, DXC, Allianz.
Extremely hands-on & cross-functional within Big Data, Cloud (Azure & AWS), Data Engineering, DevOps & MLOps space. Built and led teams of developers. Experience with greenfield projects, architecture design & delivery.
Certified in Microsoft Azure & Spark. Programming languages: Python, Java, Scala. More details in resume. References available upon request.
Extremely hands-on & cross-functional within Big Data, Cloud (Azure & AWS), Data Engineering, DevOps & MLOps space. Built and led teams of developers. Experience with greenfield projects, architecture design & delivery.
Certified in Microsoft Azure & Spark. Programming languages: Python, Java, Scala. More details in resume. References available upon request.
Languages
GermanGoodEnglishFluentPolishNative speaker
Project history
{ responsible for transforming harmonized data layer into consumption layer for Fleet project;
{ responsible for data mapping, ETL and Fleet portal back-end;
{ technologies used - PostgreSQL, Azure (AKS, Key Vault, Blob Storage), Kafka.
{ responsible for data mapping, ETL and Fleet portal back-end;
{ technologies used - PostgreSQL, Azure (AKS, Key Vault, Blob Storage), Kafka.
{ troubleshooted performance problems with Azure Databricks notebooks and tuned cluster settings;
{ implemented historical data preprocessing procedures using Azure Functions that transformed over 6TB of data;
{ introduced Databricks SQL Analytics as a reporting layer for BI;
{ technologies used - Azure (Data Lake Storage, Data Factory, Key Vault, Databricks, Functions).
{ implemented historical data preprocessing procedures using Azure Functions that transformed over 6TB of data;
{ introduced Databricks SQL Analytics as a reporting layer for BI;
{ technologies used - Azure (Data Lake Storage, Data Factory, Key Vault, Databricks, Functions).
{ part of a team designing and implementing a data lake based on Azure to ingest data from production plants, reporting
systems and other internal sources;
{ industrialized Python and R machine learning models and deployed to Azure Databricks (Spark 2.4), Data Science VM,
ML Workspace, AKS, Azure Batch;
{ introduced guidelines for data scientists working in Python and R, set up Jupyter notebooks, created dockerized
flask+scikit-learn and R Shiny environments;
{ created CI/CD Jenkins pipelines to AKS, Databricks and Azure Batch;
{ implemented near real-time messaging ETL and CDC pipelines in Data Factory;
{ technologies used - Azure (Data Lake Storage, Data Factory, Batch, AKS, ACR, Key Vault, SQL DW, Databricks,
Machine Learning Workspace, Data Science VM, ARM, IoT Hub, VNet), NATS, Cloudbreak, Python 2.7, 3.6, Jupyter
notebooks.
systems and other internal sources;
{ industrialized Python and R machine learning models and deployed to Azure Databricks (Spark 2.4), Data Science VM,
ML Workspace, AKS, Azure Batch;
{ introduced guidelines for data scientists working in Python and R, set up Jupyter notebooks, created dockerized
flask+scikit-learn and R Shiny environments;
{ created CI/CD Jenkins pipelines to AKS, Databricks and Azure Batch;
{ implemented near real-time messaging ETL and CDC pipelines in Data Factory;
{ technologies used - Azure (Data Lake Storage, Data Factory, Batch, AKS, ACR, Key Vault, SQL DW, Databricks,
Machine Learning Workspace, Data Science VM, ARM, IoT Hub, VNet), NATS, Cloudbreak, Python 2.7, 3.6, Jupyter
notebooks.