Summary
Overview
Work History
Education
Skills
Timeline
Generic

Barannik Raisa

Senior Data Engineer
Antalya

Summary

With over 14 years of specialized experience, I am an adept Data Engineer proficient in designing, developing, maintaining, and enhancing data processing solutions, including data engineering, data modeling, and ETL/ELT processes. My experience extends to detailed performance analysis and optimization, along with effectively diagnosing and resolving performance issues. I possess a proven proficiency in rapidly mastering new tools and technologies, leveraging them to meet specific business objectives.
Additionally, I bring experience in curriculum development and teaching, having authored and instructed courses on PostgreSQL, enhancing the knowledge and skills of professionals in data technology in current company.

Overview

15
15
years of professional experience
5
5
years of post-secondary education

Work History

Senior Data Engineer

Epam Systems
7 2023 - Current

Customer: The large multinational pharmaceutical and biotechnology company

Implementation of DWH and financial reporting for a large multinational pharmaceutical and biotechnology company.

Responsibilities:

  • Building an Databricks application architecture for financial reporting(Building bronze/silver/gold layers and sub-layers of data processing storing historical data)
  • Implementation of a PySpark Framework for loading data from various sources and delivering data to the target layer in SnowFlake which includes ETL/ELT data processing, data harmonization
  • Creating the solution for error handling and data quality
  • Creating monitoring of the solution of loading processes
  • Implement pipeline process in Azure Data Factory
  • Implementing a Role Model in Snowflake
  • Team Leadership 2 juniors, 1 meddle

Tools and Technologies: Databricks, PySpark, Azure Cloud, Azure Data Factory, Azure DevOps, Delta Tables, Git, Wiki

Senior Data Engineer

Epam Systems
01.2023 - 07.2023

Customer: Multinational cosmetics company

The Data Science/Machine Learning project involved refactoring the existing codebase written by data scientists, followed by the automation of all associated processes. This also included the preparation and structuring of data for efficient utilization with data models.

Responsibilities:

  • Responsible for refactoring the code of the Data Scientists team for process automation
  • Optimization ETL/ELT processes, data harmonization, analyze data quality
  • Implement pipeline process in Databricks use PySpark
  • Develop ETL processes using PySpark and Spark-SQL in Databricks for data extraction, transformation, and aggregation from multiple file formats
  • Hands-on experience on developing Python scripts for automation
  • Writing Python code to work with external sources such as Google Trends. Preparing data for the needs of the data scientist team
  • Implementation of a Framework for loading data from various sources, which includes ETL data processing, error handling, and monitoring of loading processes

Tools and Technologies: Python, PySpark, Databricks, Delta Tables, Azure Data Factory, Azure SQL Database, Azure Blob Storage, Azure DevOps, Git

Author of the course on Postgres, lecturer of the course

Epam Systems
11.2022 - 03.2023

Developed a 27-hour training program tailored for an internal department focused on employee skill enhancement. The course encompasses key aspects and modules pertinent to PostgreSQL.

The course covers essential topics and modules for PostgreSQL developers including the logical and physical architecture of PostgreSQL, Transactions & ACID properties, PostgreSQL security protocols, indexing, partitioning, utilization of EXPLAIN and EXPLAIN ANALYZE commands, and techniques for PostgreSQL performance tuning, among others.

Responsibilities:

  • Developing a comprehensive course tailored to client specifications, including the preparation of lecture materials
  • Crafting engaging presentations to facilitate learning
  • Delivering a series of educational lectures aimed at enhancing the qualifications of students and employees
  • Creating homework assignments and conducting interactive classroom exercises to reinforce learning concepts.


Senior Data Engineer

Epam Systems
5 2021 - 11.2022

Customer: The large building materials company

Building a data warehouse based on SAP HANA with the purpose of using data for the customer's reporting needs.

Responsibilities:

  • I spearheaded the implementation of an operational database designed to encapsulate information essential for managing pipelines in Azure Data Factory. This database archives data loading logs, configuration settings, metadata, and pipeline dependencies. Additionally, it incorporates an integrated exception storage system.
  • Orchestrated the deployment of a pipeline launch solution within the Data Factory to automate processes and gather monitoring data.
  • Designed and executed solutions that facilitated both ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform) processes in Azure Synapse.
  • Managed the data ingestion from Snowflake, involving complex joins across dozens of tables and views, with data preparation handled via stored procedures and ETL tools.
  • Utilized MS SQL Server as the archival source for historical data, requiring logical transformations across more than 50 tables for integration into a new Data Warehouse storage system.
  • Extracted daily data from hundreds of SAP tables and views.
  • Developed Python scripts to verify the integrity of the project architecture and master data management.

Tools and Technologies: Azure Synapse, Azure Devops, Azure Data Factory, MS SQL Server, Python, Snowflake, SAP, Git

Senior Data Engineer

Epam Systems
09.2020 - 06.2021

Customer: The multinational consumer goods corporation

A project to analyze, integrate, model, process and refine data from various sources within the customer's ecosystem and external sources to support data-driven decision making processes

Responsibilities:

  • Design and implement orchestration DB( design data model and create DDL scripts for Orchestration database)
  • Implement procedures and functions to solve Orchestration database
  • Implementation of a FrameWork for loading data from various sources, which includes ETL data processing, error handling, and monitoring of loading processes
  • Implement pipeline process in Data Factory Implement pipeline process in DataBricks(PySpark)
  • Optimization ETL/ELT processes, data harmonization, analyze data quality

Senior Software Engineer

Epam Systems
12.2019 - 09.2020

The data migration project for a group of insurance companies consolidating their operations. The requirement was to move source data from Oracle to a new AWS cloud where it would be stored in Postgres, managed and processed based on roles.

Responsibilities:

  • Migration planning and implementation
  • Migration according to the new business logic of large tables from 39 GB to 2 TB
  • Generating test data using Python for PostgreSQL in AWS Developing and maintaining automated pipelines on Jenkins
  • Uploading data to the target server for insurance group of companies using python

Tools and Technologies: PostgreSQL, DBeaver, PyCharm, Jenkins, Python, SQL, PL/pgSQL


DWH developer

Netrika
07.2017 - 09.2019

Implementation of data warehouse based on Data Vault. Data sources are various desktop applications, FHIR specifications, medical analysis information storage systems, exchanging electronic health records APIs and MS SQL Server, MongoDB, Postgres.

Responsibilities:

  • Design development from start to finish of DWH based on Data Vault
  • I implemented solutions enabling the process (ETL and ELT) of medical data from various sources and databases (PostgreSQL, MS SQL Server, MongoDB, Rest, use FAIR specification)
  • Furthermore, I enhanced data loading efficiencies by fine-tuning dependencies pipelines Apache Airflow, along with creating specialized plugins for Airflow
  • Worked as a DevOps engineer for some time oh the project
  • Tools and Technologies: PostgreSQL, MS SQL Server, MongoDB, ClickHouse, Python, Airflow, Superset, Docker, Data Vault, REST API, Ubuntu, FAIR specification

Senior PHP Backend Developer, DWH developer

Netrika
07.2015 - 08.2017

Integrated automated information system of education. Single information space of the region for implementation and management of public services in the field of education.

Responsibilities:

  • Developing a Data Warehouse (DWH) customed for an automated system that collects and stores data regarding attendees in educational institutions such as kindergartens, schools, and universities.
  • Developing and maintaining multiple ETL processes to collect, process, and analyze data from diverse sources including external databases (MySQL and PostgreSQL), SOAP packages, REST APIs, and XML data formats.
  • Ensuring data integrity and quality by implementing various optimizations and quality control measures to validate information accuracy and reliability.
  • Creating web applications using PHP (Zend and Phalcon). These applications served as primary interfaces for users to submit requests and access information.

Full-stack web developer

Center for Energy Saving Automation
02.2012 - 06.2015

Web-application system for automation of technology processes about power consumption by large enterprises.

Responsibilities:

  • Design, implementation, and refactoring of support system for automation of technology processes (power consumption management) for large enterprises
  • Developed and maintained / enhanced functionality enabling to collect, store, process and analyze operational information about the resources consumption

Software Engineer

CFT (Center of Financial Technologies)
12.2009 - 02.2012

Customer: Rosatom

Responsibilities:

  • Designed, implemented and supported software solutions for automated bookkeeping and controlling processes
  • Implemented and supported/customized existing products for automating of day-to-day operations of banks
  • Prepared financial report templates (optimized existing reports, created and optimized request forms, developed new ones)
  • Owned product onboarding and consulted end-users on implementation and usage

Software Engineer

Novosibirsk State Technical University
Novosibirsk
11.2006 - 8 2009

Responsibilities:

  • Designed and implemented databases based on Oracle RDBMS
  • Developed and optimized queries, functions, and stored procedures
  • Designed and implemented web–applications based on PHP, supported existing web applications, designed, developed, and integrated new modules
  • Covered all development deliverables by technical documentation

Education

Specialist Degree - Applied Mathematics And Information Science

Novosibirsk State Technical University
Novosibirsk
09.2006 - 07.2007

Bachelor of Science - Applied Mathematics And Information Science

Novosibirsk State Technical University
Novosibirsk
09.2002 - 07.2006

Skills

Data Modeling

Data Warehousing

ETL development

Data Pipeline Design

Data Migration

Metadata Management

Databricks

Apache Airflow

PySpark

PostgreSQL

Snowflake

Clickhouse

MS SQL Server

Azure Cloud

Azure Data Factory

Azure Synapse

Azure DevOps

FAIR

Team Leadership

Timeline

Senior Data Engineer

Epam Systems
01.2023 - 07.2023

Author of the course on Postgres, lecturer of the course

Epam Systems
11.2022 - 03.2023

Senior Data Engineer

Epam Systems
09.2020 - 06.2021

Senior Software Engineer

Epam Systems
12.2019 - 09.2020

DWH developer

Netrika
07.2017 - 09.2019

Senior PHP Backend Developer, DWH developer

Netrika
07.2015 - 08.2017

Full-stack web developer

Center for Energy Saving Automation
02.2012 - 06.2015

Software Engineer

CFT (Center of Financial Technologies)
12.2009 - 02.2012

Software Engineer

Novosibirsk State Technical University
11.2006 - 8 2009

Specialist Degree - Applied Mathematics And Information Science

Novosibirsk State Technical University
09.2006 - 07.2007

Bachelor of Science - Applied Mathematics And Information Science

Novosibirsk State Technical University
09.2002 - 07.2006

Senior Data Engineer

Epam Systems
7 2023 - Current

Senior Data Engineer

Epam Systems
5 2021 - 11.2022
Barannik RaisaSenior Data Engineer