
With over 14 years of specialized experience, I am an adept Data Engineer proficient in designing, developing, maintaining, and enhancing data processing solutions, including data engineering, data modeling, and ETL/ELT processes. My experience extends to detailed performance analysis and optimization, along with effectively diagnosing and resolving performance issues. I possess a proven proficiency in rapidly mastering new tools and technologies, leveraging them to meet specific business objectives.
Additionally, I bring experience in curriculum development and teaching, having authored and instructed courses on PostgreSQL, enhancing the knowledge and skills of professionals in data technology in current company.
Customer: The large multinational pharmaceutical and biotechnology company
Implementation of DWH and financial reporting for a large multinational pharmaceutical and biotechnology company.
Responsibilities:
Tools and Technologies: Databricks, PySpark, Azure Cloud, Azure Data Factory, Azure DevOps, Delta Tables, Git, Wiki
Customer: Multinational cosmetics company
The Data Science/Machine Learning project involved refactoring the existing codebase written by data scientists, followed by the automation of all associated processes. This also included the preparation and structuring of data for efficient utilization with data models.
Responsibilities:
Tools and Technologies: Python, PySpark, Databricks, Delta Tables, Azure Data Factory, Azure SQL Database, Azure Blob Storage, Azure DevOps, Git
Developed a 27-hour training program tailored for an internal department focused on employee skill enhancement. The course encompasses key aspects and modules pertinent to PostgreSQL.
The course covers essential topics and modules for PostgreSQL developers including the logical and physical architecture of PostgreSQL, Transactions & ACID properties, PostgreSQL security protocols, indexing, partitioning, utilization of EXPLAIN and EXPLAIN ANALYZE commands, and techniques for PostgreSQL performance tuning, among others.
Responsibilities:
Customer: The large building materials company
Building a data warehouse based on SAP HANA with the purpose of using data for the customer's reporting needs.
Responsibilities:
Tools and Technologies: Azure Synapse, Azure Devops, Azure Data Factory, MS SQL Server, Python, Snowflake, SAP, Git
Customer: The multinational consumer goods corporation
A project to analyze, integrate, model, process and refine data from various sources within the customer's ecosystem and external sources to support data-driven decision making processes
Responsibilities:
The data migration project for a group of insurance companies consolidating their operations. The requirement was to move source data from Oracle to a new AWS cloud where it would be stored in Postgres, managed and processed based on roles.
Responsibilities:
Tools and Technologies: PostgreSQL, DBeaver, PyCharm, Jenkins, Python, SQL, PL/pgSQL
Implementation of data warehouse based on Data Vault. Data sources are various desktop applications, FHIR specifications, medical analysis information storage systems, exchanging electronic health records APIs and MS SQL Server, MongoDB, Postgres.
Responsibilities:
Integrated automated information system of education. Single information space of the region for implementation and management of public services in the field of education.
Responsibilities:
Web-application system for automation of technology processes about power consumption by large enterprises.
Responsibilities:
Customer: Rosatom
Responsibilities:
Responsibilities:
Data Modeling
Data Warehousing
ETL development
Data Pipeline Design
Data Migration
Metadata Management
Databricks
Apache Airflow
PySpark
PostgreSQL
Snowflake
Clickhouse
MS SQL Server
Azure Cloud
Azure Data Factory
Azure Synapse
Azure DevOps
FAIR
Team Leadership