← В ленту

Портфолио

Deeplay

● Led the development of an enterprise scale ETL system based on Apache Airflow, Kubernetes jobs, cronjobs, and deployments with Data Warehouse, Data Lake based on ClickHouse, Kafka and Minio. ● Implemented a new Big Data ETL pipeline as a team leader, utilizing Flink, pyFlink, Apache Kafka, Google Protobufs, GRPC, and ClickHouse thus consuming a big data stream from a data provider. ● Led a team of 7 Python ETL engineer s as part of a larger team of 30 ETL engineers, including Java, TypeScript, Alerting and BI support teams. Collaborated with the team lead, tech lead/CTO, DBA, and DevOps team. ● Facilitated team growth through performance reviews, skill development sessions , and code reviews. ● Influenced agile development methodologies, including sprint planning, retrospectives, daily meetings, and backlog grooming, to ensure timely delivery of ETL projects. ● Conducted research and evaluated new technologies, resulting in the successful adoption of new versions of Airflow 2, Pandas, and a ClickHouse native tool that enhanced performance by 100% (for complex DAGs). ● Trained 6 new team members on the ETL system, leading to a reduction in development time by 500%. ● Developed and maintained 4 large, scalable parsers for complex data transformation as team leader. ● Implemented and maintained 15 DAGs, resulting in a 5% increase in department revenue per year.

Zuykov and Partners company

● Optimized website performance by implementing caching techniques, resulting in 3x faster website access and meeting Google Page Speed metrics. ● Improved search engine response time by up to 10x through the creation of views logic and partitioning of search result tables into 42 partitions. ● Reduced deployment time from 1 hour to less than 5 minutes by implementing CI/CD for the project. ● Increased readability and decreased troubleshooting time from days to less than 1 hour by integrating ElasticSearch, Logstash, and Kibana (ELK). Developed a backend using Flask, MongoDB, ElasticSearch, and Redis with a remote team of 5 developers. ● Built the backend from scratch, ensuring robust and scalable architecture. ● Implemented Continuous Integration and Continuous Deployment (CI/CD) practices for the project. ● Integrated metrics and monitoring for the backend.

RAFT

1.OpenAPI Generator System Development. 1.Developed a system from scratch using APIs and worker instances to interact with OpenAPI SDK Generator. 2.Reduced developers' wait time for SDK generation from hours to instant on demand generation. 3.Engineered end to end CI/CD CircleCI scripts from scratch f or swift deployment on Google Kubernetes Engine (GKE). 2.Google Dataflow Pipeline and Framework Development. 1.Developed a Google Dataflow pipeline with Apache Beam to parse and store XML CDA files in Google BigQuery for optimized data processing and analysis. 3.Apache Airflow development. 1.Led the ground up development of a new Apache Airflow sub framework for ETL, an alytics and business intelligence departments. 2.Created new Operators, Hooks, Sensors, and Data Quality / Data Alerts tools for Apache Airflow, enhancing code quality and delivery efficiency. 3.Provided comprehensive documentation and training to team members for the new framework. 4.Conducted a large scale migration of 100 Apache Airflow DAGs from MS SQL and ClickHouse to Greenplum to meet the deadline for MS SQL license revocation. 4.AsyncIO microservice for video files parsing. 1.Developed features for a microservi ce project using asyncio, aiohttp, and PostgreSQL for video edition. 2.Contributed to the development and maintenance of the microservice project, ensuring high quality deliverables and smooth operation.

Скиллы

Apache Airflow
Bash
Docker
Google Cloud Platform (GCP)
JavaScript
Google Dataflow
Kubernetes
Apache BEAM
Apache Flink
MySQL
Amazon Web Services (AWS)
Helm
Python
Django
FastAPI
SQL
Celery
Faust
Terraform
CircleCI
ClickHouse
Greenplum
PostgreSQL
Linux
Boto3
Google RPC (gRPC, protobuf)
ElasticSearch
MongoDB

Опыт работы

Senior Software Engineer
с 03.2022 - По настоящий момент |RAFT
Python, SQL, OpenAPI, CI/CD, ETL, XML, Apache Airflow, FastAPI, ElasticSearch, ClickHouse
● OpenAPI Generator System Development. ● Developed a system from scratch using APIs and worker instances to interact with OpenAPI SDK Generator. ● Reduced developers' wait time for SDK generation from hours to instant on demand generation. ● Engineered end to end CI/CD CircleCI scripts from scratch f or swift deployment on Google Kubernetes Engine (GKE). ● Google Dataflow Pipeline and Framework Development. ● Developed a Google Dataflow pipeline with Apache Beam to parse and store XML CDA files in Google BigQuery for optimized data processing and analysis. ● Apache Airflow development. ● Led the ground up development of a new Apache Airflow sub framework for ETL, an alytics and business intelligence departments. ● Created new Operators, Hooks, Sensors, and Data Quality / Data Alerts tools for Apache Airflow, enhancing code quality and delivery efficiency. ● Provided comprehensive documentation and training to team members for the new framework. ● Conducted a large scale migration of 100 Apache Airflow DAGs from MS SQL and ClickHouse to Greenplum to meet the deadline for MS SQL license revocation. ● AsyncIO microservice for video files parsing. ● Developed features for a microservi ce project using asyncio, aiohttp, and PostgreSQL for video edition. ● Contributed to the development and maintenance of the microservice project, ensuring high quality deliverables and smooth operation. ● Developed a FastAPI backend + LLM RAG for efficient retrieval of patent information and related data from pre-cached, augmented datasets in ElasticSearch.
Senior ETL Engineer
01.2020 - 03.2022 |Deeplay
Python, SQL, ETL system, Apache Airflow, Kubernetes, Airflow, Pandas, ClickHouse
● Led the development of an enterprise scale ETL system based on Apache Airflow, Kubernetes jobs, cronjobs, and deployments with Data Warehouse, Data Lake based on ClickHouse, Kafka and Minio. ● Implemented a new Big Data ETL pipeline as a team leader, utilizing Flink, pyFlink, Apache Kafka, Google Protobufs, GRPC, and ClickHouse thus consuming a big data stream from a data provider. ● Led a team of 7 Python ETL engineer s as part of a larger team of 30 ETL engineers, including Java, TypeScript, Alerting and BI support teams. Collaborated with the team lead, tech lead/CTO, DBA, and DevOps team. ● Facilitated team growth through performance reviews, skill development sessions , and code reviews. ● Influenced agile development methodologies, including sprint planning, retrospectives, daily meetings, and backlog grooming, to ensure timely delivery of ETL projects. ● Conducted research and evaluated new technologies, resulting in the successful adoption of new versions of Airflow 2, Pandas, and a ClickHouse native tool that enhanced performance by 100% (for complex DAGs). ● Trained 6 new team members on the ETL system, leading to a reduction in development time by 500%. ● Developed and maintained 4 large, scalable parsers for complex data transformation as team leader. ● Implemented and maintained 15 DAGs, resulting in a 5% increase in department revenue per year.
Middle Software Engineer
01.2018 - 01.2020 |Zuykov and Partners company
Django, DjangoCMS, AngularJS, JQuery, JavaScript, Celery, ElasticSearch, Redis
Upgraded a legacy website using Django, DjangoCMS, AngularJS, JQuery, JavaScript, Celery, ElasticSearch, Redis, and MySQL. ● Optimized website performance by implementing caching techniques, resulting in 3x faster website access and meeting Google Page Speed metrics. ● Improved search engine response time by up to 10x through the creation of views logic and partitioning of search result tables into 42 partitions. ● Reduced deployment time from 1 hour to less than 5 minutes by implementing CI/CD for the project. ● Increased readability and decreased troubleshooting time from days to less than 1 hour by integrating ElasticSearch, Logstash, and Kibana (ELK). Developed a backend using Flask, MongoDB, ElasticSearch, and Redis with a remote team of 5 developers. ●Built the backend from scratch, ensuring robust and scalable architecture. ● Implemented Continuous Integration and Continuous Deployment (CI/CD) practices for the project. ● Integrated metrics and monitoring for the backend.
Software Engineer
01.2017 - 01.2018 |BCMFort company
Python, SQL, Django ORM, AWS, TDD, CircleCI, PostgreSQL, Celery, pyTest, Django Rest Framework, Confluence, Jira, Slack
● Contributed at BCMFort, a Canadian startup with a remote team of 5 developers ● Upgraded Python from version 2.7 to 3.6 and Django from version 1.x to 2.x, ensuring compatibility with modern development practices. ● Utilized pytest, Test-Driven Development (TDD), Amazon Web Services (AWS), and CircleCI to deploy applicationswith increased efficiency and reliability. ● Demonstrated proficiency in both frontend (AngularJS) and backend (Django, Celery, Django Rest Framework, pyTest) development, implementing new features and tests ● Leveraged Django ORM and PostgreSQL to write SQL queries for database operations ● Managed tasks using Jira, Confluence, and Slack in a collaborative and international work environment.
Full Stack Developer
01.2015 - 01.2017 |Upwork.com
Python, MySQL, MariaDB, PostgreSQL, Redis
● Designed, implemented and maintained web application architecture from scratch using Python and Django framework. ● Developed and maintained SQL queries using MySQL, MariaDB, PostgreSQL, and Redis ● Utilized DevOps/SDLC application cycle for various clients, managing Git, GitHub, Docker, AWS, and other VPSs. ● Developed and implemented scrapers for website parsing.
Senior Systems Engineer
01.2005 - 01.2015 |Ulan Ude Aviation Plant
Pascal, C++, Assembler, PLCs, COM, LPT
● Created SCADA projects to work with various Programmable Logic Controllers (PLCs). ● Developed applications to connect to industrial machines through COM and LPT ports using Pascal, C++, and Assembler. ● Wrote a VESA video application for managing industrial machines through the ISA bus using Pascal and Assembler. ● Maintained and upgraded industrial machines and PLCs.

Образование

Computer Science and Engineering
East Siberia State University

Дополнительное образование

Unigraphics
12.2010 - 12.2010
Unigraphics

Языки

РусскийРоднойАнглийскийВыше среднего