Egor Makhov, Developer in Como, Province of Como, Italy
Egor is available for hire
Hire Egor

Egor Makhov

Verified Expert  in Engineering

Data Engineer and Developer

Como, Province of Como, Italy

Toptal member since August 6, 2024

Bio

Egor is a data engineer with extensive experience solving business problems through analytics and big data solutions. He specializes in data modeling, building scalable data architectures, and creating intuitive analytical platforms. His expertise includes transitioning legacy ETL jobs into streaming pipelines and building modern data lakes and warehouses using cloud technologies. Egor has also designed and developed an open-source data quality framework for batch processing and streaming data.

Portfolio

Raiffeisen Bank Russia
Data Lakes, Data Lakehouse, Databricks, Data Architecture, Big Data...
Oliver James Associates
Apache Kafka, Amazon S3 (AWS S3), Amazon Athena, Apache Spark, Spark Streaming...
Agile Lab
Python, Python 3, OpenCV, OpenVINO, Deep Learning, Machine Learning...

Experience

Availability

Full-time

Preferred Environment

Linux, Docker, IntelliJ IDEA, Slack, Windows, Mattermost

The most amazing...

...thing I've developed is an open-source data quality framework bundled with end-to-end no-code applications for batch processing and streaming data.

Work Experience

Principal Data Engineer

2021 - PRESENT
Raiffeisen Bank Russia
  • Built low-latency data marts across all CIB banking products, enhancing data accessibility and performance.
  • Created a fast, scalable, and user-friendly computing platform for data analytics and data operations.
  • Developed a universal data quality framework for big data processing, adopted across the entire bank.
  • Designed a GPU cluster for the rapid development of LLM-based applications.
Technologies: Data Lakes, Data Lakehouse, Databricks, Data Architecture, Big Data, Big Data Architecture, Data Warehousing, Data Warehouse Design, Databases, PostgreSQL, Spark, Apache Spark, Python, Scala, Debezium, Kafka Connect, Apache Kafka, Hadoop, HDFS, Apache Hive, Trino, Large Language Models (LLMs), Kubeflow, DataOps, Kubernetes, Docker, Docker Compose, Software Development, PySpark

Data Engineering Consultant

2023 - 2023
Oliver James Associates
  • Migrated data pipelines of a major European stock exchange to a new market protocol, ensuring seamless transitions and compliance with updated standards.
  • Optimized existing architecture to improve performance and efficiency, resulting in faster data processing and reduced operational costs.
  • Introduced data quality checks to ensure completeness and consistency of market events.
Technologies: Apache Kafka, Amazon S3 (AWS S3), Amazon Athena, Apache Spark, Spark Streaming, Delta Lake, Stock Exchange, Stock Market, Data Quality, Data Governance, Streaming Data, Scala, PySpark

Machine Learning Engineer

2019 - 2020
Agile Lab
  • Managed dataset collection and training for personal protective equipment (PPE) detection machine learning model.
  • Developed a FLANN-based pattern detection module for access control and personnel identification.
  • Led a team of three developers during the review and refactoring of a feature extraction application for a corporate banking platform.
Technologies: Python, Python 3, OpenCV, OpenVINO, Deep Learning, Machine Learning, Computer Vision, Apache Spark, Scala, PySpark

Big Data Engineer

2017 - 2019
Agile Lab
  • Developed and tested multiple ETL applications at UniCredit Services as part of a 4-member team.
  • Designed, developed, and integrated the data quality core module of an open-source big data analytical framework.
  • Built multiple pipelines during the IoT platform development for Vodafone as part of a 20-member team.
Technologies: Scala, Apache Spark, Hadoop, HDFS, Apache Impala, Akka, Cloudera

Software Engineering Intern

2016 - 2016
I&b | Innova Et Bella
  • Improved the BrandMemo platform's recommendation engine through advanced machine-learning integration using Python.
  • Reduced application load times by 30% and boosted user retention by 20% through front-end optimization and UX improvements.
  • Increased speeds by 70% by optimizing the internal data model and refactoring the most common SQL queries.
Technologies: Python 3, SciPy, Swift, Vagrant, Mobile App Development, Data Science, Data Analytics, Back-end Development, Front-end Development, JavaScript, Pandas, NumPy, XGBoost, MySQL

Checkita Data Quality Framework

http://www.checkita.org/
An open-source data quality framework bundled with end-to-end no-code applications for batch and streaming data. I created Checkita while I was at Agile Lab and seamlessly integrated it for clients. Upon joining Raiffeisen Bank, I further developed and elevated the framework to a central cornerstone of the bank's data quality platform.
2011 - 2017

Master's Degree in Mathematics and Computer Science

MIREA – Russian Technological University - Moscow, Russia

AUGUST 2020 - AUGUST 2023

Google Cloud Certified Associate Cloud Engineer

Google Cloud

JUNE 2020 - JUNE 2023

AWS Certified Machine Learning – Specialty

Amazon Web Services

Libraries/APIs

Spark Streaming, PySpark, OpenCV, SciPy, Pandas, NumPy, XGBoost

Tools

IntelliJ IDEA, Apache Airflow, Docker Compose, Slack, Mattermost, AWS IAM, Amazon SageMaker, Kafka Connect, OpenVINO, Apache Impala, Cloudera, Amazon Athena, Vagrant

Languages

SQL, Scala, Python, Python 3, Java, Swift, JavaScript

Frameworks

Apache Spark, Spark, Hadoop, Trino, Flask, Data Lakehouse, Akka

Paradigms

ETL

Platforms

Apache Kafka, Kubeflow, Kubernetes, Amazon Web Services (AWS), Linux, Docker, Windows, Google Cloud Platform (GCP), Databricks, Debezium, Azure

Storage

Databases, Data Lakes, PostgreSQL, HDFS, Apache Hive, NoSQL, JSON, Amazon S3 (AWS S3), MySQL

Other

Software Development, Big Data, Data Engineering, Data Architecture, Data Warehousing, DataOps, Cryptography, Machine Learning, Data Science, Mathematics, Big Data Architecture, Data Mining, APIs, Data Governance, Data Quality, Data Quality Analysis, Cloud Computing, Data Warehouse Design, Large Language Models (LLMs), Deep Learning, Computer Vision, Delta Lake, Stock Exchange, Stock Market, Streaming Data, Mobile App Development, Data Analytics, Back-end Development, Front-end Development

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring