Egor Makhov, Developer in Como, Province of Como, Italy
Egor is available for hire
Hire Egor

Egor Makhov

Verified Expert  in Engineering

Data Engineer and Developer

Como, Province of Como, Italy

Toptal member since August 6, 2024

Bio

Egor is a data engineer with extensive experience solving business problems through analytics and big data solutions. He specializes in data modeling, building scalable data architectures, and creating intuitive analytical platforms. His expertise includes transitioning legacy ETL jobs into streaming pipelines and building modern data lakes and warehouses using cloud technologies. Egor has also designed and developed an open-source data quality framework for batch processing and streaming data.

Portfolio

Raiffeisen Bank Russia
Data Lakes, Data Lakehouse, Databricks, Data Architecture...
Oliver James Associates
Apache Kafka, Amazon S3, AWS, Big Data Architecture, Spark Streaming...
Agile Lab
Python, Python, OpenCV, OpenVINO, Deep Learning, Machine Learning...

Experience

Availability

Full-time

Preferred Environment

Linux, Docker, IntelliJ IDEA, Slack, Windows, Mattermost

The most amazing...

...thing I've developed is an open-source data quality framework bundled with end-to-end no-code applications for batch processing and streaming data.

Work Experience

Principal Data Engineer

2021 - PRESENT
Raiffeisen Bank Russia
  • Built low-latency data marts across all CIB banking products, enhancing data accessibility and performance.
  • Created a fast, scalable, and user-friendly computing platform for data analytics and data operations.
  • Developed a universal data quality framework for big data processing, adopted across the entire bank.
  • Designed a GPU cluster for the rapid development of LLM-based applications.
Technologies: Data Lakes, Data Lakehouse, Databricks, Data Architecture, Big Data Architecture, Big Data Architecture, Data Warehouse, Data Warehouse, Database, PostgreSQL, Spark, Big Data Architecture, Python, Scala, Debezium, Kafka Connect, Apache Kafka, Hadoop, HDFS, Hadoop, Trino, LLM, Kubeflow, DataOps, Kubernetes, Docker, Docker Compose, Software Development, PySpark

Data Engineering Consultant

2023 - 2023
Oliver James Associates
  • Migrated data pipelines of a major European stock exchange to a new market protocol, ensuring seamless transitions and compliance with updated standards.
  • Optimized existing architecture to improve performance and efficiency, resulting in faster data processing and reduced operational costs.
  • Introduced data quality checks to ensure completeness and consistency of market events.
Technologies: Apache Kafka, Amazon S3, AWS, Big Data Architecture, Spark Streaming, Delta Lake, Financial Software Development, Stock Market Consulting, Data Science, Data Science, Streaming Data, Scala, PySpark

Machine Learning Engineer

2019 - 2020
Agile Lab
  • Managed dataset collection and training for personal protective equipment (PPE) detection machine learning model.
  • Developed a FLANN-based pattern detection module for access control and personnel identification.
  • Led a team of three developers during the review and refactoring of a feature extraction application for a corporate banking platform.
Technologies: Python, Python, OpenCV, OpenVINO, Deep Learning, Machine Learning, Computer Vision, Big Data Architecture, Scala, PySpark

Big Data Engineer

2017 - 2019
Agile Lab
  • Developed and tested multiple ETL applications at UniCredit Services as part of a 4-member team.
  • Designed, developed, and integrated the data quality core module of an open-source big data analytical framework.
  • Built multiple pipelines during the IoT platform development for Vodafone as part of a 20-member team.
Technologies: Scala, Big Data Architecture, Hadoop, HDFS, Hadoop, Akka, Cloudera

Software Engineering Intern

2016 - 2016
I&b | Innova Et Bella
  • Improved the BrandMemo platform's recommendation engine through advanced machine-learning integration using Python.
  • Reduced application load times by 30% and boosted user retention by 20% through front-end optimization and UX improvements.
  • Increased speeds by 70% by optimizing the internal data model and refactoring the most common SQL queries.
Technologies: Python, SciPy, Swift, Vagrant, Mobile App Design, Data Science, Data Science, Back-end Developers, Front-end Development, JavaScript, Pandas, NumPy, XGBoost, MySQL

Checkita Data Quality Framework

http://www.checkita.org/
An open-source data quality framework bundled with end-to-end no-code applications for batch and streaming data. I created Checkita while I was at Agile Lab and seamlessly integrated it for clients. Upon joining Raiffeisen Bank, I further developed and elevated the framework to a central cornerstone of the bank's data quality platform.
2011 - 2017

Master's Degree in Mathematics and Computer Science

MIREA – Russian Technological University - Moscow, Russia

AUGUST 2020 - AUGUST 2023

Google Cloud Certified Associate Cloud Engineer

Google Cloud

JUNE 2020 - JUNE 2023

AWS Certified Machine Learning – Specialty

Amazon Web Services

Libraries/APIs

Spark Streaming, PySpark, OpenCV, SciPy, Pandas, NumPy, XGBoost

Tools

IntelliJ IDEA, Apache Airflow, Docker Compose, Slack Development, Mattermost, AWS IAM, SageMaker, Kafka Connect, OpenVINO, Hadoop, Cloudera, AWS, Vagrant

Languages

SQL, Scala, Python, Python, Java, Swift, JavaScript

Frameworks

Big Data Architecture, Spark, Hadoop, Trino, Flask, Data Lakehouse, Akka

Paradigms

ETL

Platforms

Apache Kafka, Kubeflow, Kubernetes, AWS, Linux, Docker, Windows Development, Cloud Engineering, Databricks, Debezium, Azure Design

Storage

Database, Data Lakes, PostgreSQL, HDFS, Hadoop, NoSQL, JSON, Amazon S3, MySQL

Other

Software Development, Big Data Architecture, Data Engineering, Data Architecture, Data Warehouse, DataOps, Crytographer, Machine Learning, Data Science, Mathematics, Big Data Architecture, Data Mining, APIs, Data Science, Data Science, Data Quality Analysis, Cloud Computing, Data Warehouse, LLM, Deep Learning, Computer Vision, Delta Lake, Financial Software Development, Stock Market Consulting, Streaming Data, Mobile App Design, Data Science, Back-end Developers, Front-end Development

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring