Suresh Makkena, Developer in Toronto, ON, Canada
Suresh is available for hire
Hire Suresh

Suresh Makkena

Verified Expert  in Engineering

Data Engineer and Developer

Toronto, ON, Canada

Toptal member since December 2, 2024

Bio

Suresh is a senior data engineer with six years of experience in Azure and 13 years of overall IT experience. He has collaborated with top-tier firms like AT&T, Zurich Farmers, Mylan Pharmaceuticals, and Raas Infotek. He is proficient in Azure Data Factory for designing and implementing data integration pipelines and Azure Databricks for scalable data solutions. Suresh is adept at enhancing data management through Delta Live Tables and Unity Catalog.

Portfolio

Raas Infotek
Azure Cloud Services, Azure SQL, Azure Data Lake, Azure Data Factory (ADF)...
Mylan
Azure Data Lake, Azure Cloud Services, Azure Data Factory (ADF), Azure SQL...
Zurich Farmers
Azure Cloud Services, Azure Data Lake, Azure Data Factory (ADF)...

Experience

  • Azure Cloud Services - 6 years
  • Azure Databricks - 6 years
  • Azure SQL - 6 years
  • PySpark - 6 years
  • Azure Data Lake - 6 years
  • Azure Data Factory (ADF) - 6 years
  • Data Lakes - 6 years
  • Delta Live Tables (DLT) - 3 years

Availability

Full-time

Preferred Environment

Azure

The most amazing...

...team I've led leveraged a serverless SQL pool over a dedicated SQL pool for data storage in slowly changing dimensions (SCD), reducing cost significantly.

Work Experience

Azure Data Engineer

2023 - PRESENT
Raas Infotek
  • Implemented Delta Lake to improve data reliability and performance through ACID transactions and scalable metadata handling.
  • Designed and orchestrated complex ETL pipelines. Integrated them with Azure Databricks and Azure Synapse to process and transform large datasets.
  • Enabled seamless data movement and transformation across various data sources, both on premise and in the cloud.
  • Supported large-scale data processing, ensuring efficient and reliable data workflows for enterprise-level operations.
  • Implemented a proof of concept (POC) for a DevOps pipeline, enhancing automation and efficiency in data workflows. Integrated email tasks to streamline notifications and improve communication within the team.
Technologies: Azure Cloud Services, Azure SQL, Azure Data Lake, Azure Data Factory (ADF), Azure Databricks, Azure Synapse Analytics, Azure SQL Data Warehouse, Azure DevOps, Python, PySpark, Delta Live Tables (DLT), SQL Server 2007, SQL Server 2019

Azure Data Engineer

2019 - 2023
Mylan
  • Designed and orchestrated complex ETL pipelines. Integrated them with Azure Databricks and Azure Synapse to process and transform large datasets.
  • Streamlined data storage solutions by migrating legacy systems to Azure SQL Database, achieving a 50% improvement in data retrieval speeds and enhancing system reliability.
  • Supported large-scale data processing, ensuring efficient and reliable data workflows for enterprise-level operations.
Technologies: Azure Data Lake, Azure Cloud Services, Azure Data Factory (ADF), Azure SQL, Azure Synapse Analytics, Azure Databricks, PySpark, Delta Live Tables (DLT), SQL Server 2019

Azure Data Engineer

2016 - 2017
Zurich Farmers
  • Designed and orchestrated complex ETL pipelines. Integrated them with Azure Databricks and Azure Synapse to process and transform large datasets.
  • Implemented Delta Lake to improve data reliability and performance through ACID transactions and scalable metadata handling.
  • Supported large-scale data processing, ensuring efficient and reliable data workflows for enterprise-level operations.
Technologies: Azure Cloud Services, Azure Data Lake, Azure Data Factory (ADF), Azure Databricks, Azure Synapse Analytics, Azure DevOps, PySpark, SQL Server 2019

Mainfame Developer

2014 - 2015
Zurich Insurance
  • Implemented new auto policy features for Zurich Farmers, streamlining policy issuance and renewal processes and reducing manual intervention by 70%.
  • Designed and implemented an automated batch flow for the inbound process, reducing manual processing time by 80% and increasing data accuracy by 95%.
  • Implemented production fixes, resulting in a 40% reduction in incident count and a 25% decrease in downtime, ensuring high availability and reliability of critical systems.
Technologies: COBOL, JCL, CICS, IBM Db2

Mainfame Developer

2007 - 2013
AT&T
  • Implemented production fixes, resulting in a 40% reduction in incident count and a 25% decrease in downtime, ensuring high availability and reliability of critical systems.
  • Implemented a large load module in the mainframe system, processing 10 million records daily with a 99.99% success rate and reducing processing time by 30%.
  • Designed and implemented an automated batch flow for the inbound process, reducing manual processing time by 80% and increasing data accuracy by 95%.
Technologies: Mainframe, COBOL, JCL, Virtual Storage Access Method (VSAM), IBM Db2

Experience

Dynamic Load of Dimensions and Facts Using a Single Pipeline

http://www.raasinfotek.com/
Implemented dynamic loading of dimensions and facts in Azure Synapse Analytics using stored procedures. This approach ensures efficient and scalable data processing, maintaining the integrity and performance of a data warehouse.

The procedure dynamically loads data into 15 dimension tables. It handles SCD and ensures that only new or updated records are processed. Another procedure dynamically loads data into eight fact tables. It ensures that surrogate keys are correctly handled and only new or updated records are processed.

Dynamic SQL allows the procedures to handle multiple tables without hardcoding table names:
• Cursor-based looping is used to iterate the list of dimension and fact tables, ensuring each table is processed.
• The MERGE statement is used to perform upserts (update or insert), efficiently handling incremental data loads.

Dynamically Built Pipelines for Transferring Data from on Premise to the Cloud

https://www.viatris.in/
ARCHITECTURE AND COMPONENTS
• Metadata Table stores information about the tables to be processed, including queries and timestamps for incremental loading.
• Stored procedures handle data extraction, transformation, and loading based on the metadata.
• A single pipeline uses the metadata table to process multiple tables within a single pipeline dynamically. It is split into two branches to handle incremental and full load scenarios separately.
• Data is loaded into Delta tables in Azure Data Lake Storage Gen2 (ADLS Gen2). These tables provide ACID transactions, scalable metadata handling, and efficient data processing.
• Serverless SQL is utilized in Azure Synapse Analytics to query data stored in Delta tables. It avoids needing a dedicated SQL data warehouse, significantly reducing costs.

In this project, I have designed and implemented a dynamic data pipeline to migrate data from on premise systems to the cloud. By leveraging a metadata-driven approach, I efficiently manage the ETL process for multiple tables using a single pipeline, optimizing performance and cost.

Inbound-outbound Job Creation Tool

https://dxc.com/
Customized the inbound-outbound tool for commercial line applications to suit requirements. I automated this tool by creating the skeleton for the most frequently used jobs required to run on commercial line applications. I utilized the data acquired through extensive analysis and research on the technology and concepts best suited for implementing this functionality.

This tool provides a 'skeleton' framework for the job flows and allows for quickly tailoring each flow to specific requirements. With this framework in place, batch flows can be prepared and deployed rapidly, ensuring faster time to market and minimizing production delays. Since adopting this tool, the process of implementing batch flows has become notably quicker and more efficient.

Education

1998 - 2021

Master's Degree in Computer Science

National Institute of Technology Karnataka - Mangalore, India

Skills

Libraries/APIs

PySpark

Tools

JCL

Languages

COBOL, Easytrive, C, C++, Python, CICS

Frameworks

Delta Live Tables (DLT)

Platforms

Azure SQL Data Warehouse, Azure Synapse Analytics, Azure

Storage

Database Management Systems (DBMS), Data Lakes, Azure Cloud Services, Azure SQL, IBM Db2, Virtual Storage Access Method (VSAM), IBM Mainframe, SQL Server 2007, SQL Server 2019

Paradigms

Compiler Design, Azure DevOps

Other

Azure Data Lake, Azure Data Factory (ADF), Azure Databricks, Unity Catalog, Operating Systems, Networks, Data Structures, Mainframe

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring