
Suresh Makkena
Verified Expert in Engineering
Data Engineer and Developer
Toronto, ON, Canada
Toptal member since December 2, 2024
Suresh is a senior data engineer with six years of experience in Azure and 13 years of overall IT experience. He has collaborated with top-tier firms like AT&T, Zurich Farmers, Mylan Pharmaceuticals, and Raas Infotek. He is proficient in Azure Data Factory for designing and implementing data integration pipelines and Azure Databricks for scalable data solutions. Suresh is adept at enhancing data management through Delta Live Tables and Unity Catalog.
Portfolio
Experience
- Azure Cloud Services - 6 years
- Azure Databricks - 6 years
- Azure SQL - 6 years
- PySpark - 6 years
- Azure Data Lake - 6 years
- Azure Data Factory (ADF) - 6 years
- Data Lakes - 6 years
- Delta Live Tables (DLT) - 3 years
Availability
Preferred Environment
Azure
The most amazing...
...team I've led leveraged a serverless SQL pool over a dedicated SQL pool for data storage in slowly changing dimensions (SCD), reducing cost significantly.
Work Experience
Azure Data Engineer
Raas Infotek
- Implemented Delta Lake to improve data reliability and performance through ACID transactions and scalable metadata handling.
- Designed and orchestrated complex ETL pipelines. Integrated them with Azure Databricks and Azure Synapse to process and transform large datasets.
- Enabled seamless data movement and transformation across various data sources, both on premise and in the cloud.
- Supported large-scale data processing, ensuring efficient and reliable data workflows for enterprise-level operations.
- Implemented a proof of concept (POC) for a DevOps pipeline, enhancing automation and efficiency in data workflows. Integrated email tasks to streamline notifications and improve communication within the team.
Azure Data Engineer
Mylan
- Designed and orchestrated complex ETL pipelines. Integrated them with Azure Databricks and Azure Synapse to process and transform large datasets.
- Streamlined data storage solutions by migrating legacy systems to Azure SQL Database, achieving a 50% improvement in data retrieval speeds and enhancing system reliability.
- Supported large-scale data processing, ensuring efficient and reliable data workflows for enterprise-level operations.
Azure Data Engineer
Zurich Farmers
- Designed and orchestrated complex ETL pipelines. Integrated them with Azure Databricks and Azure Synapse to process and transform large datasets.
- Implemented Delta Lake to improve data reliability and performance through ACID transactions and scalable metadata handling.
- Supported large-scale data processing, ensuring efficient and reliable data workflows for enterprise-level operations.
Mainfame Developer
Zurich Insurance
- Implemented new auto policy features for Zurich Farmers, streamlining policy issuance and renewal processes and reducing manual intervention by 70%.
- Designed and implemented an automated batch flow for the inbound process, reducing manual processing time by 80% and increasing data accuracy by 95%.
- Implemented production fixes, resulting in a 40% reduction in incident count and a 25% decrease in downtime, ensuring high availability and reliability of critical systems.
Mainfame Developer
AT&T
- Implemented production fixes, resulting in a 40% reduction in incident count and a 25% decrease in downtime, ensuring high availability and reliability of critical systems.
- Implemented a large load module in the mainframe system, processing 10 million records daily with a 99.99% success rate and reducing processing time by 30%.
- Designed and implemented an automated batch flow for the inbound process, reducing manual processing time by 80% and increasing data accuracy by 95%.
Experience
Dynamic Load of Dimensions and Facts Using a Single Pipeline
http://www.raasinfotek.com/The procedure dynamically loads data into 15 dimension tables. It handles SCD and ensures that only new or updated records are processed. Another procedure dynamically loads data into eight fact tables. It ensures that surrogate keys are correctly handled and only new or updated records are processed.
Dynamic SQL allows the procedures to handle multiple tables without hardcoding table names:
• Cursor-based looping is used to iterate the list of dimension and fact tables, ensuring each table is processed.
• The MERGE statement is used to perform upserts (update or insert), efficiently handling incremental data loads.
Dynamically Built Pipelines for Transferring Data from on Premise to the Cloud
https://www.viatris.in/• Metadata Table stores information about the tables to be processed, including queries and timestamps for incremental loading.
• Stored procedures handle data extraction, transformation, and loading based on the metadata.
• A single pipeline uses the metadata table to process multiple tables within a single pipeline dynamically. It is split into two branches to handle incremental and full load scenarios separately.
• Data is loaded into Delta tables in Azure Data Lake Storage Gen2 (ADLS Gen2). These tables provide ACID transactions, scalable metadata handling, and efficient data processing.
• Serverless SQL is utilized in Azure Synapse Analytics to query data stored in Delta tables. It avoids needing a dedicated SQL data warehouse, significantly reducing costs.
In this project, I have designed and implemented a dynamic data pipeline to migrate data from on premise systems to the cloud. By leveraging a metadata-driven approach, I efficiently manage the ETL process for multiple tables using a single pipeline, optimizing performance and cost.
Inbound-outbound Job Creation Tool
https://dxc.com/This tool provides a 'skeleton' framework for the job flows and allows for quickly tailoring each flow to specific requirements. With this framework in place, batch flows can be prepared and deployed rapidly, ensuring faster time to market and minimizing production delays. Since adopting this tool, the process of implementing batch flows has become notably quicker and more efficient.
Education
Master's Degree in Computer Science
National Institute of Technology Karnataka - Mangalore, India
Skills
Libraries/APIs
PySpark
Tools
JCL
Languages
COBOL, Easytrive, C, C++, Python, CICS
Frameworks
Delta Live Tables (DLT)
Platforms
Azure SQL Data Warehouse, Azure Synapse Analytics, Azure
Storage
Database Management Systems (DBMS), Data Lakes, Azure Cloud Services, Azure SQL, IBM Db2, Virtual Storage Access Method (VSAM), IBM Mainframe, SQL Server 2007, SQL Server 2019
Paradigms
Compiler Design, Azure DevOps
Other
Azure Data Lake, Azure Data Factory (ADF), Azure Databricks, Unity Catalog, Operating Systems, Networks, Data Structures, Mainframe
How to Work with Toptal
Toptal matches you directly with global industry experts from our network in hours—not weeks or months.
Share your needs
Choose your talent
Start your risk-free talent trial
Top talent is in high demand.
Start hiring