Abhijeet Gupta, Developer in Bengaluru, Karnataka, India
Abhijeet is available for hire
Hire Abhijeet

Abhijeet Gupta

Verified Expert  in Engineering

Data Engineer and Developer

Bengaluru, Karnataka, India

Toptal member since June 18, 2020

Bio

Abhijeet is a seasoned data engineer with more than 10 years of industry experience. He has considerable hands-on experience in Microsoft Azure tech stacks such as Azure Data Factory, Databricks, Azure SQL Data Warehouse, Azure Logic Apps, Key Vault, Delta Lake, and Data Lake. Abhijeet has completed several projects with Fortune 500 companies. He also has valuable knowledge in SQL, ETL, Unix shell scripting, and Agile.

Portfolio

Johnson & Johnson
Azure, Azure Data Factory (ADF), Dedicated SQL Pool (formerly SQL DW)...
Maersk
Scala, SQL, Python, Spark, Big Data, Microsoft SQL Server
Accenture Services
Informatica, Unix, Teradata, Azure, Azure PaaS, Azure Blobs

Experience

  • SQL - 7 years
  • Data Engineering - 6 years
  • Data Warehouse Design - 4 years
  • Data Warehousing - 4 years
  • Spark SQL - 3 years
  • Azure - 3 years
  • Spark - 3 years
  • Unix Shell Scripting - 3 years

Availability

Part-time

Preferred Environment

SQL, Unix, Azure, Spark, Azure Data Factory (ADF), Azure Data Lake, Azure SQL Data Warehouse, Dedicated SQL Pool (formerly SQL DW), Databricks, PySpark

The most amazing...

...thing I've coded are complex stored procedures where complex business logic is implemented to fulfill the user's requirements.

Work Experience

Lead Data Engineer

2020 - PRESENT
Johnson & Johnson
  • Authored several complex Azure data factory pipelines to migrate the data from on-prem to Azure Cloud.
  • Wrote complex Databricks notebooks to apply business transformation logic to build several KPIs.
  • Used Informatica to load historical data from on-prem to the Azure Cloud.
  • Wrote views in Azure SQL Data Warehouse to apply data governance rules.
  • Completed Performance Tuning on tables and view in Azure SQL Data Warehouse.
  • Used PolyBase mechanism to load data from Azure Data Lake Storage to Azure SQL Data Warehouse.
Technologies: Azure, Azure Data Factory (ADF), Azure SQL Data Warehouse, Dedicated SQL Pool (formerly SQL DW), Databricks, Big Data, Azure Blob Storage API, Informatica ETL, Teradata, SQL, Migration, Microsoft SQL Server, Databases, Pandas, Data Engineering, Microsoft Excel, JSON, ETL

Senior Data Engineer

2018 - PRESENT
Maersk
  • Created a unified data platform.
  • Worked on Plantir Foundry to build business use cases.
  • Created generic data bricks notebooks to cleanse or wrangle the data.
  • Worked extensively with big data technologies like Spark, Scala, Pyspark, Hive, HDFS, and Sqoop.
  • Developed with the following tools: MPP database (Teradata), ETL tool (Informatica) and Informatica Developer tool, Unix, shell scripting, Autosys, and Control-M.
  • Contributed analytical skills, programming, problem-solving, troubleshooting, and mentoring to various projects.
Technologies: Scala, SQL, Python, Spark, Big Data, Microsoft SQL Server

Senior Data Engineer

2015 - 2018
Accenture Services
  • Supported a master data management (MDM) application for a client.
  • Wrote stored procedures to apply business logic to data.
  • Worked on optimizing and tuning the Teradata views and SQL to improve the performance of batch and data response time for users.
  • Created Informatica (ETL) mappings and workflows to extract and load source data to the target system.
  • Developed Unix shell scripts to automate BTEQ scripts.
  • Coordinated with the change management team in code deployments.
  • Managed tickets and incidents by using an HP service manager.
  • Provided quick production fixes and was proactively involved in fixing production support issues.
  • Implemented Borland's StarTeam as a versioning tool.
  • Prepared basic reports and supporting documents using Microsoft Excel.
Technologies: Informatica, Unix, Teradata, Azure, Azure PaaS, Azure Blobs

Data Engineer

2014 - 2015
Capgemini India Pvt Ltd.
  • Wrote stored procedures to apply business logic to data.
  • Used Teradata-loading utilities to load data from the source to a Teradata database.
  • Authored very complex Unix shell scripts to preprocess the staging data.
  • Created database objects such as tables, views, and indexes.
  • Used a BTEQ utility to access and query databases.
Technologies: PL/SQL, Unix, Teradata

Associate Developer

2013 - 2014
Cognizant Technology Solution, India
  • Improved stored procedures and enhanced their performance.
  • Created database objects.
  • Used BTEQ in a Unix environment.
  • Tuned SQL queries to overcome spool space errors and improve performance.
Technologies: Unix, Teradata

Software Engineer

2011 - 2013
Exilant Technologies, India
  • Wrote complex procedures to perform aggregations in a semantic layer and to written result sets to the front-end user.
  • Created database objects like tables, view, and indexes.
  • Contributed to different portions of the development lifecycle such as development, testing, UAT, and production deployment.
  • Developed Informatica sessions and workflows.
  • Wrote Unix shell scripts.
  • Connected an AIX file system to the Teradata database in the network attached system.
  • Developed AutoSys jobs (JIL).
  • Extensively worked on Teradata utility tools like BTEQ, FastLoad, Multiload, TPUMP, FastExport, and TPT (Teradata Parallel Transporter).
  • Used SQL assistant to query Teradata tables; worked with BTEQ in a Unix environment.
  • Implemented performance tuning, bug fixing, and error handling; set up and supported, in a team, the data for a UAT environment.
Technologies: Informatica, Subversion (SVN), Autosys, Unix, PL/SQL, Teradata

Experience

Lead Data Engieer

A major project of one of the largest pharmaceutical companies where was involved in the following assignments:

• Migrated a considerable amount of data from on-prem to Azure Cloud.
• Authored several data pipelines for data movement.
• Wrote very complex Databricks notebooks to apply business transformation logic.
• Used Azure dedicated SQL pools as a Data warehouse.
• Completed performance tuning of SQL Data Warehouse tables and views.

Maestro (Unified Data Platform)

Tasks Accomplished:

• Worked on the Azure stack of big data and its components.
• Developed Spark programs in Scala to perform data transformation, creating data frames, and running Spark SQL.
• Took part in converting Teradata queries into Spark transformations using Spark RDD and Scala.
• Created Hive tables, then loaded, and analyzed data using Hive queries.
• Designed and created data models for business using Hive and Spark SQL.
• Constructed and scheduled notebooks in Azure Databricks to perform data cleansing and to do transformations.
• Contributed to the creation and scheduling of data pipelines using Azure Data Factory.
• Created the L1 canonical model for GCSS sources.

SustainTeam at Accenture

Client: PepsiCo
Position: Senior Software Engineer
Technologies Used: Teradata, PL/SQL, Unix

Description:

I built a master data management (MDM) system that deals with various source systems and domains. It manages customers, materials, locations, finances, vendors, information, and reference data. There are various stages to manage and publish data to the down streams, respectively. The MDM provides the user interface to maintain the global attributes not found in any source.

Tasks Accomplished:

• Investigated production environment failure.
• Worked on user requests.
• Wrote stored procedures for tasks.
• Investigated QA environment failures.

Education

2006 - 2010

Bachelor's Degree in Electronics and Instrumentation Engineering

Shri Govindram Seksaria Institute of Technology and Science - Indore, India

Certifications

OCTOBER 2020 - PRESENT

Azure 900

Microsoft

Skills

Libraries/APIs

Pandas, Azure Blob Storage API, PySpark

Tools

Spark SQL, Oracle GoldenGate, Informatica PowerCenter, Microsoft Excel, Subversion (SVN), Control-M, Autosys, Informatica ETL, BigQuery, Apache Airflow

Languages

Python, SQL, Scala

Storage

PL/SQL, Teradata, Oracle PL/SQL, Databases, JSON, Apache Hive, HDFS, Microsoft SQL Server, Azure Blobs, Master Data Management (MDM)

Frameworks

Spark

Paradigms

Agile Software Development, Dimensional Modeling, ETL

Platforms

Databricks, Windows, Windows XP, Unix, Azure, MacOS, Azure PaaS, Azure SQL Data Warehouse, Dedicated SQL Pool (formerly SQL DW)

Other

Data Warehouse Design, Data Warehousing, Data Engineering, Informatica, Big Data, UNIX Utilities, Unix Shell Scripting, Logistics, Blob Storage, Azure Data Factory (ADF), Migration, Azure Data Lake, Electronics, Engineering, Software Engineering, Data Modeling, Google BigQuery

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring