Bibeksheel Kaur, Developer in Kalimpong, West Bengal, India
Bibeksheel is available for hire
Hire Bibeksheel

Bibeksheel Kaur

Verified Expert  in Engineering

Data Engineer and Developer

Location
Kalimpong, West Bengal, India
Toptal Member Since
June 24, 2020

Bibeksheel has over 12 years of experience in the IT industry, specializing in the design, development, maintenance, and support of database applications across various platforms and cloud implementations. She is skilled in Microsoft SQL Server, SSIS, Azure Data Factory, Azure Databricks, and Azure Data Lake Storage, and proficient with Snowflake, AWS, and Amazon S3. A quick learner and self-motivated professional, Bibeksheel excels in multitasking and consistently delivers high-quality results.

Portfolio

PepsiCo Global - PepsiCo International Limited
Data Engineering, Azure Data Factory, Python, Scala, R, Databricks, PySpark...
HCL Technologies
Apache Spark, Azure Data Lake, Storage, Databricks, Azure Data Factory...
Infosys, Ltd.
Java, SQL Server 2012, SQL Server 2008, T-SQL (Transact-SQL)...

Experience

Availability

Part-time

Preferred Environment

Microsoft Azure, Microsoft SQL Server, Amazon Web Services (AWS), Snowflake, PySpark, Azure DevOps, T-SQL (Transact-SQL), Azure Databricks, Azure Data Factory, Data Engineering

The most amazing...

...thing I've redesigned is the statement mailing process for a banking client to reduce the overall cost by $0.9 million.

Work Experience

Senior Data Engineer (via Toptal)

2020 - PRESENT
PepsiCo Global - PepsiCo International Limited
  • Collaborated with stakeholders to understand business requirements and design effective data architectures. This involved creating data models, defining data storage strategies, and ensuring data quality and integrity.
  • Implemented data ingestion pipelines to extract data from various sources, transform it, and load it into Azure data storage solutions.
  • Developed data transformation workflows to clean, enrich, and aggregate data to perform data wrangling, integration, and orchestration tasks.
  • Monitored data pipelines and workflows to ensure data integrity, performance, and reliability. Identified and resolved data ingestion, transformation, storage, and processing issues.
  • Automated and optimized data solutions for cost-efficiency and performance.
  • Collaborated with cross-functional teams, such as data scientists, business analysts, and software developers, to understand their needs and provide the required data solutions.
  • Communicated effectively with stakeholders to gather requirements, provide updates, and present insights.
  • Developed an ROI solution comprising marketing and sales data. It involves multiple data sources to help the business team understand net spending versus revenue for PepsiCo products.
  • Developed several dbt models and orchestrated pipelines using Airflow. Optimized DAGs and models two times to improve the overall execution time.
  • Designed final reporting tables in Snowflake. Played an important role in the data vault architecture of initial tables and evolved models over time to satisfy business needs.
Technologies: Data Engineering, Azure Data Factory, Python, Scala, R, Databricks, PySpark, Data Analysis, Data Warehousing, T-SQL (Transact-SQL), Azure Data Lake

Technical Lead

2018 - 2020
HCL Technologies
  • Migrated the existing on-prem ETL solutions residing locally to a global cloud platform. Implemented end-to-end data solutions (storage, integration, processing, and visualization) in Azure.
  • Recreated the functionality of the current on-prem ETL solution by scripting the data transformations in Databricks using Spark SQL, PySpark, and Scala.
  • Created end-to-end database solutions in accordance with business requirements. Optimized existing database solutions and enhanced the performance of databases.
  • Created analytical reports for database solutions to help business users make future decisions accordingly.
  • Supported various BAU activities such as defect prevention, change, incident, and problem management.
  • Worked in an interactive Agile and DevOps environment.
Technologies: Apache Spark, Azure Data Lake, Storage, Databricks, Azure Data Factory, SQL Server Integration Services (SSIS), T-SQL (Transact-SQL), Microsoft SQL Server, Python, Data Engineering, PySpark, Data Analysis, Data Warehousing, HCLSoftware

Senior Systems Engineer

2012 - 2020
Infosys, Ltd.
  • Redesigned the statement mailing process for the client to reduce the overall cost by US $0.9 million. Supported the critical process of sending the annual statements to 11 million savings account stakeholders.
  • Optimized and reengineered the ETL solution to implement the travel insurance process for a client. Wrote and tuned stored procedures, subqueries, functions, triggers, and views to maintain referential integrity and implement complex business logic.
  • Handled business queries and CSR and was involved in 120+ mart development and enhancement. Made recommendations for performance improvement in hosted databases involving partitioning, index creation, index removal, and index modification.
Technologies: Java, SQL Server 2012, SQL Server 2008, T-SQL (Transact-SQL), SQL Server Integration Services (SSIS), Data Analysis, Data Warehousing

BI Modernization

A project to migrate an on-prem data warehouse reporting solution to Microsoft Azure. It contains 100+ tables and 50+ stored procedures, which have to be rescripted in Spark SQL, PySpark, and Scala. End reporting tables will be presented as a Microsoft Power BI report to the business users who will consume the data for analysis and prediction purposes.

ACCOMPLISHMENTS
• Used Azure Data Factory to build pipelines for extracting source data from SQL tables and FlatFiles.
• Rescripted the transformation logic by scripting it in Spark SQL, PySpark, and Scala.
• Recreated existing business logic and functionality using Azure stack.
• Designed, developed, and deployed the solution.
• Implemented end-to-end data solutions (storage, integration, processing, visualization) in Azure.

GSS-DW

An enterprise-wide data warehouse platform to migrate the existing on-prem data residing locally to a global cloud platform and run analytics on top of the data, which will be used for decision-making and predictive analysis.
- Migrate the existing ETL stack from SSIS to the Azure platform.
- Redesign the existing DW solution by writing it in Databricks and integrating it.
- Architect and implement medium to large scale BI solutions on Azure using Azure Data Platform services (Azure Data Lake, Data Factory, Azure SQL DW, and HDInsight/Databricks)
- Recreating existing application logic and functionality in the Azure Data Lake, Data Factory, SQL Database and SQL data warehouse environment. experience in DWH/BI project implementation using Azure DF.
- Propose architectures considering cost/spend in Azure and develop recommendations to right-size data infrastructure
- Implement end-to-end data solutions (storage, integration, processing, visualization) in Azure
- Migrate on-premise data (Oracle/SQL Server/Mainframe) to Azure Data Lake Storage using Azure Data Factory.
- Create dashboards on the delta tables in Azure Databricks and present the analytics per requirements

Ropes And Repairs

Migration of 30 years of historical data from Mainframe files to Azure where I migrated 3000 files from the mainframe system to Azure Data Lake storage.
Populated the data in Azure tables (CosmosDB).
- Created Python scripts to refine the Mainframe historical files so that they are eligible for processing in Azure.
- Created Delta tables from the flat files in Azure Databricks and created a Data model on top of it.
- Set up the ETL Process for rendering of monthly reports from Azure for future use.

FMG Data Warehouse

Worked as the offshore lead in this project for managing data warehouse for a mining company. Deployed and uploaded the SSRS reports to SharePoint Server for the end-users and involved in enhancements and modifications. Also managed the subscription reports for the end-users. Created databases and schema objects including tables, indexes, and applied constraints, connected various applications to the database and written functions, and stored procedures and triggers. Extensively used SSIS transformations such as lookup, derived column, data conversion, aggregate, conditional splits, SQL tasks, script tasks, send mail tasks, and more. Used execution plans, SQL profiler, and database engine tuning advisor to optimize queries and enhance the performance of databases. Supported various BAU activities such as defect prevention, change management, incident management, and problem management. Responsible for monitoring and making recommendations for performance improvement in hosted databases involving partitioning, index creation, index removal, index modification, and adding scheduled jobs to re-index and update statistics in databases. Monitoring 100+ jobs as well as troubleshooting and reporting in case of failure.

Flex Accounts Travel Insurance

Worked as an onsite coordinator in the UK for a year supporting travel insurance annual review processing according to the eligibility rules provided by the business and informing users regarding entitlement, confirmation, and revocation of their travel insurance. Optimized and re-engineered the travel insurance process for current accounts nationwide. Supported flex account travel insurance processing and answered business queries related to customers' data.

Statement Efficiency

Re-designed the statement mailing process for savings customers nationwide to reduce the overall cost by US $0.9 million. Identified the loopholes in the existing logic and provided solutions for them. Supported the critical process of annual savings mailings to send the annual statements to 11 million savings account stakeholders.

Team Tidbit

Served as an offshore coordinator and acted as a bridge between the client and developers. Also worked as a SQL developer and developed stored procedures, functions, table scripts, and rewrites of stored procedures to improve performance. Worked on unit testing by applying various positive and negative test cases to provide defect-free solutions. Created SSIS packages for the smooth flow of data between different sources.

Corporate Analytical Warehouse

Supported critical business activities and created SQL script queries to resolve business queries. Supported various BAU activities such as defect prevention, change management, incident management, and problem management. Supported incident monitoring activities and resolved incidents quickly. Handled business queries and CSR (service requests), involved in the development and enhancement of 20+ marts (SQL stored procedures). Monitored and made recommendations for performance improvement in hosted databases involving partitioning, index creation, index removal, index modification, and adding scheduled jobs to re-index and update statistics in databases. Involved in troubleshooting and fine-tuning of databases for its performance and concurrency.

Aetna Quoting Centre (AQC)

AQC is a Java/J2EE web-based application to provide rating and quotation services for Aetna's medical, life, and dental insurance product application development in Java using Spring MVC and Struts framework with Waterfall and Agile SDLC models. Worked on a wide variety of database issues that plague a complex software system related to concurrency, caching, tuning of databases, and query optimization. Analyzed and resolved existing bugs in the application.

Microsoft IT (MSIT)

Maintained and supported 80+ applications and tools developed in .NET. Made necessary updates and generated reports using SQL queries, and worked in Active Directory. Created T-SQL stored procedures and traced and analyzed the stored procedures for performance and cost optimization. Developed tools in .NET to automate various infrastructure management service activities to deliver exceptional client service.
2007 - 2011

Bachelor's Degree in Electronics and Communication Engineering

Panjab University - Chandigarh

JULY 2013 - PRESENT

70-433 Microsoft SQL Server 2008, Database Development

Microsoft Technologies

NOVEMBER 2012 - PRESENT

70-461 Querying Microsoft SQL Server 2012

Microsoft Technologies

Libraries/APIs

PySpark

Tools

Spark SQL

Languages

T-SQL (Transact-SQL), SQL, Scala, Python, Java, Snowflake, R

Platforms

Databricks, Azure PaaS, Azure, Amazon Web Services (AWS), Jakarta EE, HCLSoftware

Paradigms

Agile, ETL, DevOps, Azure DevOps

Storage

SQL Server Integration Services (SSIS), SQL Server 2012, SQL Server Reporting Services (SSRS), Microsoft SQL Server, Azure SQL, Azure Table Storage, SQL Server 2008, Amazon S3 (AWS S3)

Frameworks

Apache Spark

Other

Azure Data Lake, Data Warehousing, Data Analysis, Azure Data Factory, Microsoft Azure, Data Warehouse Design, Data Engineering, ServiceNow, Performance Tuning, Storage, Electronics, Azure Databricks

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring