Muhammad Bilal, Developer in Islamabad, Islamabad Capital Territory, Pakistan
Muhammad is available for hire
Hire Muhammad

Muhammad Bilal

Verified Expert  in Engineering

Database Engineer and Developer

Location
Islamabad, Islamabad Capital Territory, Pakistan
Toptal Member Since
September 19, 2022

Muhammad has seven years of experience transforming raw data into useful insights for enterprises. An enterprise resource planning (ERP) developer turned into a data engineer, he is an expert in database (DB) and data warehouse (DWH) design, writing complex SQL queries, and ETL development using native and 3rd-party tools and utilities. Muhammad is a data-focused person who profoundly understands every step of designing data ingestion pipelines and ETL processing.

Portfolio

Philip Morris International
Amazon S3 (AWS S3), Amazon Athena, Snowflake, SQL, Query Optimization...
IBM
ETL, SQL, IBM InfoSphere (DataStage), Oracle, PostgreSQL, Migration...
China Mobile Pakistan | Zong 4G
SPSS Modeler, ETL, Data Science, Data Analysis, SQL, Oracle, Python...

Experience

Availability

Part-time

Preferred Environment

SQL, Data Engineering, ETL, Azure Databricks, Spark, Modeling, Data Architecture, Databricks

The most amazing...

...ETL data modernization program I've led transformed a legacy DWH into a modern data platform and enabled a telco to access a single version of the truth.

Work Experience

Data Architect

2023 - PRESENT
Philip Morris International
  • Worked on the architectural enhancement of existing data products. Added new functionalities, fixed bugs, and extended the functionalities to new markets. I own several data marts within EDW. I perform model changes and update transformation logic.
  • Distributed assigned tickets to my team. I distributed these tickets among my team members and myself as per the nature of the task. My team includes data engineers and Microsoft Power BI experts.
  • Interfaced with customers to understand their requirements and document them with the help of a business analyst. Later, I implemented changes to the model and transformational logic, followed up by testing in the QAS environment.
  • Deployed, after QAS, changes in the production environment.
Technologies: Amazon S3 (AWS S3), Amazon Athena, Snowflake, SQL, Query Optimization, SQL Stored Procedures, Data Modeling, Database Modeling, Dimensional Modeling

Data Engineer

2020 - 2022
IBM
  • Developed data pipelines as required by the mapping documents in IBM InfoSphere DataStage.
  • Refined the build architecture to streamline the ETL pipeline development and optimize the development cycle.
  • Created and maintained the technical documentation required to support solutions.
Technologies: ETL, SQL, IBM InfoSphere (DataStage), Oracle, PostgreSQL, Migration, Star Schema, Dimensional Modeling, Data Modeling, Data Analytics, Data Architecture, Business Intelligence (BI), Performance Tuning, Query Composition, Datasets, Data Profiling, Data Cleaning, Data Cleansing, Technical Writing, Technical Documentation, Oracle Database, Databases, Pipelines, Database Modeling, PL/SQL Tuning, Query Optimization, Data, Modeling, Data Pipelines, Data Warehousing, Database Development, Data Analysis, Data Manipulation Language (DML), Data Definition Languages (DDL), Data Warehouse Design, Oracle Exadata, SQL Stored Procedures, Stored Procedure, Data Management, OLAP

Assistant Manager in Fraud Management

2014 - 2017
China Mobile Pakistan | Zong 4G
  • Implemented an end-to-end grey traffic identification system, from the provided requirements to implementation. Saved approximately 30 million international direct dialing (IDD) minutes per month terminated via grey traffic.
  • Developed controls to monitor Global System for Mobile Communication (GSM) data streams and stats to identify any abnormal patterns suspected of fraudulent activity.
  • Prevented daily loss of 0.98 million due to an auto re-subscription process.
  • Prevented PKR323,000 commission fraud in mobile financial service (MFS) peer-to-peer (P2P) transactions during one month.
  • Identified 476,000 fake activations during Q3 2014 and saved the commission.
  • Identified fraudulent transactions in balance sharing and saved millions of PKR.
  • Identified fraudulent transactions in the usage of a social bundle and saved potential revenues for the company.
Technologies: SPSS Modeler, ETL, Data Science, Data Analysis, SQL, Oracle, Python, Data Analytics, Business Intelligence (BI), Performance Tuning, Query Composition, Datasets, Data Profiling, Data Cleaning, Data Cleansing, Oracle Database, Databases, Normalization, Database Modeling, Data, Data Warehousing, Database Development, Star Schema, Data Manipulation Language (DML), Data Definition Languages (DDL), SQL Stored Procedures, Stored Procedure

Team Lead of Data Warehouse Development

2012 - 2014
China Mobile Pakistan | Zong 4G
  • Implemented query optimization, table partitions, and parallel execution schemes to reduce total batch execution time by three hours.
  • Implemented housekeeping scripts to monitor and maintain execution logs.
  • Worked on ETL development for 3G/4G implementation. Modified ETL routines and revised a value-added service (VAS) data model following new business dynamics, i.e., new dimensions derived by tax structure, 3G/4G sites, and 3G/4G data users.
Technologies: ETL, Dimensional Modeling, Oracle PL/SQL, SQL, Oracle, Data Analysis, Star Schema, Data Analytics, Data Architecture, Business Intelligence (BI), Performance Tuning, Query Composition, Datasets, Data Profiling, Data Cleaning, Data Cleansing, Oracle Database, Databases, Pipelines, Normalization, Database Modeling, PL/SQL Tuning, Query Optimization, Data, Modeling, Database Design, Data Pipelines, Data Warehousing, Database Development, Data Manipulation Language (DML), Data Definition Languages (DDL), Data Warehouse Design, SQL Stored Procedures, Stored Procedure, Data Management, OLAP

Senior ERP Developer

2009 - 2012
National University of Sciences and Technology (NUST)
  • Designed and developed a payroll module and telephone extension billing module as part of an internal ERP application managed by the NUST ICT department.
  • Collaborated with technical and commercial groups to understand and document their requirements by analyzing existing workflows and procedures.
  • Reviewed and rewrote older programs to increase operating efficiency, enhance the customer experience, and adapt to new requirements. Investigated and resolved open bugs.
Technologies: Database Design, Oracle PL/SQL, Oracle Forms & Reports, SQL, PL/SQL, Oracle, Database Development, Relational Databases, Data Modeling, erwin Data Modeler, Data Architecture, Query Composition, Oracle Database, Databases, Normalization, Database Modeling, PL/SQL Tuning, Query Optimization, Data, Modeling, Programming, Performance Tuning, Data Manipulation Language (DML), Data Definition Languages (DDL), SQL Stored Procedures, Stored Procedure, Data Management, OLTP

ERP Developer

2007 - 2009
Government of Pakistan
  • Designed a material procurement module for an ERP application developed and managed internally by the department.
  • Designed an inventory management module for an ERP application developed and managed internally by the department.
  • Implemented the business logic using Oracle PL/SQL.
Technologies: Database Design, Database Development, Oracle PL/SQL, SQL, PL/SQL, Oracle, Relational Databases, Data Modeling, Data Architecture, erwin Data Modeler, Query Composition, Oracle Database, Databases, Normalization, Database Modeling, PL/SQL Tuning, Query Optimization, Modeling, Programming, Performance Tuning, Data Manipulation Language (DML), Data Definition Languages (DDL), SQL Stored Procedures, Stored Procedure, Data Management, OLTP

Oracle Developer

2006 - 2007
AI Soft | System Developers
  • Conducted a root cause analysis (RCA). Debugged and fixed issues in codes suggested by the client's tickets.
  • Implemented changes in business logic and client requirements.
  • Developed business reports in Oracle Reports 10g.
Technologies: Oracle PL/SQL, Oracle Forms & Reports, PL/SQL, SQL, Oracle, Relational Databases, Query Composition, Oracle Database, Databases, Normalization, PL/SQL Tuning, Programming, Data Manipulation Language (DML), Data Definition Languages (DDL), SQL Stored Procedures, Stored Procedure, OLTP

Data Warehouse Modernization

Served as the lead ETL developer in a data modernization program for a Gulf Cooperation Council's leading telecom company to transform their legacy data warehouse into a modern data platform and enable the organization to access a single version of the truth.

The project deliverables included:
• Migration of existing data from a legacy data warehouse platform to an enterprise data warehouse, including data pipelines developed on the IBM InfoSphere DataStage platform.
• Understanding and documentation of the existing data architecture.
• Transformation of existing data architecture to modern data architecture, including new business requirements.
• Design documentation followed by business reviews and a sign-off.
• ETL development guided by the design document and followed by tech and business user acceptance testing.
• Deployment with handover training sessions to AMS and operations teams.
• Technical training on new data models for business.
• Data quality framework development and implementation to ensure data quality concerning the enterprise data warehouse.

Advanced Data Analytics Platform

Developed an advanced data analytics platform to identify fraudulent suspicious activities in the telecom network. While working on the project, I contributed to optimizing company processes, prevented fraud and daily losses totaling over 1 million Pakistani rupees, and identified fraudulent transactions saving the company millions of Pakistani rupees.

Data Warehouse Optimization

Contributed to a project for China Mobile, Pakistan's second-largest telecom operator, with 70 terabytes of legacy data warehouse in Oracle Database 11g. They had over 1 billion call detail records in data sessions, calls, text messages, recharges, unstructured supplementary service data, and value-added service from 25 million active subscribers, which were processed in daily batches.

I led the optimization of the ETL batch and saved two hours. We started with the implementation of logging for each job to find out which one was taking the longest.

To optimize query executions, we performed these actions:
• Partitioning and subpartitioning of tables.
• Creating proper local and global indexes.
• Gathering stats.
• Applying data retention.
• Rewriting queries.
• Developing materialized views for the business.

Migration of ETL Logic from ADF to Azure Databricks

Radix is an online, cloud-based multifamily market research and analytics platform with built-in tools designed to help users make sense of ambiguous data and drive performance. My role was to migrate ETL logic from ADF pipelines to Azure Databricks Notebooks.

• Develop pipelines in Azure Databricks using PySpark notebooks to ingest data from multiple sources, including Azure Eventhub, AWS Containers, SFTP, and MongoDB.
• Save data as parquet files in the enterprise landing layer, i.e., Azure Data Lake Storage
• Transform data and sink it into a dimensional model implemented in Azure Synapse Analytics

Data Migration

I migrated the data from the Azure SQL data warehouse to Postgre DB. I developed PySpark notebooks on Azure Databaricks to migrate data from Azure SQL Server to Postgre DB. I also developed a reconciliation script to reconcile data between both databases.

Data Warehouse Enhancements

I was part of China Mobile (Pakistan) when Zong acquired a 3G/4G license. I implemented a data model that was not rich enough to produce BI reports as per company requirements post-3G/4 G network rollout.

I led a team of data engineers to implement changes in the data model by introducing new facts and dimensions, followed by changes in ETL pipelines to sink data into a new model. We also develop new ETL pipelines to ingest 3G/4G network equipment data sources. BI reports were also created per company requirements to report 3G/4 G-related business KPIs.

Data Warehouse Data Quality Roadmap Implementation

This assignment was for Vodafone Qatar EDW Operations. The business used to report and lock several data-related issues related to data quality. The ops team spent 90% of their time debugging/RCA and fixing these issues by applying ad-hoc data fixes.

As the lead data engineer, I analyzed why so many issues are being reported daily.

We learned that 70% of issues are related to source data and can be avoided if data quality checks are implemented on the landing layer. So we implemented around 150 reviews on the landing layer to check the quality of source data with acceptance threshold to variation in trend.

Since implementation, the ops team and business team have been at peace.

Data Architect at PMI AMS Team

Phillips Morris International maintaining the EDW application in Snowflake with ETL in WhereScape RED. Data vault models were developed in WhereScape 3D.

I was responsible for the architectural enhancement of all existing data products. The main supporting areas were the Sales and Digital Sustainability Program for any architectural enhancement and modifications.

My contribution:
• Data vault 2.0 design approach for integrated data platform of Data Ocean.
• WhereScape 3D for data vault designing.
• Standardized the implementation framework for data pipelines.
• The delivery of trusted data products to enable data-driven decision-making for the business.
2001 - 2004

Master's Degree in Computer Science

International Islamic University Islamabad (IIUI) - Islamabad, Pakistan

JANUARY 2023 - PRESENT

Microsoft Certified: Azure Fundamentals

Microsoft

JANUARY 2023 - JANUARY 2024

Microsoft Certified: Azure Data Engineer Associate

Microsoft

OCTOBER 2022 - PRESENT

Hands On Essentials - Data Warehouse

Snowflake

OCTOBER 2020 - SEPTEMBER 2023

DP-200 Implementing an Azure Data Solution

Microsoft

Libraries/APIs

PySpark

Tools

IBM InfoSphere (DataStage), Oracle Exadata, SPSS Modeler, Amazon Athena

Paradigms

Database Design, Dimensional Modeling, ETL, Database Development, OLAP, Business Intelligence (BI), Data Science

Storage

PL/SQL, Relational Databases, Data Pipelines, Databases, Database Modeling, Oracle PL/SQL, Data Definition Languages (DDL), SQL Stored Procedures, OLTP, JSON, PostgreSQL, MySQL, Azure SQL Databases, Database Security, Azure Cloud Services, Data Lakes, MongoDB, Oracle 11g, Amazon S3 (AWS S3)

Languages

SQL, Data Manipulation Language (DML), Stored Procedure, Python, Snowflake, Python 3

Platforms

Oracle, Oracle Database, Databricks, Amazon Web Services (AWS), Azure Synapse, Azure, Azure SQL Data Warehouse, Azure Event Hubs, Dedicated SQL Pool (formerly SQL DW)

Frameworks

Spark

Other

Data Engineering, Oracle Forms & Reports, Data Analysis, Star Schema, Data Analytics, Data Modeling, Performance Tuning, Query Composition, Datasets, Data Profiling, Data Cleaning, Data Cleansing, Modeling, PL/SQL Tuning, Query Optimization, Normalization, Data, Data Warehousing, Data Warehouse Design, Data Management, Migration, Data Architecture, erwin Data Modeler, Data Migration, Azure Databricks, Technical Documentation, Data Wrangling, Parquet, Programming, Data Visualization, Linear Regression, GSM, Pipelines, Azure Data Lake, Data Security, Technical Writing, Azure Data Factory, Big Data, Cloud, Data Processing, Azure Stream Analytics, Cloud Security, Storage, PySQL, WhereScape, Data-level Security

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring