Muhammad Bilal
Verified Expert in Engineering
Database Engineer and Developer
Muhammad has seven years of experience transforming raw data into useful insights for enterprises. An enterprise resource planning (ERP) developer turned into a data engineer, he is an expert in database (DB) and data warehouse (DWH) design, writing complex SQL queries, and ETL development using native and 3rd-party tools and utilities. Muhammad is a data-focused person who profoundly understands every step of designing data ingestion pipelines and ETL processing.
Portfolio
Experience
Availability
Preferred Environment
SQL, Data Engineering, ETL, Azure Databricks, Spark, Modeling, Data Architecture, Databricks
The most amazing...
...ETL data modernization program I've led transformed a legacy DWH into a modern data platform and enabled a telco to access a single version of the truth.
Work Experience
Data Architect
Philip Morris International
- Worked on the architectural enhancement of existing data products. Added new functionalities, fixed bugs, and extended the functionalities to new markets. I own several data marts within EDW. I perform model changes and update transformation logic.
- Distributed assigned tickets to my team. I distributed these tickets among my team members and myself as per the nature of the task. My team includes data engineers and Microsoft Power BI experts.
- Interfaced with customers to understand their requirements and document them with the help of a business analyst. Later, I implemented changes to the model and transformational logic, followed up by testing in the QAS environment.
- Deployed, after QAS, changes in the production environment.
Data Engineer
IBM
- Developed data pipelines as required by the mapping documents in IBM InfoSphere DataStage.
- Refined the build architecture to streamline the ETL pipeline development and optimize the development cycle.
- Created and maintained the technical documentation required to support solutions.
Assistant Manager in Fraud Management
China Mobile Pakistan | Zong 4G
- Implemented an end-to-end grey traffic identification system, from the provided requirements to implementation. Saved approximately 30 million international direct dialing (IDD) minutes per month terminated via grey traffic.
- Developed controls to monitor Global System for Mobile Communication (GSM) data streams and stats to identify any abnormal patterns suspected of fraudulent activity.
- Prevented daily loss of 0.98 million due to an auto re-subscription process.
- Prevented PKR323,000 commission fraud in mobile financial service (MFS) peer-to-peer (P2P) transactions during one month.
- Identified 476,000 fake activations during Q3 2014 and saved the commission.
- Identified fraudulent transactions in balance sharing and saved millions of PKR.
- Identified fraudulent transactions in the usage of a social bundle and saved potential revenues for the company.
Team Lead of Data Warehouse Development
China Mobile Pakistan | Zong 4G
- Implemented query optimization, table partitions, and parallel execution schemes to reduce total batch execution time by three hours.
- Implemented housekeeping scripts to monitor and maintain execution logs.
- Worked on ETL development for 3G/4G implementation. Modified ETL routines and revised a value-added service (VAS) data model following new business dynamics, i.e., new dimensions derived by tax structure, 3G/4G sites, and 3G/4G data users.
Senior ERP Developer
National University of Sciences and Technology (NUST)
- Designed and developed a payroll module and telephone extension billing module as part of an internal ERP application managed by the NUST ICT department.
- Collaborated with technical and commercial groups to understand and document their requirements by analyzing existing workflows and procedures.
- Reviewed and rewrote older programs to increase operating efficiency, enhance the customer experience, and adapt to new requirements. Investigated and resolved open bugs.
ERP Developer
Government of Pakistan
- Designed a material procurement module for an ERP application developed and managed internally by the department.
- Designed an inventory management module for an ERP application developed and managed internally by the department.
- Implemented the business logic using Oracle PL/SQL.
Oracle Developer
AI Soft | System Developers
- Conducted a root cause analysis (RCA). Debugged and fixed issues in codes suggested by the client's tickets.
- Implemented changes in business logic and client requirements.
- Developed business reports in Oracle Reports 10g.
Experience
Data Warehouse Modernization
The project deliverables included:
• Migration of existing data from a legacy data warehouse platform to an enterprise data warehouse, including data pipelines developed on the IBM InfoSphere DataStage platform.
• Understanding and documentation of the existing data architecture.
• Transformation of existing data architecture to modern data architecture, including new business requirements.
• Design documentation followed by business reviews and a sign-off.
• ETL development guided by the design document and followed by tech and business user acceptance testing.
• Deployment with handover training sessions to AMS and operations teams.
• Technical training on new data models for business.
• Data quality framework development and implementation to ensure data quality concerning the enterprise data warehouse.
Advanced Data Analytics Platform
Data Warehouse Optimization
I led the optimization of the ETL batch and saved two hours. We started with the implementation of logging for each job to find out which one was taking the longest.
To optimize query executions, we performed these actions:
• Partitioning and subpartitioning of tables.
• Creating proper local and global indexes.
• Gathering stats.
• Applying data retention.
• Rewriting queries.
• Developing materialized views for the business.
Migration of ETL Logic from ADF to Azure Databricks
• Develop pipelines in Azure Databricks using PySpark notebooks to ingest data from multiple sources, including Azure Eventhub, AWS Containers, SFTP, and MongoDB.
• Save data as parquet files in the enterprise landing layer, i.e., Azure Data Lake Storage
• Transform data and sink it into a dimensional model implemented in Azure Synapse Analytics
Data Migration
Data Warehouse Enhancements
I led a team of data engineers to implement changes in the data model by introducing new facts and dimensions, followed by changes in ETL pipelines to sink data into a new model. We also develop new ETL pipelines to ingest 3G/4G network equipment data sources. BI reports were also created per company requirements to report 3G/4 G-related business KPIs.
Data Warehouse Data Quality Roadmap Implementation
As the lead data engineer, I analyzed why so many issues are being reported daily.
We learned that 70% of issues are related to source data and can be avoided if data quality checks are implemented on the landing layer. So we implemented around 150 reviews on the landing layer to check the quality of source data with acceptance threshold to variation in trend.
Since implementation, the ops team and business team have been at peace.
Data Architect at PMI AMS Team
I was responsible for the architectural enhancement of all existing data products. The main supporting areas were the Sales and Digital Sustainability Program for any architectural enhancement and modifications.
My contribution:
• Data vault 2.0 design approach for integrated data platform of Data Ocean.
• WhereScape 3D for data vault designing.
• Standardized the implementation framework for data pipelines.
• The delivery of trusted data products to enable data-driven decision-making for the business.
Education
Master's Degree in Computer Science
International Islamic University Islamabad (IIUI) - Islamabad, Pakistan
Certifications
Microsoft Certified: Azure Fundamentals
Microsoft
Microsoft Certified: Azure Data Engineer Associate
Microsoft
Hands On Essentials - Data Warehouse
Snowflake
DP-200 Implementing an Azure Data Solution
Microsoft
Skills
Libraries/APIs
PySpark
Tools
IBM InfoSphere (DataStage), Oracle Exadata, SPSS Modeler, Amazon Athena
Paradigms
Database Design, Dimensional Modeling, ETL, Database Development, OLAP, Business Intelligence (BI), Data Science
Storage
PL/SQL, Relational Databases, Data Pipelines, Databases, Database Modeling, Oracle PL/SQL, Data Definition Languages (DDL), SQL Stored Procedures, OLTP, JSON, PostgreSQL, MySQL, Azure SQL Databases, Database Security, Azure Cloud Services, Data Lakes, MongoDB, Oracle 11g, Amazon S3 (AWS S3)
Languages
SQL, Data Manipulation Language (DML), Stored Procedure, Python, Snowflake, Python 3
Platforms
Oracle, Oracle Database, Databricks, Amazon Web Services (AWS), Azure Synapse, Azure, Azure SQL Data Warehouse, Azure Event Hubs, Dedicated SQL Pool (formerly SQL DW)
Frameworks
Spark
Other
Data Engineering, Oracle Forms & Reports, Data Analysis, Star Schema, Data Analytics, Data Modeling, Performance Tuning, Query Composition, Datasets, Data Profiling, Data Cleaning, Data Cleansing, Modeling, PL/SQL Tuning, Query Optimization, Normalization, Data, Data Warehousing, Data Warehouse Design, Data Management, Migration, Data Architecture, erwin Data Modeler, Data Migration, Azure Databricks, Technical Documentation, Data Wrangling, Parquet, Programming, Data Visualization, Linear Regression, GSM, Pipelines, Azure Data Lake, Data Security, Technical Writing, Azure Data Factory, Big Data, Cloud, Data Processing, Azure Stream Analytics, Cloud Security, Storage, PySQL, WhereScape, Data-level Security
How to Work with Toptal
Toptal matches you directly with global industry experts from our network in hours—not weeks or months.
Share your needs
Choose your talent
Start your risk-free talent trial
Top talent is in high demand.
Start hiring