Innocent Musanzikwa, Developer in Calgary, AB, Canada
Innocent is available for hire
Hire Innocent

Innocent Musanzikwa

Verified Expert  in Engineering

Bio

Inno is a seasoned data engineer and developer who's worked at IRI—a top retail data analytics company—in Africa and North America for the past decade and as a freelance consultant for the past couple of years. As a SQL and ETL developer, he has created quality data warehouses using industry-standard techniques like Kimball and DataVaults. As a data engineer, Inno has built highly robust and scalable data pipelines both on-premise and on the cloud using several latest cutting-edge technologies.

Portfolio

Darwill, Inc.
SQL, Tableau Development, Python, Data Engineering, Data Science, ETL...
SFL Scientific LLC
SQL, SSIS, MariaDB, SQL Server, Data Transformation, Python...
Airiam Holdings, LLC
Business Intelligence Development, SQL, APIs, SQL Server, Dimensional Modeling...

Experience

Availability

Part-time

Preferred Environment

SQL, PySpark, Python, Hadoop, Azure Data Factory, Data Warehousing, Snowflake, Databricks, Amazon Web Services (AWS), Informatica

The most amazing...

...big data warehousing and data integration solution I've designed—using Python, SQL, ADF, Hadoop, Hive, and Spark—won an RFP in Canada out of six competitors.

Work Experience

Data Engineer

2022 - 2024
Darwill, Inc.
  • Built Tableau dashboards and visualizations using AWS Redshift and Aurora databases.
  • Created AWS Lambda functions running Python for custom ETL tasks and ad-hoc requests.
  • Managed AWS Redshift and Aurora databases and designed data warehouses and data migrations.
  • Redesigned the client's data warehouse using the AWS tech stack and improved their migration process by introducing federated queries and Lambda functions running Python pipelines, as well as overhauling their Tableau dashboards.
Technologies: SQL, Tableau Development, Python, Data Engineering, Data Science, ETL, Data Warehouse, AWS, Relational Databases, Data Science, Data Science, Database, PostgreSQL, AWS Lambda, Database, Data Visualization, Dedicated SQL Pool (formerly SQL DW), Azure SQL Data Warehouse, Database Modeling, MySQL, Entity Relationships, Business Analysis, Database Design, Informatica, Databricks, Snowflake

Data Engineer

2022 - 2022
SFL Scientific LLC
  • Consulted on an existing SSIS poorly designed data integration project and helped identify bottlenecks and inefficiencies.
  • Redesigned the existing data pipeline using SSIS to be efficient and scalable.
  • Performed SQL tuning and SQL code review for process efficiencies.
Technologies: SQL, SSIS, MariaDB, SQL Server, Data Transformation, Python, Database Schema Design, iPaaS, CI/CD Pipelines, Relational Databases, Stored Procedure, Data Analysis, T-SQL, SQL DML, Database, Data Science, Data Visualization, Dedicated SQL Pool (formerly SQL DW), Azure SQL Data Warehouse, Database Modeling, Entity Relationships, Tableau Development, Business Analysis, Database Design

BI and Data Warehouse Expert

2021 - 2022
Airiam Holdings, LLC
  • Designed and developed data pipelines to integrate data from Quickbooks API, Sage Intacct API, and spreadsheets into Azure SQL.
  • Designed and developed a data warehouse in Azure SQL.
  • Designed and created business reports and KPI dashboards using Power BI.
  • Developed complex SQL scripts to manage data transformations and speed up integration.
Technologies: Business Intelligence Development, SQL, APIs, SQL Server, Dimensional Modeling, Relational Databases, Business Intelligence Development, Cloud Engineering, Git, REST API, Synapse, DAX, Dashboard Design, Dashboard, Stored Procedure, Tableau Development, Data Analysis, T-SQL, SQL DML, Database, Data Science, Microsoft Power Automate, Data Visualization, Database Modeling, Entity Relationships, Business Analysis, Database Design

Data Analyst for Migration Project

2021 - 2021
JLL - JLLT Data
  • Developed the data pipeline to integrate data from Salesforce to Microsoft SQL.
  • Designed advanced SQL code, e.g., CTE, stored procedures, and functions to manage data transformations.
  • Performed SQL tuning to improve ETL efficiencies and process scalability.
  • Consulted on standard operating procedures and best case scenarios.
Technologies: SQL, T-SQL, ETL, Salesforce Design, Data Migration, Relational Databases, Business Intelligence Development, SQL Server, Stored Procedure, Data Analysis, Google Sheets Development, SQL DML, Database, Data Science, Database Modeling, Entity Relationships, Tableau Development, Business Analysis, Database Design

Director | Data Engineering

2019 - 2021
IRI
  • Developed Azure Data Factory pipelines to integrate data from Apache Hive, HDFS, OAuth 2 APIs, and various flat-file types into Azure SQL.
  • Managed a team of onshore and offshore big data developers, assigning tasks and tracking the progress on Jira.
  • Oversaw data strategy and recommendations for new data sources and ongoing projects.
  • Mentored big data engineers to help them develop their skills.
  • Architected new data models and upgraded old data warehouses as per client request or technology change.
Technologies: Python, Hadoop, Hadoop, Azure Synapse, Azure Data Factory, Bash Script, SQL, SQL, Databricks, Data Engineering, ETL, Data Modeling, Database, Azure Design, Data, Data Architecture, Business Intelligence Development, Database, Apache Airflow, Data Integration, Big Data Architecture, T-SQL, Data Migration, Snowflake, Data Build Tool (dbt), Apache Kafka, ELT, SSIS, Data Transformation, Dimensional Modeling, Relational Databases, Business Intelligence Development, Cloud Engineering, SQL DML, Database, Azure SQL Data Warehouse, Dedicated SQL Pool (formerly SQL DW), Database Modeling, Entity Relationships, Database Design, Informatica

ETL Architect

2016 - 2019
IRI
  • Developed SQL-based data warehouses on-premise and on the cloud.
  • Integrated various data sources from flat files to cloud-based data sources like Snowflake, AWS and data lakes into Azure Data Warehouse, and Apache Hive on Hadoop.
  • Created scalable data pipelines and improved efficiencies on the existing ones.
  • Trained and upskilled new data developers and participated in code reviews.
  • Maintained system documentation of all business data components and strategies.
Technologies: SSIS, Azure Synapse, Azure Data Factory, Databricks, PySpark, SQL, Oracle Development, Hadoop, Hadoop, Data Warehouse, Data Engineering, ETL, Data Modeling, SQL, Database, Data, Data Architecture, Business Intelligence Development, Database, Data Integration, Big Data Architecture, BigQuery, JavaScript, T-SQL, Data Migration, Snowflake, AWS, Amazon EMR, ELT, APIs, Data Transformation, MariaDB, SQL Server, Dimensional Modeling, Relational Databases, Business Intelligence Development, Cloud Engineering, REST API, SQL DML, Database, Azure SQL Data Warehouse, Dedicated SQL Pool (formerly SQL DW), Database Modeling, Entity Relationships, Performance Tuning, SQL

SQL Lead Developer

2012 - 2016
IRI
  • Developed SQL-based data warehouses and data marts.
  • Wrote SQL queries to provide data for SSRS reports.
  • Used SSIS, Talend, and DataStage for ETL processes depending on the client's requirements.
  • Created custom business reports using SQL Server Reporting Services (SSRS).
  • Managed junior developers and ran stand-up development meetings.
Technologies: SQL, SSIS, SQL Server, PSQL, MySQL, Data Warehouse, Data Engineering, ETL, Data Modeling, SQL, Database, Data, Data Architecture, Business Intelligence Development, Database, Data Integration, Big Data Architecture, T-SQL, Data Migration, ELT, Data Transformation, Dimensional Modeling, Relational Databases, Business Intelligence Development, REST API, SSAS, Dashboard Design, Dashboard, SQL DML, Database, SSRS Reports, Azure SQL Data Warehouse, Dedicated SQL Pool (formerly SQL DW), Database Modeling, SQL Server 2015, Entity Relationships, Business Analysis, Performance Tuning, SQL

SQL/ETL Developer and Consultant

2010 - 2012
Mi9 Retail (formerly JustEnough Software Corporation)
  • Managed SQL replication between mobile devices and SQL Server.
  • Created SQL data warehouses using the Kimball methodology for reporting purposes.
  • Designed and developed ETL packages using SQL Server Integration Services (SSIS).
  • Designed and developed reports in SQL Server Reporting Services (SSRS).
  • Performed database tuning and code reviews for any code being deployed to production.
Technologies: SQL, SSIS, SQL Server, SQL Server, Data Engineering, ETL, Data Modeling, SQL, Database, Data, Data Architecture, Business Intelligence Development, Database, Data Integration, Big Data Architecture, T-SQL, Data Migration, Data Transformation, Relational Databases, Business Intelligence Development, SSAS, SQL DML, Database, SSRS Reports, Database Modeling, SQL Server 2015, Entity Relationships

Data Migration from Azure SQL to Snowflake

https://github.com/innowarue/ADF
This project involved migrating data from an Azure SQL database to a Snowflake data warehouse using an Azure Data Factory data pipeline. It took me minutes to create it based on my skill set and proficiency in Data Factory.

I replaced the authentic data sources with my Azure and Snowflake accounts to make the project publicly available without compromising confidentiality.

Data Integration from OAuth2 API

I created an automated data pipeline to integrate data accessible via an OAuth2-based API in JSON format into a cloud-based data warehouse solution. The solution used Python and Spark on Databricks integrated into an Azure Data Factory pipeline.

SQL Server Replication to Mobile Devices

I created a replication system that synced data between mobile devices and Microsoft SQL Server. Field sales representatives would collect information from the field, upload it to SQL Server using SQL CE and download any updates from SQL Server via the mobile replication I set up.

In-place Data Integration for an Acquisition

I created an in-place ETL integration for a company acquisition and merger, bringing the two companies' data into a single warehouse while continuously delivering weekly reports to the client services and retail service teams.

Kafka Streaming and Data Integration

I created an automated data pipeline to integrate data accessible via a Kafka stream, ingesting it into Spark Streaming using Spark and Python and loading it into a Cloudera Hadoop file system accessible using a Hive data warehouse solution.
2013 - 2015

Bachelor's Degree in Information Technology

University of South Africa - Pretoria, South Africa

AUGUST 2023 - AUGUST 2025

Databricks Certified Data Engineer Associate

Databricks

AUGUST 2023 - AUGUST 2025

SnowPro Core

Snowflake

DECEMBER 2020 - DECEMBER 2022

Certified Apache Spark and Hadoop Developer

Cloudera

DECEMBER 2019 - PRESENT

Analyzing Big Data with Hive

LinkedIn Learning

DECEMBER 2019 - PRESENT

Advanced NoSQL for Data Science

LinkedIn Learning

Libraries/APIs

PySpark, REST API, Spark Streaming

Tools

Business Intelligence Development, Tableau Development, BigQuery, Synapse, SSAS, Apache Airflow, Amazon EMR, Git, Google Sheets Development

Languages

SQL, Python, Bash Script, T-SQL, Snowflake, Stored Procedure, SQL DML, Scala, JavaScript, Bash

Frameworks

Hadoop, Spark, Windows PowerShell, ADF

Paradigms

ETL, Business Intelligence Development, Dimensional Modeling, Database, Database Design

Platforms

AWS, AWS Lambda, Azure SQL Data Warehouse, Dedicated SQL Pool (formerly SQL DW), Azure Design, Microsoft Power Automate, Azure Synapse, Oracle Development, Databricks, Apache Kafka, Salesforce Design, Zeppelin

Storage

Hadoop, MySQL, SSIS, SQL Server, PSQL, SQL Server, SQL, PostgreSQL, Database, Database, Data Integration, Relational Databases, Database, RDBMS, Database Modeling, SQL, NoSQL, SQL Server, Database Replication, SQL, MariaDB

Other

Azure Data Factory, Data Warehouse, Data Analysis, Data Engineering, Data, Data Architecture, Big Data Architecture, Data Migration, ELT, Data Warehouse, Data Transformation, Database Schema Design, ETL Tools, Scripting Languages, Data Science, Data Visualization, SSRS Reports, SQL Server 2015, Entity Relationships, Business Analysis, Performance Tuning, Informatica, Data Modeling, Cloud Engineering, APIs, Dashboard Design, Dashboard, Web Scraping, Data Build Tool (dbt), iPaaS, CI/CD Pipelines, DAX, Data Science, Data Science, Azure Databricks

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring