Hao Luo, Developer in London, United Kingdom
Hao is available for hire
Hire Hao

Hao Luo

Verified Expert  in Engineering

Bio

Hao is a data engineering manager who's led complex and innovative projects in the fields of banking/investment. When a new challenge arises, Hao can always identify paths to solutions while dealing with uncertainties. Although Hao's main focus is in data engineering, he has broad knowledge from full-stack development to DevOps engineering.

Portfolio

Top UK Financial Institution
Data Marts, Data Lakes, Quality Assurance (QA), Data Cleansing, Data Cleaning...
Top UK Financial Institution
Data Marts, Data Lakes, Quality Assurance (QA), Data Cleansing, Data Cleaning...
University of Greenwich
Data Cleansing, Data Cleaning, REST APIs, Database Modeling, SQL, Python...

Experience

  • Python - 8 years
  • SQL - 8 years
  • Data Engineering - 6 years
  • Azure - 4 years
  • Hadoop - 4 years
  • Spark - 4 years
  • Docker - 4 years
  • Google Cloud Platform (GCP) - 2 years

Availability

Full-time

Preferred Environment

Python, Azure, Google Cloud Platform (GCP), React, Go, Hadoop, Spark, Python 3, Kubernetes, Docker

The most amazing...

...thing I've done was to redesign a large data system to fit with the project's modern architecture.

Work Experience

Data Engineering Manager

2018 - PRESENT
Top UK Financial Institution
  • Delivered six financial data ELT (extract, load, transform) data projects with reporting visualization. The data comes from multiple sources such as APIs, folders, and emails.
  • Designed and developed a high-performance time-series aggregation engine to support highly complex equation aggregation rules with custom functions. It supports different data layers such as files, RDBMS, Parquet, Hive, BigQuery, etc.
  • Reduced the data point modeling (XBRL) schema update effort from three months every year to only one week by delivering a low-maintenance solution.
  • Reduced the cost of training new developers by developing a chatbot with BERT.
  • Designed the organization's first cloud big data solution with an options paper (Google Cloud and Azure).
  • Supported two high-performance data engineers who are now technical leads.
Technologies: Data Marts, Data Lakes, Quality Assurance (QA), Data Cleansing, Data Cleaning, Complex Data Analysis, Machine Learning Operations (MLOps), DevOps, REST APIs, Apache Hive, Big Data, Google BigQuery, Data Modeling, Database Modeling, SQL, ELT, Python, Data Analysis, Data Pipelines, ETL, Data Architecture, Data Engineering, Machine Learning, MongoDB, Apache Airflow, Flask, Azure, Google Cloud Platform (GCP), PostgreSQL, Hadoop, Spark, Microsoft Power BI, Team Leadership, Azure Data Factory, Architecture, Data Visualization, Data Aggregation, ETL Tools, Snowflake, Data Warehousing, Performance Tuning, Database Migration, Firebase, NoSQL

Data Engineer

2014 - 2018
Top UK Financial Institution
  • Optimized a data processing pipeline that is now five times faster.
  • Delivered a critical functional change to an end-user application within one day of the request from production.
  • Built an automation functionality in a time series data system to save three days of business effort every quarter.
  • Developed a continuous integration and deployment pipeline that saved five hours of manual build and deployment every month.
  • Led the delivery of a data analysis reporting sharing platform by using React and .NET Core.
Technologies: Data Marts, Data Lakes, Quality Assurance (QA), Data Cleansing, Data Cleaning, Complex Data Analysis, SharePoint, Machine Learning Operations (MLOps), DevOps, REST APIs, Apache Hive, Big Data, Google BigQuery, Data Modeling, Database Modeling, SQL, ELT, Python, Data Analysis, Data Pipelines, ETL, Data Architecture, Flask, Data Engineering, MongoDB, Apache Airflow, React, CSS, HTML, PostgreSQL, Microsoft SQL Server, C#, Spark, Architecture, Data Visualization, Data Aggregation, ETL Tools, Snowflake, Data Warehousing, Performance Tuning, Database Migration, Firebase, NoSQL

Software Developer

2012 - 2014
University of Greenwich
  • Developed an inventory management system by using C# and WPF to replace Excel.
  • Reduced the asset support daily effort from three hours to ten minutes by establishing a SharePoint workflow to automate the asset support process.
  • Built a motivation system to encourage individuals to join events by using C# and MVC.
Technologies: Data Cleansing, Data Cleaning, REST APIs, Database Modeling, SQL, Python, MVC Razor, Windows Presentation Foundation (WPF), ASP.NET, CSS, HTML, SharePoint

Software Developer

2011 - 2012
Qin Dynasty Ltd.
  • Delivered an automatic postage tracking system by consuming APIs.
  • Automated the dispatch process by integrating the APIs from eBay, Amazon, and other custom shopping portals.
  • Developed a web scraping tool to obtain competitor information.
Technologies: Python, Windows Forms (WinForms), ASP.NET, C#

AI-driven Social Distancing Tool

A tool developed to help people live in London to understand the congestion of the street before going out for the purpose of maintaining social distance.

It uses:
1. Python and Flask as the web API to serve data.
2. Python and Image AI as the image process engine.
3. React as the front end.
4. MongoDB as the data storage.

Big Data Cluster Docker Images

http://github.com/sporule/big-data-cluster
A big data cluster that can run in a development laptop. It simulates multiple nodes in a cluster without the need for a cluster. This makes the development and testing journey much easier compared to a standalone development environment.

It supports key big data components:
• Airflow
• Hadoop
• Hive
• Spark
• Nifi
• Livy
• Kafka

Data Basket

http://github.com/sporule/data-basket
A test data generation tool by using Go. It can generate random data and relational data by using a configuration with regular expression. It can generate 1 million rows of relational test data in 16 seconds in a machine with four logical CPU cores.

Sporule

A simple static blog system that can dynamically render markdown files. The system was built by using React. Three Markdown-related plugins are also published to the NPM.

Key Features:
1. RSS output
2. Schedule article publishing
3. Theme integration
4. Offline search
5. Progressive web app

Werewolf Chatbot

This is a social game built by using Python and Telegram Bot API. It plays the role of judge for the full Werewolf game and supports unlimited players by customizing the roles and rules. It implements a flexible modular design.

Grater | Distributed Web Scraping Tool

https://github.com/sporule/grater
A distributed web scraping tool by using Go, Gin, and Colly.

Key Features:
1. It supports unlimited scale-out worker nodes.
2. It supports transformation and validation in the process of scraping.
3. It supports complex custom scraping rules.
4. All events are configuration based.
5. It supports automatic proxy obtaining through scraping and APIs.
2016 - 2019

Master's Degree in Central Banking and Financial Regulation

Warwick Business School - London, United Kingdom

2010 - 2014

Bachelor's Degree in Computer Science

University of Greenwich - London, United Kingdom

JANUARY 2021 - JANUARY 2023

Azure Solutions Architect Expert

Microsoft

NOVEMBER 2020 - NOVEMBER 2022

Professional Cloud Architect

Google Cloud

OCTOBER 2020 - OCTOBER 2022

Professional Data Engineer

Google Cloud

JUNE 2013 - PRESENT

ITIL Foundation Certification

British Computer Society

Libraries/APIs

REST APIs, React, Windows Forms (WinForms), Telegram Bot API

Tools

Tableau, Microsoft Power BI, Apache Airflow, Webpack

Languages

C#, Python, SQL, Snowflake, HTML, CSS, JavaScript, R, Go, Java, C++, UML

Frameworks

Spark, Hadoop, Flask, ASP.NET, Windows Presentation Foundation (WPF), MVC Razor, Gin-Gonic

Paradigms

ETL, Web Architecture, DevOps

Platforms

Docker, Google Cloud Platform (GCP), Azure, SharePoint, Firebase, Kubernetes

Storage

Data Pipelines, Database Modeling, Database Migration, NoSQL, PostgreSQL, Apache Hive, Data Lakes, Microsoft SQL Server, MongoDB

Other

Data Engineering, ELT, Data Modeling, Big Data, Data Cleaning, Data Cleansing, Team Leadership, Architecture, Data Analytics, ETL Tools, Data Warehousing, Data Architecture, Data Analysis, Google BigQuery, Complex Data Analysis, Quality Assurance (QA), Data Marts, Dashboards, Data Visualization, Azure Data Factory, Data Aggregation, Performance Tuning, Finance, Credit Risk, Leadership, Risk, Economics, Projects, ITIL V3 Foundation Certified, Machine Learning, Chatbots, Scraping, APIs, Machine Learning Operations (MLOps)

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring