Bukhbayar (Buku) Purevsuren, Developer in Sydney, New South Wales, Australia
Bukhbayar is available for hire
Hire Bukhbayar

Bukhbayar (Buku) Purevsuren

Verified Expert  in Engineering

Data Engineer and Developer

Location
Sydney, New South Wales, Australia
Toptal Member Since
June 24, 2020

Buku is a data engineer who has 11 years of experience in IT including data and software engineering and full-stack development in eCommerce, financial, and sports industries. He's done data modeling, designed data warehouses, and architected data analytics solutions on the cloud. Skilled in Nix OS, CI/CD, DevOps practices, web scraping, data collection transformation, and data engineering techniques, Buku is a self-starter who hits the ground running, especially picking up alternative tools.

Portfolio

Xendit Incorporated
Databricks, PySpark, Delta Lake, Apache Airflow, Presto, Trino...
Albayanat Alwasphiah LTD
Data Lakes, Data Lake Design, Data Modeling, Data Architecture, Metabase...
Recko
Amazon Web Services (AWS), AWS Lake Formation, PostgreSQL, SQL, Hue, AWS Glue...

Experience

Availability

Part-time

Preferred Environment

Amazon Web Services (AWS), Zsh, Docker, Azure, Slack, Command-line Interface (CLI), Visual Studio Code (VS Code), Unix

The most amazing...

...project I've built is the data warehouse, BI system, along with all of the data integration for a bank's new core banking system.

Work Experience

Data Engineer

2021 - 2022
Xendit Incorporated
  • Developed a data quality report that reconciled billing data across different platform that has different production versions. The platform was developing a new version but at a same time still maintaining the old version with old customers.
  • Worked in Agile team with 2-week sprints, Ant tracking, task management in Jira, and document storage in Confluence.
  • Developed 12 main data quality reports as well as building a pipeline to generate billing invoices for customers.
  • Performance tuned existing and new Databricks notebooks and SQL queries due to change of data engineering team changing PrestoDB to TrinoDB and implementing delta lake.
Technologies: Databricks, PySpark, Delta Lake, Apache Airflow, Presto, Trino, Amazon Web Services (AWS), Data Lakes, Data Lake Design, Data Quality, SQL, Redash, Jira, Confluence, Python, Spark, ETL Implementation & Design

Senior Data Engineer

2021 - 2021
Albayanat Alwasphiah LTD
  • Worked as a senior data engineer—driving and supporting an existing data team, data modeling in the data lake for BI while working on KPI requirements and architecting data lake and data pipeline on-premise.
  • Used GitHub as a source tracking tool, but my main focus was modeling and guiding the existing team so it was mainly POC scripts that I worked on.
  • Modeled a data lake in a star schema model conceptually and mentored a data scientist and data engineers in the team for testing KPIs based on the created model; also modeled data, trained the team, and prepared POC scripts.
Technologies: Data Lakes, Data Lake Design, Data Modeling, Data Architecture, Metabase, DocumentDB, Data Warehouse Design, Data Warehousing, Star Schema, ETL Implementation & Design

Data Engineer

2020 - 2020
Recko
  • Gathered business requirements and documented and finalized the microservice API contracts.
  • Architected the data platform for multi-tenancy—building a highly scalable data pipeline and data lake and documenting and presenting to the teams and stakeholders.
  • Modeled a data lake and data warehouse including reporting data marts; also did modeling in a data vault and star schema.
  • Developed a Spark application in Scala, Spark, and AWS EMR; debugged and deployed in Hue and an Oozie workflow.
  • Built stored procedures and optimized queries in Amazon Redshift.
Technologies: Amazon Web Services (AWS), AWS Lake Formation, PostgreSQL, SQL, Hue, AWS Glue, Amazon S3 (AWS S3), Oozie, Docker, Redshift, Data Lake Design, Data Lakes, Spark, Amazon Elastic MapReduce (EMR), Scala, Database Architecture, Data Warehouse Design, Data Warehousing, Star Schema, ETL Implementation & Design

Data and DevOps Engineer (Remote)

2020 - 2020
TomYo
  • Deployed a Kubernetes cluster on AWS EKS along with handling maintenance and monitoring.
  • Developed different Python functions and deployed them in AWS Lambda with a Serverless framework.
  • Automated a CI/CD pipeline with the Bamboo pipeline and Terraform.
  • Built a data lake in S3 and developed a pipeline in Apache Airflow.
Technologies: Spark, Apache Airflow, Amazon Web Services (AWS), Amazon EC2, Serverless, AWS Lambda, Python, Terraform, Amazon Elastic Container Registry (ECR), Amazon EKS, Docker, Kubernetes, Database Architecture, Data Warehouse Design, Data Warehousing, ETL Implementation & Design

Data Engineer

2018 - 2020
National Rugby League
  • Remodeled and architected an on-premise data warehouse on Azure Synapses and built a data integration from scratch on Azure Data Factory.
  • Developed a record-linkage application on Azure Databricks (PySpark) and integrated it into a data warehouse and a Salesforce marketing cloud.
  • Built a fully automated data integration pipeline on Azure Data Factory that integrated with Azure Databricks, Azure Polybase, Azure Function, Azure Logic Apps, and EventHubs.
  • Documented a solution on Jira Confluence, as well as used the markup language in Bitbucket for the developer's documentation.
  • Used Jira and Bamboo tools to build a CI/CD pipeline and track issues in an Agile team.
Technologies: Amazon Web Services (AWS), Data Flows, Azure Data Factory, PySpark, Azure Synapse, SQL, C#, Python, Synapse, Databricks, Azure Functions, Azure Logic Apps, ADF, Azure, Spark, Database Architecture, Data Warehouse Design, Data Warehousing, Azure SQL Databases, ETL Implementation & Design

Full-stack Developer

2016 - 2018
Stelly Clothing
  • Developed new features on the website and integrated it into Mailchimp, Mailgun, the Australian Post, Google and Facebook Analytics, Commission Factory, and many other third-party platforms.
  • Created new plugins on the CMS for affiliate marketing, referral marketing, customer segmentation, Instashop, personalized pages, customer behavior tracking, gift cards, abandoned cards, and many others.
  • Deployed the website into AWS EC2 and maintained multiple testing and staging services and a MySQL database on AWS RDS.
  • Used GitHub for source and issue tracking and Trello for planning and task tracking.
  • Worked alone, planned the management time, and took responsibility.
Technologies: Python, Amazon S3 (AWS S3), Amazon EC2, jQuery, JavaScript, HTML, MySQL, PHP, PrestaShop

Senior Data Engineer | BI Engineer

2012 - 2015
Golomt Bank of Mongolia
  • Architected data analytics and data warehouse systems with a team of two junior data developers led by me.
  • Modeled a data warehouse in Oracle 11g and modeled in many other data marts for different purposes (e.g., star schema modeling for OBIEE).
  • Managed stakeholders by myself, taking systems requirements from the head of business department, and planning and assigning tasks to the team using Jira.
  • Built a data integration pipeline from core banking, middleware (API), MongoDB, MySQL, MSSQL, Oracle, and many others.
  • Oversaw the development process and deadline, along with helping and training the team.
  • Tuned and maintained Big SQL queries, BI system, and data warehouse for replication and backups.
  • Developed internal tools by myself for automating the development process, documentation, and data dictionary using the Java Play framework.
Technologies: Unix, SQL Server Integration Services (SSIS), MongoDB, MySQL, Microsoft SQL Server, Java, C#, .NET, Oracle Data Integrator (ODI), Oracle BI Publisher, Oracle Business Intelligence Enterprise Edition 11g (OBIEE), Oracle 11g, Database Architecture, Data Warehouse Design, Data Warehousing, Oracle, Star Schema, ETL Implementation & Design

Database Administrator (DBA)

2011 - 2012
Golomt Bank of Mongolia
  • Maintained and tuned reporting databases, backups, and replication.
  • Planned new tasks and tracked current system issues on Jira.
  • Researched BI and data warehouse systems and compared OBIEE, Microsoft Business Intelligence Studio, and Pentaho for a future BI system for the bank.
  • Integrated a new core banking system and planned the migration of the current reporting system to a proper data warehouse.
Technologies: Oracle GoldenGate, Data Vaults, MySQL, MongoDB, Microsoft SQL Server, Oracle 11g, Oracle

Software Engineer

2008 - 2011
Golomt Bank of Mongolia
  • Took ownership of the Grammis reporting system using SQL and Crystal Reports.
  • Maintained the system and developed many new reports.
  • Gathered reporting requirements from the finance department and managed stakeholders.
  • Tracked new reporting requirements in Jira, trained the customer to use the reporting system, and presented new features on the reporting system.
  • Performed a system analysis on the finance system (assets, payroll, inventory, and HR) that later over 100 branches used.
  • Developed assets, payroll, and inventory modules of a finance system using .NET C# WinForms.
  • Planned and tracked issues in Jira, using Microsoft team foundation and SVP for source control.
  • Worked on a SOAP integration with the Central Bank of Mongolia for central loan information.
Technologies: SOAP, XML, JavaScript, HTML, Symfony, PHP, Play Framework, Java, C#, .NET, Microsoft SQL Server, Oracle

Junior C# Developer

2007 - 2008
Shandas Soft (Mongolia)
  • Worked as a junior C# developer on the finance system during my last year at university.
  • Implemented a system analysis and database modeling of the assets, inventory, payroll, and human resource modules.
  • Planned our tasks on a Redmine open-source project management system and used SVN for source control.
  • Developed a .NET WinForm application using C# and Oracle Database.
Technologies: Stored Procedure, PL/SQL, XML, Windows Forms (WinForms), .NET, C#, Oracle

Re-architecture and Migration of an Existing On-premise Data Analytics Platform on Azure Stack

Throughout this project, I:

• Re-architected the infrastructure.
• Remodeled a data warehouse (data vault) and a data mart (star schema) on Azure Synapses.
• Built a data integration pipeline on Azure Data Factory.
• Developed a record-linkage (fuzzy matching) application on Azure Data Bricks.
• Integrated APIs and the Azure Polybase.

Automation of a CI/CD Pipeline and Deployment of a Kubernetes Cluster on the Product

http://www.tomyo.mn
Throughout this project, I:

• Automated a CI/CD pipeline on Bamboo with Terraform.
• Deployed a Kubernetes cluster on AWS EKS along with horizontal auto-scaling, load balancing, and monitoring.
• Developed Python functions and the deployment to AWS Lambda with a Serverless framework.

Development of New Features on the Web, Data Collection, Maintenance of Server and Deployment on AWS

http://www.stelly.com.au
ROLE: Full-stack Developer

TASKS ACCOMPLISHED:
• Developed new modules and features on an existing eCommerce website.
• Built from scratch of several modules on PrestaShop CMS, including the referral and affiliate modules.
• Implemented the customer segmentation and integrated it into Mailchimp.
• Developed the gift card and voucher code.
• Built a data integration module for Google Analytics, Facebook, and so on.
• Developed a customer-behavior-tracking module.
• Implemented page personalization.
• Enabled reporting.

Creation of a Data Warehouse From Scratch and Implementation of an Oracle BI System at Golomt Bank

While employed at Golomt Bank, I worked as a new member of the core banking migration team. Due to the cost issue, the bank decided to buy a core banking system without the BI module and data warehouse.

I led a team of two and built a data warehouse on Oracle 11g and OBIEE for a data analytics system.

TASKS ACCOMPLISHED:
• Modeled a data warehouse (OLTP and OLAP).
• Modeled a data mart for OBIEE (star schema).
• Architected an on-premise solution with a custom data integration tool, SSIS, and ODI.
• Gathered customer requirements from the head of the finance department, then planned and managed our technical tasks on Jira.
• Trained a team as a technical lead and trained not-tech-savvy stakeholders about the usages and benefits of a BI system.
• Built a data pipeline on the .NET C# customer tool, SSIS, and ODI (later on).
• Developed a dynamic data dictionary of the entire internal system, which helps not-tech-savvy customers use our BI system and create reports for themselves, reducing ad-hoc requests.
• Integrated a fully dynamic pipeline between the general ledger and the BI system, where the finance team can create a formula to generate reports and custom data visualizations.

Development of a WinForm Finance Application in C#

I worked with a team on a finance system that used .NET C# WinForms. I mainly focused on the development of several modules (payroll, assets, inventory) and the integration of the core banking system by the end of the day (EOD).

TASKS ACCOMPLISHED:
• Performed a system analysis on modules (payroll, assets, and inventory).
• Developed the modules primarily.
• Planned out an EOD process for the core banking integration.
• Provided maintenance and testing.

Languages

T-SQL (Transact-SQL), SQL, Java, PHP, HTML, XML, Stored Procedure, Scala, Python, C#, JavaScript

Paradigms

ETL Implementation & Design, ETL, Agile, REST, Azure DevOps

Storage

Data Pipelines, Database Architecture, Data Lake Design, Redshift, PostgreSQL, SQL Server DBA, Azure Cloud Services, Azure SQL, Oracle 11g, Microsoft SQL Server, Azure SQL Databases, Amazon S3 (AWS S3), SQL Server Integration Services (SSIS), NoSQL, MySQL, MongoDB, PL/SQL, Data Lakes

Other

Data Modeling, Software Engineering, Query Optimization, Azure Data Factory, Data Engineering, Cloud, Performance Tuning, PolyBase, AWS Lake Formation, Shell Scripting, Data Warehousing, Build Pipelines, Data Architecture, ELT, Data Warehouse Design, Unix Shell Scripting, Star Schema, Data Flows, Command-line Interface (CLI), Data Vaults, SOAP, Serverless, Metabase, DocumentDB, Delta Lake, Data Quality, ETL Tools, APIs

Tools

Terraform, Amazon EKS, AWS Glue, Visual Studio, Bitbucket, Bamboo, Jira, *nux Shells, Slack, Zsh, Azure Logic Apps, Synapse, PrestaShop, Oracle Business Intelligence Enterprise Edition 11g (OBIEE), Oracle BI Publisher, Oracle GoldenGate, Amazon Elastic Container Registry (ECR), Amazon Elastic MapReduce (EMR), Oozie, Hue, GitHub, GitLab, Apache Airflow, Redash, Confluence, Microsoft Power BI, Azure App Service

Platforms

Kubernetes, Azure Synapse, Linux, Azure, Databricks, Docker, Amazon EC2, Oracle, Amazon Web Services (AWS), Unix, Azure Functions, Oracle Data Integrator (ODI), AWS Lambda, Visual Studio Code (VS Code), Azure Event Hubs

Frameworks

ADF, .NET, Symfony, Play Framework, Laravel, Spark, Apache Spark, Presto, Trino

Libraries/APIs

jQuery, Windows Forms (WinForms), Liquibase, PySpark

2016 - 2018

Master's Degree in Information Systems

Central Queensland University - Sydney, Australia

2005 - 2008

Bachelor's Degree in Computer Science

Mongolian University of Science and Technology - Ulaanbaatar, Mongolia

AUGUST 2019 - PRESENT

Data Science

Institute of Data

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring