Breno Costa Dolabela Dias, Developer in São José dos Campos - State of São Paulo, Brazil
Breno is available for hire
Hire Breno

Breno Costa Dolabela Dias

Verified Expert  in Engineering

Data Engineer and Developer

São José dos Campos - State of São Paulo, Brazil

Toptal member since August 13, 2024

Bio

Breno is a data engineer with six years of experience, specializing in data engineering for the last three years. His skills encompass infrastructure management, Python development, monitoring, governance, big data, and machine learning deployment. In his most recent experience at Thoughtworks, Breno developed CI/CD processes for a large food chain, created data pipelines, and migrated to infrastructure as code using Terragrunt, all on the Google Cloud Platform.

Portfolio

ThoughtWorks
Google Cloud Platform (GCP), Python 3, Terraform, Terragrunt, Linux...
Tembici
Python 3, Google Cloud Platform (GCP), BigQuery, SQL, Pub/Sub, Bash, Terraform...
Frontier Capital
Python 3, BigQuery, SQL, MySQL, Excel VBA, Data Warehousing, Google Cloud

Experience

  • Python 3 - 6 years
  • SQL - 6 years
  • BigQuery - 4 years
  • Google Cloud Platform (GCP) - 4 years
  • Apache Airflow - 3 years
  • Terraform - 3 years
  • APIs - 3 years
  • Terragrunt - 1 year

Availability

Part-time

Preferred Environment

Python 3, Google Cloud Platform (GCP), APIs, FastAPI, PySpark, Terraform, Terragrunt

The most amazing...

...thing I've developed is a data lake from scratch that combines data from multiple sources. I created the CI/CD processes and managed the orchestration tool.

Work Experience

Data Platform Engineer

2024 - PRESENT
ThoughtWorks
  • Implemented infrastructure as code (IaC) using Terragrunt.
  • Developed and maintained batch pipelines using Airflow.
  • Established CI/CD processes for data engineering and infrastructure development.
  • Built data monitoring and alerts using Airflow and cloud monitoring.
  • Created and maintained the data catalog and applied data access rules for data governance.
Technologies: Google Cloud Platform (GCP), Python 3, Terraform, Terragrunt, Linux, Google BigQuery, Pub/Sub, Apache Airflow, Google Cloud Composer, Docker, Data Management, Python, Data Pipelines, ETL, API Databases, Data Quality, Data Modeling, Data Architecture, Data Warehouse Design, Data Warehousing, BigQuery, Google Cloud

Data and MLOps Engineer

2021 - 2024
Tembici
  • Developed and maintained 190+ batch pipelines using Apache Airflow.
  • Built and maintained 10+ streaming pipelines using Apache Beam.
  • Created and maintained machine learning operations (MLOps) pipelines using Airflow.
  • Constructed CI/CD processes for the data and MLOps engineering area using Cloud Build.
  • Migrated data and MLOps engineering components with Terraform to obtain infrastructure as code (IaC).
  • Created APIs for machine learning real-time prediction using FastAPI.
Technologies: Python 3, Google Cloud Platform (GCP), BigQuery, SQL, Pub/Sub, Bash, Terraform, Terragrunt, Google Cloud Functions, Google Cloud Composer, Cloud Run, Docker, Apache Airflow, Cloud Dataflow, FastAPI, API Databases, Data Management, Data Pipelines, ETL, Data Engineering, Python, Data Quality, Data Warehouse Design, Data Warehousing, Data Build Tool (dbt), Machine Learning Operations (MLOps), Google Cloud, Business Analysis, Business Intelligence (BI)

Data Engineer

2020 - 2021
Frontier Capital
  • Maintained and developed the back-office system using VBA and SQL.
  • Developed batch pipelines using Python for daily reports.
  • Structured databases using MySQL to maintain the back office system.
Technologies: Python 3, BigQuery, SQL, MySQL, Excel VBA, Data Warehousing, Google Cloud

Data Scientist

2020 - 2020
Itaú Unibanco Holding
  • Worked as an analyst and data scientist in the customer service area.
  • Developed projects and data structures for data analysis.
  • Built a call prediction system for a 3-month-long summer master’s program.
  • Improved machine learning models for people analytics.
Technologies: Python 3, SQL, Scikit-learn, Tableau, Data Science, Data Analytics, Data Visualization

Data Engineer

2019 - 2019
Sympla
  • Developed a system for collecting and monitoring market data.
  • Architectured and constructed extract, transform, and load (ETL) and data warehouse loads.
  • Applied pre-processing techniques for text classification.
  • Built data transformations for the business areas.
Technologies: Python 3, SQL, Tableau, Amazon Web Services (AWS), Data Warehouse Design, Data Warehousing, Data Analytics

Data Analyst

2018 - 2018
Sympla
  • Developed data research and decision-making studies.
  • Understood processes and created KPIs for various business areas.
  • Created processes and routine optimizations using Python.
Technologies: Python 3, SQL, Tableau, Amazon Web Services (AWS), Data Warehouse Design, Data Warehousing, Data Analytics, Business Intelligence (BI), Business Analysis

Experience

Financial Engineering Simulation

The repository financial engineering project focuses on developing a simulation and risk analysis system for strategies in the American capital market. It employs optimization methods to model, simulate, and analyze financial strategies, providing insights into potential risks and returns.

Education

2019 - 2021

Master's Degree in Computational Intelligence

Federal University of Minas Gerais (UFMG) - Belo Horizonte, Minas Gerais, Brazil

2014 - 2018

Engineer's Degree in Industrial Engineering

IBMEC - Belo Horizonte, Minas Gerais, Brazil

Certifications

APRIL 2022 - APRIL 2024

Professional Data Engineer

Google Cloud

Skills

Libraries/APIs

Terragrunt, PySpark, Scikit-learn

Tools

Apache Airflow, Terraform, BigQuery, Google Cloud Composer, Cloud Dataflow, Tableau

Languages

Python, Python 3, SQL, Bash, Excel VBA

Paradigms

ETL, Business Intelligence (BI)

Platforms

Google Cloud Platform (GCP), Cloud Run, Docker, Linux, Amazon Web Services (AWS)

Storage

Data Pipelines, Google Cloud, API Databases, MySQL

Other

Data Quality, Data Modeling, Data Architecture, Data Engineering, Data Warehousing, Data Analytics, APIs, Data Management, Data Science, Data Warehouse Design, FastAPI, Machine Learning, Optimization, Cloud, Pub/Sub, Google Cloud Functions, Google BigQuery, Stock Market, Data Build Tool (dbt), Data Visualization, Machine Learning Operations (MLOps), Business Analysis

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring