Eduardo Bartolomeu
Verified Expert in Engineering
Software Developer
Eduardo is a senior data engineer with over 12 years of experience in the data field. He has worked as an Oracle and SQL Server database administrator and a PL/SQL and T-SQL (Transact-SQL) developer. With experience in the financial, retail, health, and education industries, Eduardo has most recently worked as a data engineer, specializing in AWS and Google Cloud Platform (GCP) environments, creating data pipelines, data lakes, ETLs, and data warehouses.
Portfolio
Experience
Availability
Preferred Environment
Amazon Web Services (AWS), SQL, Python, PySpark, Google Cloud Platform (GCP), Apache Airflow, ETL, Data Modeling, Big Data, Data Lake Design
The most amazing...
...feature I've created for data science teams predicts hospitalizations in healthcare plans, helping to save lives.
Work Experience
Senior Data Engineer
DataArt
- Developed stored procedures on Snowflake to consolidate data assets consumed by dashboards.
- Made Azure Data Factory pipelines for ETL Excel business files in Snowflake.
- Created Logic Apps to get email attachments and save them in Blob Storage.
- Created Azure Functions to run after a file arrives in the storage container doing transformation and load a new CSV to be consumed as a Snowflake stage.
Senior Data Engineer and SQL Developer
2am.tech
- Translated procedures from T-SQL (Transact-SQL) to SnowSQL.
- Created tables, stages, pipes, streams, procedures, and functions in the Snowflake data lake, taking data from SQL Server, PostgreSQL, and MySQL.
- Maintained and tested the SQL scripts using Liquibase.
Senior Data Engineer
Truelogic Software
- Created AWS Glue jobs using PySpark to transform data between the data lake zones.
- Performed dimensional modeling for data warehouses stored on Redshift.
- Documented processes using Confluence linked to Jira tickets.
- Built data pipelines from scratch from databases to the data lake and Redshift.
Senior Data Engineer
Neurotech
- Created data pipelines using Composer, Apache Airflow, and BigQuery, building datasets to be used by data science teams to predict hospitalizations and people with chronic diseases.
- Imported database files from the Brazilian public healthcare system to our data lake on AWS using EMR clusters and PySpark.
- Oversaw other data engineers on their tasks, helping them to achieve the company's expectations.
- Improved the performance of PySpark jobs running on EMR clusters.
Senior Database Administrator
Nyx Soluções
- Installed Oracle and SQL Server's database environment from scratch.
- Oversaw environment health statuses using monitoring tools.
- Improved the query performance for several clients, particularly in the retail industry.
- Created monthly environment health status reports for clients to monitor KPIs, including disk space, tablespace usage, heavy queries, and processor usage.
Experience
Snowflake Data Lake
http://www.emsmc.comData Lake and Data Warehouse
https://www.vectorsolutions.com/Cancer Identifier
http://portal.sulamericaseguros.com.brInitially, we migrated data sheets to BigQuery, focusing on the most common procedures for beneficiaries with cancer. Subsequently, we developed a Python algorithm to export the results for storage as both CSV files and tables. Everything was orchestrated using Composer and Apache Airflow.
Education
Master of Business Administration (MBA) in Business Intelligence
Institute of Management in Information Technology (IGTI) - Belo Horizonte, Brazil
Bachelor's Degree in Information Systems
Faculdade Estácio do Recife - Recife, Brazil
Certifications
AWS Certified Cloud Practitioner
Amazon Web Services
Splunk Core Certified Power User
Splunk
ITIL Foundation Certificate in IT Service Management
Axelos
Oracle Database 11g Administrator Certified Professional
Oracle
Skills
Libraries/APIs
PySpark, Liquibase, SQLAlchemy
Tools
AWS Glue, BigQuery, Amazon Elastic MapReduce (EMR), Apache Airflow, Splunk, Terraform, Amazon Athena, AWS CodeBuild, Google Cloud Dataproc, Google Compute Engine (GCE), Google Cloud Composer, Git, Composer, Azure Logic Apps
Languages
SQL, T-SQL (Transact-SQL), Python, Snowflake
Storage
Databases, PL/SQL, Data Lake Design, Oracle 11g, Oracle Database Tuning, MySQL, PostgreSQL, Microsoft SQL Server, Amazon S3 (AWS S3), Database Administration (DBA), Redshift, Data Pipelines, Google Cloud Storage, SQL Performance, Data Lakes, Azure SQL
Frameworks
ADF
Paradigms
ETL, Business Intelligence (BI), ITIL
Platforms
AWS Lambda, Jupyter Notebook, Oracle Database, Amazon Web Services (AWS), Google Cloud Platform (GCP), Amazon EC2, Linux, Windows Server, Azure SQL Data Warehouse, Azure Functions
Other
Google BigQuery, Data Modeling, Big Data, Oracle Performance Tuning, Amazon RDS, ELT, Software Development, IT Project Management, Product Management, Data Warehousing, AWS CodePipeline, AWS Database Migration Service (DMS), Amazon API Gateway, Metabase, Google Data Studio, Relational Database Services (RDS), CSV Import, CSV Export, Information Systems, SnowSQL, Document Management Systems (DMS), Lambda Functions, API Gateways, PL/SQL Tuning, Performance Tuning, Data Engineering, Azure Data Factory, Azure Data Lake, Azure Virtual Machines, Blob Storage
How to Work with Toptal
Toptal matches you directly with global industry experts from our network in hours—not weeks or months.
Share your needs
Choose your talent
Start your risk-free talent trial
Top talent is in high demand.
Start hiring