Victor Martins, Developer in Curitiba - State of Paraná, Brazil
Victor is available for hire
Hire Victor

Victor Martins

Verified Expert  in Engineering

Bio

Victor is a data engineer with six years of experience developing cloud-based data pipelines. He specializes in big data applications, data privacy, data-driven software, cloud architecture and he excels at extracting value from data. Two of Victor's largest projects involved creating a data lake for one of Latin America's top fintechs from scratch and developing financial intelligence for one of the largest Brazilian e-commerce startups.

Portfolio

SimpleTire, LLC - Via Toptal
Amazon Web Services (AWS), Snowflake, Python, Containers, Apache Kafka...
Carta Healthcare
Python 3, HL7 FHIR Standard, Jenkins, Docker, Apache Kafka, PostgreSQL...
Pipefy
Amazon Web Services (AWS), Python 3, Kubernetes, Big Data, Data Lakes...

Experience

  • Python - 5 years
  • SQL - 5 years
  • Amazon Web Services (AWS) - 4 years
  • Data Engineering - 4 years
  • Tableau - 4 years
  • Data Analytics - 4 years
  • Snowflake - 3 years
  • Big Data - 3 years

Availability

Full-time

Preferred Environment

Amazon Web Services (AWS), Linux, PostgreSQL, MySQL, SQL, Python, Spark, Redshift, Snowflake

The most amazing...

...data lake I've developed was for one of Latin America's largest fintechs—completely serverless, seamlessly scales to 50x its size, and is 100% cloud-based.

Work Experience

Data Engineering Specialist

2021 - PRESENT
SimpleTire, LLC - Via Toptal
  • Developed a MySQL (AWS Aurora) to Snowflake near real-time replication pipeline. It replicates business analytics-specific data from the core database to the current Snowflake implementation.
  • Constructed a near-real-time order processing pipeline that uses AWS serverless architecture to deploy a train, test, and serve machine learning model environment. It's currently used to optimize the logistics tied to the business model.
  • Built a Snowflake integration with third-party tools and providers, such as CRM systems, data providers, and other sales channels. This integration ships data from those services and models them accordingly.
Technologies: Amazon Web Services (AWS), Snowflake, Python, Containers, Apache Kafka, Data Modeling, Data Architecture, Relational Databases, Data Integration, CI/CD Pipelines

Senior Back-end Engineer

2021 - PRESENT
Carta Healthcare
  • Optimized Jenknis test infrastructure and code, reducing overall test time by 20% on the first iteration, and prepared test structure to further improvements.
  • Implemented deploy metrics such as test coverage, code complexity, and other code quality-related KPIs. It covers both Python and Typescript codebases.
  • Implemented a default restful Python API abstraction for all business entities.
Technologies: Python 3, HL7 FHIR Standard, Jenkins, Docker, Apache Kafka, PostgreSQL, Elasticsearch, MinIO, Relational Databases, Data Integration, CI/CD Pipelines

Data Engineering Technical Lead

2021 - 2021
Pipefy
  • Planned the technical roadmap to migrate from legacy BI systems (BigQuery and Airflow) to an Apache Kafka stream-based approach.
  • Oversaw multiple customer-facing data product iterations, most notably improvements on the self-service analytics capability of the core product.
  • Created an Apache Kafka-based near real-time analytics solution to replicate databases across the company's entire ecosystem.
Technologies: Amazon Web Services (AWS), Python 3, Kubernetes, Big Data, Data Lakes, Streaming, Python, Relational Databases, Data Integration, CI/CD Pipelines

Data Engineering Specialist

2020 - 2021
Pipefy
  • Developed the second version of the back end serving the platform analytical tool that allows the customer to build custom dashboards with built-in OLAP capabilities. Built with AWS Redshift and Lambda.
  • Led a project to deliver customer data into a segmented Redshift database, allowing every company to have a high-performance analytical database with all of its data.
  • Architected the integration between the billing and analytical systems to understand better how customers behave before and after each billing or subscription change.
Technologies: Amazon Web Services (AWS), Big Data Architecture, Big Data, Data Engineering, Python 3, Relational Databases, Data Integration, CI/CD Pipelines

Senior Data Engineer

2018 - 2020
EBANX
  • Developed the company's data lake from scratch. The first version of the system had 30+ integrations and 1,500 tables and processed around 2TB of compressed data per day, built on top of AWS S3 and Redshift, with Spark as the core processing engine.
  • Built a database replication system to import all the database events using a CDC architecture. Built using AWS Kinesis and containers hosted on AWS ECS.
  • Migrated a 1,500 dashboard and 200+ daily active users Tableau deployment from a legacy data warehouse infrastructure to a data lake approach.
Technologies: Amazon Web Services (AWS), PostgreSQL, Tableau, MySQL, Python, Relational Databases, Data Integration

Business Intelligence Analyst

2018 - 2018
Rentcars.com
  • Architected the company's data warehouse and ETL workflows using AWS Lambda and AWS Redshift.
  • Redesigned and standardized the Tableau structure, visualization layouts and access management hierarchy completely.
  • Refactored the legacy ETL from a cron job-based system to an event batch-driven serverless architecture.
Technologies: Amazon Web Services (AWS), Tableau, Redshift, Python, Relational Databases, Data Integration

Business Intelligence Analyst

2016 - 2018
MadeiraMadeira
  • Developed the company's financial data mart, enabling it to deliver faster and more accurate insights to its stakeholders. Made on top of MariaDB using Pentaho as the ETL engine.
  • Created a process to ensure data quality in the customer satisfaction pipelines, helping the manager drive up CSAT by 10%.
  • Structured the company dashboard tool (Tableau), making data distribution more reliable and governable across all stakeholders.
Technologies: Amazon Web Services (AWS), Tableau, Python, Pentaho, MySQL, Relational Databases, Data Integration

Experience

Database CDC Engine

The goal of this Python-based project was to capture all the events (inserts, deletes, and updates) that happen on MySQL databases and stream them to AWS Kinesis and later to Parquet in S3.

I was the lead developer and architect, working with a team of six. The project took six months, from the initial idea to production deployment. The deliverables allowed the company to integrate new systems easily into its existing environments, effectively scale the query workload, and centralize the core of its pipeline in a homegrown application.

Analytical Engine

http://www.pipefy.com
A cloud-native, serverless, and Python-based application to deliver curated and modeled data to thousands of customers. The project involved embedding a dashboard creation tool into the product—a business process management SaaS. The main focus was to make it easier for the end-user to explore and extract value from the data already sitting in the relational database.

The whole software was developed on top of AWS, using Redshift, S3, Lambda, DynamoDB, and EventBridge. It is completely event-driven and serverless, and it has seamless scaling capabilities.

If you want to check it out, head over to pipefy.com, create a Pipe, and explore its data on the Dashboards tab.

Certifications

NOVEMBER 2019 - NOVEMBER 2022

AWS Certified Developer Associate

AWS

SEPTEMBER 2019 - SEPTEMBER 2022

AWS Certified Solutions Architect Associate

AWS

Skills

Libraries/APIs

Node.js

Tools

Git, Tableau, Jenkins

Languages

SQL, Python, Snowflake, Python 3

Paradigms

ETL, Business Intelligence (BI), OLAP, Lambda Architecture, HL7 FHIR Standard

Platforms

Amazon Web Services (AWS), AWS Lambda, Linux, Pentaho, Kubernetes, Docker, Apache Kafka

Storage

Database Modeling, Redshift, Amazon Aurora, MySQL, Amazon DynamoDB, Relational Databases, Data Integration, PostgreSQL, Amazon S3 (AWS S3), Data Lakes, Elasticsearch

Frameworks

Django, Spark

Other

Data Visualization, Data Transformation, Data Analytics, Data Engineering, Tableau Server, Big Data, Cloud, Data Modeling, Serverless, CI/CD Pipelines, Big Data Architecture, Dashboard Design, Dashboards, Containers, Streaming, MinIO, Amazon Kinesis, Parquet, Data Architecture

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring