Jorge Andres Robiola, Developer in Miami, FL, United States
Jorge is available for hire
Hire Jorge

Jorge Andres Robiola

Verified Expert  in Engineering

Data Engineer and Developer

Miami, FL, United States

Toptal member since November 7, 2023

Bio

Jorge is a senior data engineer and architect with vast expertise in cloud back-end systems. He has consistently focused on designing data mesh solutions to save costs and improve execution performance, incorporating automation technologies and advanced AI techniques. In addition, Jorge has a proven track record as a project manager dedicated to infrastructure and software engineering endeavors.

Portfolio

Entrustody
Azure Synapse, Azure SQL Databases, Postman, Azure Data Factory (ADF), Azure...
EY
Azure Databricks, Neo4j, Azure Data Factory (ADF), Azure Synapse...
Newsweek
Artificial Intelligence (AI), Kubernetes, Google BigQuery, Azure Data Lake...

Experience

  • Data Analysis - 10 years
  • Spark - 7 years
  • Python - 7 years
  • PySpark - 7 years
  • Python 3 - 7 years
  • Snowflake - 5 years
  • Delta Lake - 2 years
  • Artificial Intelligence (AI) - 1 year

Availability

Part-time

Preferred Environment

IT Systems Engineering

The most amazing...

...project I've worked on is the successful migration of an IT database support service for 84 countries to a single global delivery center.

Work Experience

Data Engineer

2023 - PRESENT
Entrustody
  • Integrated subscriptions and customer workspaces into a single consolidated data mesh in Azure.
  • Supported an existing data lake model installed on Microsoft Azure.
  • Migrated data-driven models to Fabric, supported data warehouse pipelines, and collaborated with real-time applications.
Technologies: Azure Synapse, Azure SQL Databases, Postman, Azure Data Factory (ADF), PostgreSQL, Azure, Azure Cosmos DB, Python, Spark, Scala, PySpark, Delta Lake, Microsoft Power BI, APIs, Microsoft Azure, Fabric, Microsoft SQL Server, Data Engineering, Data Analytics

Data Engineer

2021 - PRESENT
EY
  • Applied different Databricks techniques to re-design data warehousing models, saving costs in the cloud and improving the data ingestion performance.
  • Implemented new data mesh systems using cloud-based data-driven tools.
  • Started a project to implement AI for a large language model (LLM) to identify privacy text included in audit documents.
Technologies: Azure Databricks, Neo4j, Azure Data Factory (ADF), Azure Synapse, Azure SQL Databases, Azure VDI, Azure Functions, Applications, Python 3, PySpark, Large Language Models (LLMs), Supervised Machine Learning, Microsoft SQL Server, Data Engineering, Data Analytics

Senior Data Engineer

2022 - 2023
Newsweek
  • Saved costs in the data-driven process using materialized views in Azure's dedicated pool and delta lake architecture.
  • Built connected machine learning models to monetarize subscribers, identifying pattern behaviors.
  • Performed data-driven orchestration among cloud environments, including Azure and Google Cloud Platform (GCP), to build a robust analysis of subscribers.
Technologies: Artificial Intelligence (AI), Kubernetes, Google BigQuery, Azure Data Lake, Python, PySpark, Synapse, Data Integration, Analytical Thinking, Data Warehousing, SQL, Microsoft Azure, Microsoft Power BI, APIs, Azure Data Factory (ADF), MongoDB, Google Cloud Platform (GCP), Microsoft SQL Server, Data Engineering, Data Analytics

Data Engineer and DevOps Developer

2019 - 2023
Enve
  • Incorporated a cloud data platform to an existing AWS architecture for data ingestion and final storage into massively parallel processing environments.
  • Worked with cloud-based providers, including Snowflake, Databricks, and Redshift, to migrate on-premise data warehouse platforms into AWS and Azure.
  • Introduced fundamental automation tools to create an infrastructure as code (IaC) environment in AWS using HashiCorp tools.
Technologies: APIs, Amazon EC2 API, AWS Lambda, Amazon DynamoDB, Amazon Aurora, PostgreSQL, Relational Database Services (RDS), Redis, Redshift, Document Management Systems (DMS), AWS Glue, ETL, Apache Airflow, Datadog, Grafana, Terraform, Packer, Azure Logic Apps, Containers, Azure Data Lake, Blob Storage, Azure Data Factory (ADF), Azure SQL, SQL, Databricks, Synapse, Google Cloud Platform (GCP), BigQuery, Virtual Machines, Amazon EC2, Amazon Virtual Private Cloud (VPC), AWS IAM, Amazon S3 (AWS S3), Amazon RDS, Ansible, Vagrant, MongoDB, Kubernetes, Docker Swarm, Infrastructure as Code (IaC), HashiCorp, Microsoft SQL Server, Data Engineering, Data Analytics, Parallels Business Automation

Data Engineer and BI Analyst

2021 - 2022
PayPal
  • Collaborated with the analytics payment checkout department to identify, analyze, and generate data pipelines to create a root-cause analysis for errors within web and mobile tools.
  • Leveraged Google Cloud Platform (GCP), BigQuery, Google Cloud Functions, and Apache Airflow to collect data from an HDFS ecosystem and load a data warehouse model connected to payroll.
  • Maintained a complex Tableau reporting platform to visualize the analysis performed on BigQuery.
Technologies: Google BigQuery, Machine Learning, Python, PySpark, Data Integration, Analytical Thinking, Tableau, Data Warehousing, SQL, Hadoop, HDFS, BigQuery, Google Cloud Platform (GCP), Google Cloud Functions, Apache Airflow, Microsoft SQL Server, Data Engineering, Data Analytics

Data Engineer and BI Architect

2018 - 2021
TradeHelm, Inc.
  • Provided services for Hightower Advisors, an RIA enterprise located in Chicago.
  • Maintained an existing data warehouse system mounted in an SQL Server environment and migrated it to a cloud platform located in AWS.
  • Leveraged Snowflake as the final data warehouse system for this project, with the ongoing objective of creating pipelines and re-designing the data architecture model based on the new standards.
Technologies: Snowflake, REST APIs, Machine Learning, AWS DevOps, Tableau Server, CRM Implementation (Salesforce), HashiCorp, Packer, Ansible, Terraform, PySpark, Amazon Redshift, Tableau, Data Warehousing, SQL, Amazon Web Services (AWS), Docker, Amazon RDS, Microsoft SQL Server, Financial Risk Management, Finance, Investment Funds, Banking & Finance, Data Engineering, Data Analytics

Data Engineer and BI Analyst

2018 - 2019
ProKarma
  • Replaced a data warehouse model stored in Relational Database Services (RDS) and PostgreSQL with the Hadoop ecosystem based on Apache Hive and Impala.
  • Analyzed the data consumed from Apache Hive, compared the values against PostgreSQL, and loaded the data into Hadoop, replacing the data sources in the Tableau server.
  • Adjusted the Tableau dashboards and sheets after replacing PostgreSQL with Hadoop and Impala.
Technologies: ETL, Cloudera, Spark, Scala, Hadoop, Python, Jira, PostgreSQL, Data Integration, Amazon Redshift, Data Warehousing, SQL, Python 3, Amazon RDS, Apache Impala, Apache Hive, Relational Database Services (RDS), Tableau, Microsoft SQL Server, Data Engineering, Data Analytics

Data Engineer and BI Architect

2017 - 2018
Zenfolio
  • Worked on a financial data warehouse and reporting platform, creating a common repository to extract and transform the information coming from different sources.
  • Loaded the final Snowflake schema model into a visualization tool, creating different marketing and financial reports refreshed incrementally throughout an automated ETL process.
  • Built the final data warehouse in an Amazon RDS cloud environment, while the visualization platform was constructed in Sisense ElastiCube.
Technologies: Amazon RDS, Sisense ElastiCube, Python, Microsoft SQL Server, Data Engineering, Data Analytics

Data Engineer and BI Team Leader

2016 - 2018
Lighthouse Investment Partners
  • Built a flexible ETL platform to integrate information from various risk management applications, including Axioma, Thomson Reuters, Beta, Barra, and SunGard. The platform shows principal statistics in an efficient reporting tool.
  • Started work on analytic scripts created initially in RStudio. The output information had to be automated and integrated into the existing data warehouse model.
  • Joined the data science team to create complex machine learning models to identify patterns in the stock exchange transactions and load the data into a data warehouse system, keeping required runtimes.
Technologies: RStudio, Python 3, Amazon Redshift, Microsoft SQL Server, Financial Risk Management, Finance, Investment Funds, Banking & Finance, Data Engineering, Data Analytics

Data Engineer and BI Developer

2016 - 2017
Kestra Financial
  • Created an ETL process and database design to integrate different CRM and financial management applications.
  • Updated the existing reporting platform and analyzed a potential migration to Microsoft Power BI tools mounted on a cloud environment.
  • Integrated different financial platforms into a consolidated data warehouse system.
Technologies: SQL Server Integration Services (SSIS), Microsoft Power BI, C#, Microsoft SQL Server, Financial Risk Management, Finance, Data Engineering, Data Analytics

Experience

IBM and Novartis Migration Project

I participated in a global project that comprised the migration of an IT service in 84 countries. The project's goal was to take over the ITSCM environment in place, including operating system, networking, data management, and monitoring tools, and migrate the system to perform IT activities remotely from the global delivery center located in Buenos Aires, Argentina. To achieve this goal, I covered the Asia-Pacific region. I traveled to Singapore and Japan to complete the due diligence phase, settle deliverables, and provide vital information about the database systems.

Education

2000 - 2002

Bachelor's Degree in Information Technology

Interamerican Open University (UAI) - Buenos Aires, Argentina

Certifications

MARCH 2023 - DECEMBER 2027

Data Engineering with Databricks

Databricks

JANUARY 2023 - PRESENT

Containers & Kubernetes Essentials

IBM

NOVEMBER 2020 - PRESENT

Microsoft Certified: Azure Fundamentals

Microsoft

AUGUST 2020 - PRESENT

AWS Certified Cloud Practitioner

Amazon Web Services

NOVEMBER 2018 - PRESENT

Big Data Hadoop and Spark Developer

Simplilearn

MAY 2018 - PRESENT

Python for Data Science

IBM

MAY 2018 - PRESENT

Machine Learning with Python

cognitiveclass IBM

MAY 2018 - PRESENT

Data Analysis with Python

IBM

APRIL 2014 - PRESENT

Scrum Grand Master

National Technological University (UTN)

MAY 2011 - PRESENT

Information Technology Infrastructure Library (ITIL) Foundations V3

EXIN

SEPTEMBER 2010 - SEPTEMBER 2013

Project Management Professional (PMP)

Project Management Institute

DECEMBER 2007 - PRESENT

Microsoft Certified Professional (MCP)

Microsoft

Skills

Libraries/APIs

PySpark, Amazon EC2 API, REST APIs, Fabric

Tools

Tableau, BigQuery, Postman, Microsoft Power BI, Azure Kubernetes Service (AKS), Synapse, AWS Glue, Apache Airflow, Grafana, Terraform, Packer, Azure Logic Apps, Ansible, Vagrant, HashiCorp, Jira, AWS IAM, Amazon Virtual Private Cloud (VPC), Apache Impala, Docker Swarm, Cloudera

Languages

SQL, Python 3, Python, Scala, Snowflake, C#

Storage

Azure SQL Databases, SQL Server Integration Services (SSIS), Microsoft SQL Server, PostgreSQL, Data Integration, Amazon Aurora, Azure SQL, Neo4j, Azure Cosmos DB, MongoDB, Amazon DynamoDB, Redis, Redshift, Datadog, Amazon S3 (AWS S3), HDFS, Apache Hive

Frameworks

Spark, Hadoop

Platforms

Databricks, Azure Synapse, Azure Functions, Azure, Kubernetes, AWS Lambda, Google Cloud Platform (GCP), Amazon Web Services (AWS), Docker, RStudio, Amazon EC2, Microsoft

Industry Expertise

Banking & Finance, Project Management

Paradigms

ETL, ITIL

Other

Data Analysis, Data Engineering, Data Analytics, IT Systems Engineering, Programming, Azure Databricks, Delta Lake, Scrum Master, IT Project Management, Google BigQuery, Analytical Thinking, Relational Database Services (RDS), Financial Risk Management, Finance, Investment Funds, Azure Data Factory (ADF), Azure VDI, Applications, Large Language Models (LLMs), Supervised Machine Learning, APIs, IT Management, Machine Learning, Data Science, ITIL V3 Foundation Certified, Artificial Intelligence (AI), Azure Data Lake, Data Warehousing, Microsoft Azure, Document Management Systems (DMS), Containers, Blob Storage, Virtual Machines, AWS DevOps, Tableau Server, CRM Implementation (Salesforce), Amazon Redshift, Sisense ElastiCube, Amazon RDS, Information Technology, Systems Analysis, Infrastructure as Code (IaC), Google Cloud Functions, Parallels Business Automation

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring