Satish Basetty, Developer in Los Angeles, CA, United States
Satish is available for hire
Hire Satish

Satish Basetty

Verified Expert  in Engineering

Bio

Satish is a senior data engineer with over 14 years of experience in database and data warehouse projects in both on-premises and cloud. He is an expert in the design and development of ETL pipelines using Python and SQL over Cloud Dataflow orchestration with Apache Airflow. He automated processing data of royalties and copyrights for Universal Music Group. Satish has provided solutions encompassing reports and visualizations, real-time data processing, migrations, and performance tuning.

Portfolio

The Estee Lauder Companies Inc.
Apache Airflow, Python, Google Cloud Platform (GCP), BigQuery, SQL, Bash...
LA City
Apache Airflow, BigQuery, Cloud, Docker, Kubernetes, Microsoft Power BI...
Kitchen United
Amazon Athena, PostgreSQL, AWS Glue, Python 3, ETL, Google Cloud, Docker...

Experience

Availability

Part-time

Preferred Environment

Slack, PyCharm, Windows, Linux, MacOS, Docker, Docker Hub, Oracle, Streaming Data, HIPAA Compliance, Amazon Web Services (AWS), Data Analytics, Tableau, Data Modeling, PostgreSQL, SQL, Google Analytics, Data Engineering, Data Analysis, APIs

The most amazing...

...solution I've provided was a highly scalable daily sales reporting process.

Work Experience

Senior Data Engineer

2022 - PRESENT
The Estee Lauder Companies Inc.
  • Developed data pipeline jobs to ingest web search and sales data into BigQuery.
  • Fixed and tracked issues that were occurring in the data pipeline using the Jira tool.
  • Oversaw deployments using Git and CI/CD. Created a high-level process flow and technical design specification document.
Technologies: Apache Airflow, Python, Google Cloud Platform (GCP), BigQuery, SQL, Bash, Data Pipelines, Data Engineering, HIPAA Compliance, Data Analytics, Google Analytics

SR Data Engineer

2021 - 2022
LA City
  • Deployed the web application from on-prem to Google Cloud, developed Dataflow pipelines, and implemented CI/CD in the Cloud environment.
  • Tracked bugs using Jira and troubleshot pipeline-related errors and performance tuning.
  • Oversaw the workload processed from the pipeline jobs.
Technologies: Apache Airflow, BigQuery, Cloud, Docker, Kubernetes, Microsoft Power BI, PostgreSQL, Google Analytics, JSON

Senior Data Engineer

2019 - 2021
Kitchen United
  • Created a daily reporting process to send out reports to members. This daily process ingests the data into the data lake then the "send email" process sends the reporting emails to all members.
  • Developed the ETL pipeline to ingest the purchase data into the data lake. Created the batch job using PySpark and Apache Beam to load the third-party sales data into the data lake.
  • Designed and developed the data mart that provides insights and visualization.
  • Automated the process for onboarding and offboarding members.
Technologies: Amazon Athena, PostgreSQL, AWS Glue, Python 3, ETL, Google Cloud, Docker, Apache Beam, ETL Pipelines, Python, Business Intelligence (BI), Data Engineering, Streaming Data, Amazon Web Services (AWS), Data Analytics, Data Modeling, SQL

Senior Data Engineer

2018 - 2019
Fabfitfun
  • Designed a data mart to track the sales, CPA, and churns across various sales channels—provided a solution for automated AB testing.
  • Developed the ETL pipeline to ingest data related to the add on purchases and seasonal box delivery to members across Fabfitfun.
  • Developed the ETL pipeline for survey data ingestion.
  • Designed and developed the style data mart that provides visualizations across top-selling SKUs.
Technologies: PostgreSQL, Python 3, Apache Airflow, ETL, Apache Beam, ETL Pipelines, Python, Business Intelligence (BI), Data Engineering, Migration, Amazon Web Services (AWS), Tableau

Senior Data Engineer

2017 - 2018
Machinima
  • Developed a process that provides video data insights.
  • Designed and developed the data mart that provides visualizations on the best performing videos across channels.
  • Configured the Goofys file system used as a primary source/target for most of the ELT/ETL process.
Technologies: Redshift, PostgreSQL, Python 3, ETL Pipelines, Python, Business Intelligence (BI), Data Engineering, Oracle, Data Modeling

Data Engineer

2015 - 2017
PennyMac
  • Gathered requirements and completed data analysis, design, and development of the ELT/ETL process using Pentaho and Python.
  • Designed a data lake on AWS for various processes with data ingestion into the data warehouse Redshift and Snowflake. Worked with stakeholders in resolving issues and completing requirements.
  • Oversaw performance tuning of the queries and provided operations support.
Technologies: Snowflake, Python 2, PostgreSQL, Python, Data Engineering

Senior Database Developer

2014 - 2015
BeachMint
  • Designed and developed ELT/ETL processes using Python.
  • Designed a sales data mart of complex queries.
  • Oversaw performance tuning of queries.
Technologies: Redshift, Bash, Python 2, PostgreSQL, Tableau

Senior Developer

2013 - 2014
Bank of America
  • Designed and developed the ETL process. Collaborated with stakeholders to resolve issues and clarify requirements.
  • Designed the order data-mart and loaded the data using the ETL Pentaho and SQL.
  • Managed the performance tuning of the queries.
Technologies: Oracle, PostgreSQL, Python 2

Database developer

2011 - 2013
Universal Music Group
  • Developer ETL processes using Oracle PL/SQL to extract the legacy data and load it into the data mart.
  • Oversaw the performance tuning of complex queries. Gathered requirements from end-users and designed the data mart for royalties and copyrights.
  • Performed data analysis for royalties and copyrights. Created an automation process for processing the data.
Technologies: Bash Script, Oracle PL/SQL

ETL Developer

2007 - 2010
Prokarma
  • Oversaw the data migration project from the legacy system to SAP.
  • Developed the ETL process to handle the car's data.
  • Collaborated with stakeholders on requirements gathering. Performed data analysis.
Technologies: SAP FICO, Shell, Oracle PL/SQL

Senior Developer

2006 - 2007
RapidIgm Consulting
  • Developed an ETL process to perform data integration from various sources. Peformed analysis on the Rx and DDD data.
  • Designed the sales data mart and assisted with complex queries and performance tuning.
  • Collaborated with stakeholders to gather requirements and develop the data modeling.
Technologies: SQL, Oracle

Sales Data Ingestion

I developed and architected a data pipeline to ingest sales data into the data mart for Kitchen United. This process keeps track of the sales performance and supply chain across several kitchen centers and provides near real-time data insights for Menumix to the members. The data pipeline process ingests the data for purchase and menu events into the data lake, and also keeps track of the daily sales by location and members.

I collaborated worked with the finance, marketing, data science, and BI teams and provided solutions accordingly. I helped build the data modeling that enabled the BI team to create reports and dashboards. I created a reconciliation process to keep track of the orders, a cloud to watch alerts, error reporting, and an outbound process to various third-party vendors.

Gaming/Video Data Ingestion for Machinima

Assisted in the development of an end-to-end process for gaming and video data ingestion. The real-time process ingests the live gaming data into the data lake.

I designed the data mart to track insights at the video-id grain from various channels and collaborated with the finance, email marketing and BI teams. I developed a process to ingest the sentiment data events into the data mart and configured the Goofys file system used as the primary source/target for most of the ELT/ETL process.

Sentiment Data Analysis

Developed an ETL pipeline to ingest the survey data into the data lake and created a data mart for the sentiment data analysis. I collaborated with business users and designed the database views for the reporting.
2002 - 2005

Master's Degree in Computer Science

Texas A & M University - College Station, Texas, USA

Libraries/APIs

Scikit-learn

Tools

AWS Glue, Apache Airflow, Amazon Athena, Apache Beam, BigQuery, Cloud Dataflow, Google Cloud Composer, Tableau, Google Analytics, PyCharm, Slack, Docker Hub, Microsoft Power BI

Languages

SQL, Python, Snowflake, Python 3, Python 2, Bash

Paradigms

ETL, Business Intelligence (BI), HIPAA Compliance

Platforms

Azure, AWS Lambda, Amazon Web Services (AWS), Docker, Oracle, Google Cloud Platform (GCP), Kubernetes

Storage

PostgreSQL, Redshift, Google Cloud, Data Pipelines, JSON

Other

Data Warehousing, Query Optimization, Data Warehouse Design, Data Engineering, Migration, Data Analytics, ETL Pipelines, Indexing, Streaming Data, Data Modeling, Data Analysis, APIs, Slurm Workload Manager, Cloud

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring