Sait Dogan, Developer in London, United Kingdom
Sait is available for hire
Hire Sait

Sait Dogan

Verified Expert  in Engineering

Bio

Sait is a data engineer with 15+ years of ETL development experience. He implemented many greenfield data warehouse projects in diverse sectors, such as telecommunications, banking, and insurance systems. For the last 5+ years, he's also been involved in big data projects.

Portfolio

Gauntlet Networks Inc
Python, SQL, ETL, Databases, Blockchain, BigQuery, Docker
Groupon
Python, Apache Sqoop, Hadoop, Amazon RDS, Amazon S3 (AWS S3), PostgreSQL, MySQL...
FIS Global (Worldpay)
Oracle Data Integrator (ODI), Python, PL/SQL, Shell, Spark, Performance Tuning...

Experience

Availability

Part-time

Preferred Environment

ETL, Data Engineering, Python, SQL, Informatica, Oracle Data Integrator (ODI)

The most amazing...

...thing I've designed and implemented is an enterprise data warehouse and self-service BI environment that serves more than 300 head office users.

Work Experience

Data Engineer

2022 - 2022
Gauntlet Networks Inc
  • Refactored the ETL packages written for lending protocols to modularize and make them scalable.
  • Config-driven ETL design has been applied to current modules. With this new design, integrating new customers into the system became flexible and much easier.
  • Tested and validated newly written ETL flows by comparing their output with production data.
Technologies: Python, SQL, ETL, Databases, Blockchain, BigQuery, Docker

Senior Data Engineer

2021 - 2022
Groupon
  • Acted as part of the data ingestion team and was in charge of a replication flow of MySQL and Postgres data to HDFS, Hive, and Teradata through in-house developed ETL tools.
  • Developed an audit package that compares record counts between source and target systems and sends the differences greater than a threshold to a dashboard.
  • Added a feature to the in-house developed ETL tool to incrementally get data from Postgres sources.
  • Contributed cloud migration of on-prem environments and services to GCP and AWS. Set up Dataproc clusters, configured jobs on Airflow, implemented audit scripts with Python on GCP, and configured hosts via Terraform on AWS.
Technologies: Python, Apache Sqoop, Hadoop, Amazon RDS, Amazon S3 (AWS S3), PostgreSQL, MySQL, Teradata, Apache Airflow, Terraform, Google Cloud Platform (GCP), Packer, Amazon Machine Images (AMI), Scala

Data Engineer

2018 - 2021
FIS Global (Worldpay)
  • Served as a data engineer in the data integration team that mainly uses ODI in the new acquiring platform project.
  • Made various ODI knowledge module customizations by writing Python codes to optimize workflows.
  • Conducted performance tuning after data volumes increased due to migration projects by using SQL and PL/SQL.
Technologies: Oracle Data Integrator (ODI), Python, PL/SQL, Shell, Spark, Performance Tuning, Data Analytics

Lead Data Warehouse Developer

2014 - 2018
Akbank
  • Created three new data marts and analyzed the needs of business teams using Informatica.
  • Increased the use of the ODS environment by improving validation jobs using Informatica and PL/SQL.
  • Established ETL offloading for performance tuning using Hadoop and Spark.
  • Continuously conducted performance tuning on ETL jobs using SQL, PL/SQL, and Informatica.
Technologies: Informatica, Oracle, SQL, PL/SQL, Shell, Hadoop, Spark, Apache Sqoop, Apache Hive, Performance Tuning, Data Analytics

Senior ETL Developer

2014 - 2014
KKB
  • Validated and integrated all credit card-related data from multiple banks in Turkey into the KKB system by using Informatica.
  • Performed performance tuning on ETL jobs using SQL, PL/SQL, and Informatica.
  • Designed and developed a map-based dashboard for reporting Corporate Score Query results by using OBIEE.
Technologies: Informatica, Oracle Business Intelligence Enterprise Edition 11g (OBIEE), PL/SQL, SQL

Data Warehouse Architect

2012 - 2014
Türkiye Sigorta
  • Established an enterprise data warehouse and self-service BI environment for head office users using ODI and PL/SQL.
  • Automatized regulatory reporting through a regulatory reporting data mart using OBIEE.
  • Provided training on ODI and OBIEE to various departments in the company.
  • Made customizations by writing Python codes to optimize an ETL workflow.
Technologies: Oracle Data Integrator (ODI), Oracle Business Intelligence Enterprise Edition 11g (OBIEE), PL/SQL, Python, Performance Tuning

Optimus ETL Tool

A config file-based, highly scalable ETL pipeline project has been implemented in Python and SQL. No code change is needed when adding a new table to the flow. Adding any table to the flow is as easy as creating a simple YAML file. ETL metadata tables track all processes. Any anomaly in the flow can be detected easily by querying these tables.
1997 - 2002

Bachelor of Science Degree in Computer Engineering

Ege University - Izmir, Turkey

SEPTEMBER 2020 - PRESENT

SQL (Advanced)

HackerRank

MAY 2020 - NOVEMBER 2021

Oracle Cloud Infrastructure 2019 Certified Architect Professional

Oracle

JANUARY 2019 - PRESENT

PowerCenter Data Integration 10: Developer

Informatica

Libraries/APIs

PySpark

Tools

Apache Airflow, GitHub, PyCharm, Apache Sqoop, Shell, Oracle Business Intelligence Enterprise Edition 11g (OBIEE), Spark SQL, BigQuery, AWS Glue, Amazon Athena, Terraform, Packer, Informatica PowerCenter

Languages

SQL, Stored Procedure, Python, Bash, Scala

Paradigms

ETL, OLAP, Business Intelligence (BI), Database Design, Database Development, Database-driven Web App Development

Platforms

Oracle Data Integrator (ODI), Oracle, Oracle Database, Visual Studio Code (VS Code), Amazon Web Services (AWS), Blockchain, Docker, Google Cloud Platform (GCP)

Storage

PL/SQL, Oracle PL/SQL, Databases, Data Pipelines, Oracle Cloud, OLTP, Database Management Systems (DBMS), Amazon S3 (AWS S3), PostgreSQL, MySQL, Teradata, Apache Hive, MongoDB

Frameworks

Hadoop, Spark, Apache Spark

Other

Informatica, Data Warehousing, Database Optimization, Query Optimization, Data Cleaning, Data Cleansing, Performance Tuning, ELT, Data Engineering, Data Migration, Parquet, Data Analysis, ETL Testing, Data Manipulation, Data Quality, Data Warehouse Design, Data Profiling, Datasets, Query Composition, Data Analytics, Data Modeling, PL/SQL Tuning, AWS Cloud Architecture, Amazon RDS, Big Data, Dashboards, CSV, Google BigQuery, Data Matching, Cloud Architecture, Amazon Machine Images (AMI)

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring