
Ceyhun Kerti
Verified Expert in Engineering
Data Engineering Developer
Istanbul, Turkey
Toptal member since March 10, 2022
Ceyhun is a data engineer with experience in various industries and technologies. His primary goal is to introduce the latest technology and methodologies to his customers at the minimum cost and maximum stability. Ceyhun enjoys working in interdisciplinary fields using hybrid technology stacks.
Portfolio
Experience
- Python - 17 years
- Data Engineering - 17 years
- ETL - 16 years
- SQL - 15 years
- JavaScript - 8 years
- Scala - 4 years
- Go - 3 years
- Apache Beam - 1 year
Availability
Preferred Environment
Python, SQL, ETL, Data Engineering
The most amazing...
...project I've developed is a score calculation platform for sustainable development growth that represents a step forward for a better future for humanity.
Work Experience
Data Engineer
Plan A
- Developed data pipelines to calculate carbon emissions from business activity data.
- Created API endpoints for dynamic carbon emissions calculation and management of metadata for emission conversion factors.
- Implemented reference pipelines in Snowflake and dbt.
- Designed a real-time emission calculation system using Elixir/Flow to handle concurrency with minimal resource overhead.
Data Engineer
Spotify - Main
- Developed an ad tracking system that measures how much of an ad is listened to by the unique user and how many ads are listened to in the content. Calculated related metrics and impressions used by downstream systems like finance and reporting.
- Fixed bugs in the system and worked on the usual small-scale tasks on the sprint.
- Prepared RFCs and reference implementations to improve the stability and reliability of the pipelines.
- Involved in planning to add dbt to our data pipeline stack, which was already used by other teams with whom we interact.
Senior Data Engineer Consultant
Screen Seventeen
- Developed ETL pipelines from different data providers to AWS S3.
- Created an SDK for the data science team so they can easily access and wrangle data from different sources.
- Maintained ETL pipelines in Apache Airflow and helped the data science team to package their flows.
Senior Data Engineer Consultant
BBVA Garanti Bank
- Created and implemented data pipelines for loan risk reports.
- Implemented a couple of fast-track projects related to customer interactions.
- Analyzed the underlying infrastructure and existing pipelines and created possible restructuring guidelines.
Senior Data Engineer Consultant
Fiat
- Created an ETL pipeline between IoT source systems and Azure Blob Storage.
- Identified problems in the existing ETL pipelines and developed and maintained optimized data pipelines.
- Processed unstructured data in Azure with Spark and created structured models.
- Conducted training sessions for employees of the company.
Full-stack Engineer Consultant
ING Bank
- Implemented a web UI for the internal lending grid optimization tool.
- Maintained and packaged the application to be easily deployed to different platforms.
- Made the demo and showcase of the application for different company branches in other countries.
Senior Data Engineer Consultant
Medical Park
- Developed ETL pipelines for real-time reporting and BI infrastructure.
- Made implementation data transfer modules between SQL Server and PostgreSQL.
- Implemented data ingestion and real-time reporting features in KSQL/Kafka.
Senior Data Engineer Consultant
ING Bank
- Developed real-time ETL pipelines for the client.
- Created dynamic AI model visualization infrastructure in Python with Plotly.
- Helped to maintain and enhance existing ETL pipelines between the core system and Spark cluster.
Senior Data Engineer Consultant
CK Energy
- Implemented high-performance ETL pipelines for the company and its subsidiaries.
- Managed an outsourced team for the business intelligence and data warehouse tasks.
- Created a web-based management software for the ETL pipeline manager so that the company can manage all the flow once for all subsidiaries.
- Optimized existing ETL pipelines and reduced the total ETL time.
Senior Data Engineer Consultant
Aksigorta
- Developed highly scalable ETL pipelines for the client in SQL, Python, and Oracle Data Integrator (ODI).
- Made the transition from old ETL pipelines to a new data integrator.
- Involved in preprocessing and masking customers' sensitive healthcare data.
- Implemented highly optimized data extraction utility.
- Provided training sessions in SQL, ODI, and Python to various departments in the company.
- Optimized the current ETL pipelines and reduced the total ETL time.
- Implemented a near-realtime ETL pipeline so the company and its agencies can see the reports and insights with a 2-minute delay.
- Created a basic implementation of data integration with other companies where the client has data exchange agreements, like some hospitals, subsidiary banks, etc. Used DBT and Airflow for integration and workflow management.
Senior Data Engineer Consultant
Dicle Energy
- Developed ETL pipelines in SQL and Oracle Data Integrator (ODI).
- Managed and trained a team of between 10 and 15 people.
- Implemented highly optimized data extraction utility.
Data Engineer
Oracle
- Developed highly scalable ETL pipelines with big data for different companies in the telecommunication industry, including Vodafone, Turkish Telecom, Avea, and Turkcell.
- Created both internal and client-facing applications.
- Mentored junior developers to enjoy software development and produce high-quality applications.
Experience
Fast Oracle to CSV Extraction Utility
https://github.com/ceyhunkerti/oracle_tocsvWeb-based SQL Query Editor
https://github.com/ceyhunkerti/owlIt can be used to:
• Query local cloud storage files (GCS, S3).
• Query PostgreSQL and DuckDB databases.
You can create multiple users, and each user can manage their user space (databases, connections, etc.). It also has built-in user authentication and Google OAuth support.
Cross-platform Plugin-based Data Transfer Tool
https://github.com/bluecolor/tractorOnline Data Redaction Platform
https://github.com/bluecolor/redactTask Scheduler
https://bluecolor.github.io/octopus/Education
Bachelor's Degree in Computer Science Engineering
Yeditepe University - Istanbul, Turkey
Skills
Libraries/APIs
Pandas, REST APIs, Luigi, PySpark, Terragrunt
Tools
Apache Airflow, Postman, Plotly, Apache Sqoop, AWS CloudFormation, Prefect, Apache Beam, Terraform, Git, GitHub, BigQuery
Languages
Python, Go, JavaScript, SQL, Snowflake, Java, Scala, Bash Script, Elixir, TypeScript, Zig, C
Frameworks
Flask, Apache Spark, Presto, Spark, Ruby on Rails (RoR), Streamlit
Paradigms
ETL, ETL Implementation & Design, Dataflow Programming, Business Intelligence (BI)
Platforms
Oracle, Oracle Data Integrator (ODI), Amazon Web Services (AWS), Kubernetes, Oracle Database, Databricks, MapR, Azure, AWS Lambda, Docker, Linux, Google Cloud Platform (GCP)
Storage
Databases, PL/SQL, Druid.io, MySQL, Data Pipelines, Data Integration, Oracle PL/SQL, PostgreSQL, Amazon S3 (AWS S3), Cassandra
Other
Data Engineering, Scripting, Google BigQuery, Data Build Tool (dbt), API Integration, Data Analysis, Data Modeling, Data Analytics, Data Transformation, Data Migration, Data Science, Delta Lake, Cloud, Azure Databricks, Data Warehouse Design, Data Warehousing, Dash, DuckDB, FastAPI, Computer Science, Polaras, GitHub Actions, CSV, Big Data
How to Work with Toptal
Toptal matches you directly with global industry experts from our network in hours—not weeks or months.
Share your needs
Choose your talent
Start your risk-free talent trial
Top talent is in high demand.
Start hiring