Yunus Yünel, Developer in London, United Kingdom
Yunus is available for hire
Hire Yunus

Yunus Yünel

Verified Expert  in Engineering

AWS Solution Architect, Data Engineer, and DWH Developer

Location
London, United Kingdom
Toptal Member Since
March 15, 2021

Yunus is a data engineer who sees data as Lego. It is meaningless without other pieces, and when put together, something valuable is created. He excels with automation, data warehouse, BI, and big data projects. He cares about modularity in scripts and delivers bespoke solutions for clients. Yunus developed a loyalty score project and moved a semi-structured mainframe log file to a big data environment with PySpark for a bank and maintained around 7,000 jobs (mostly Ab Initio) in the ETL system.

Portfolio

GfK - Growth from Knowledge
Spark, Python, Kubernetes, Docker, Apache Hive, YARN, Amazon Web Services (AWS)...
Data Reply
Ab Initio, Cloudera, Ansible, Amazon Web Services (AWS), Oracle, Apache Kafka...
Hopi
Apache Hive, Informatica, Talend, ETL, Oracle, HBase, RabbitMQ, PostgreSQL...

Experience

Availability

Part-time

Preferred Environment

Google Cloud Platform (GCP), Amazon Web Services (AWS), Python, SQL, Apache Airflow, Apache Hive, ETL, Apache Beam, Spark, Apache Kafka

The most amazing...

...system I implemented was a centralized data delivery infrastructure for sharing knowledge between teams and reducing similar tasks.

Work Experience

Senior Data Engineer

2020 - PRESENT
GfK - Growth from Knowledge
  • Oversaw new product development for running machine learning algorithms.
  • Built a science platform for data scientist teams.
  • Designed and implemented new data ingestions for the data scientist team.
  • Created impact analysis report for operation team. Prepared dashboards in Google Data Studio for their analysis.
  • Gathered data from different on-premises databases into GCP and prepared reports of the journey of the production time.
Technologies: Spark, Python, Kubernetes, Docker, Apache Hive, YARN, Amazon Web Services (AWS), Amazon S3 (AWS S3), Amazon Elastic MapReduce (EMR), Amazon Elastic Container Registry (ECR), Amazon Simple Notification Service (Amazon SNS), Amazon Simple Queue Service (SQS), Redshift, Google Cloud Platform (GCP), Google BigQuery, Apache Airflow, Oracle, Catalog Data Entry Services, Google Cloud, Pub/Sub

Consultant

2017 - 2018
Data Reply
  • Prepared the data platform for regulatory requirements with using Ab Initio. (GDE, BRE, EME) and Oracle DB.
  • Maintained three existing and two new Cloudera Hadoop Clusters with Ansible scripts. Kept clusters up-to-date with the latest versions. Secured clusters with Kerberos, TLS/SSL, and Sentry.
  • Added new services to the Cloudera Cluster (Cloudera Data Science. Workbench, Sentry) and installed new modules (Spark).
Technologies: Ab Initio, Cloudera, Ansible, Amazon Web Services (AWS), Oracle, Apache Kafka, Spark, Apache Hive, HBase, Impala, Apache ZooKeeper, Kerberos, SSL Certificates

Data Engineer

2017 - 2017
Hopi
  • Designed and implemented new projects in the data warehouse with Talend and Informatica.
  • Maintained data flow from Kafka to Hive and Oracle and PostgreSQL with. Storm and HiveQL.
  • Maintained hourly, daily, and monthly jobs in Informatica, Python, and HiveQL.
Technologies: Apache Hive, Informatica, Talend, ETL, Oracle, HBase, RabbitMQ, PostgreSQL, Python, Jenkins, Java, Apache Kafka, Hortonworks Data Platform (HDP), Git

DWH and Big Data Developer

2014 - 2017
Kredi Kayıt Bürosu (Credit Bureau)
  • Designed and implemented new projects in a data warehouse system with Informatica.
  • Developed a loyalty score project. Calculated a score based on how loyal customers were to their banks.
  • Designed and implemented a Factoring Datamart. Communicated report needs and built end-to-end projects.
  • Moved a semi-structured mainframe log file to a big data environment with PySpark.
Technologies: Oracle, Informatica, Microsoft SQL Server, Spark, PySpark

BI Consultant – Certified Ab Initio Technician

2010 - 2014
i2i Systems
  • Designed and developed an automated data migration and data masking tool in Ab Initio. Gathered metadata from Ab Initio EME and a data masking algorithm implementation with PDL.
  • Redesigned and re-engineered a revenue Datamart with PL/SQL. Re-implemented Ab Initio graphs in PL/SQL.
  • Maintained ~7,000 job (mostly Ab Initio) ETL system.
  • Improved the performance of tasks in Ab Initio and PL/SQL.
  • Led a team with five colleagues. Managed team communication with customers and upper management. Participated in the hiring process of this and other teams in the company.
Technologies: Ab Initio, Oracle, Microsoft SQL Server, Oracle PL/SQL, Data Modeling, Data Transformation, ETL, Algorithms, Data Encryption, Metadata

Automated Data Migration and Data Masking Tool

Migrating data for test environments from production data sources. For security reasons, data should be masked. I built a tool that masks data while being migrated with a unique algorithm that I designed. The tool also detects what kind of data is needed for testing specific code. For Ab Initio codes, we queried the Ab Initio metadata tool (EME).

Science Platform

I built a science platform used for research purposes. Data scientists need a platform where they can run research trials and machine learning algorithms. We ingested data from different data sources and built ETL data pipelines and used Data Catalog for labeling, which enabled users to easily find material. On GCP we used Big Query for data storing and Airflow for orchestrating the data pipelines.
2011 - 2014

Master's Degree in Engineering Management

Bahcesehir University - Istanbul, Turkey

2005 - 2010

Bachelor's Degree in Mathematics Engineering

Istanbul Technical University - Istanbul, Turkey

AUGUST 2019 - AUGUST 2020

Certified SAFe 4 Practitioner

Scaled Agile

APRIL 2018 - APRIL 2020

CCA Administrator

Cloudera

OCTOBER 2017 - OCTOBER 2020

AWS Solution Architect Certificate - Associate

AWS

JANUARY 2012 - JANUARY 2014

Ab Initio Technician

Ab Initio

Libraries/APIs

PySpark

Tools

Apache Airflow, Ab Initio, Git, Cloudera, RabbitMQ, Jenkins, Ansible, Impala, Apache ZooKeeper, Amazon Elastic MapReduce (EMR), Amazon Elastic Container Registry (ECR), Amazon Simple Notification Service (Amazon SNS), Amazon Simple Queue Service (SQS), BigQuery, Apache Beam, Composer

Paradigms

ETL, Agile

Storage

Databases, Oracle PL/SQL, Apache Hive, Microsoft SQL Server, PostgreSQL, Data Pipelines, Google Cloud, HBase, Amazon S3 (AWS S3), Redshift, Google Cloud Storage

Languages

SQL, Python, Java

Platforms

Oracle, Google Cloud Platform (GCP), Amazon Web Services (AWS), Talend, Apache Kafka, Hortonworks Data Platform (HDP), Kubernetes, Docker, Amazon EC2

Industry Expertise

Project Management

Frameworks

Spark, YARN, Hadoop, Scaled Agile Framework (SAFe)

Other

Analytical Thinking, Data Modeling, Algorithms, Data Transformation, Data Engineering, Data, ETL Development, Data Warehousing, Data Warehouse Design, Informatica, Software Development, Google BigQuery, Catalog Data Entry Services, Data Architecture, Google Data Studio, Human Resources (HR), Entrepreneurship, Data Mining, Finance, Data Encryption, Metadata, Kerberos, SSL Certificates, Pub/Sub, Google Pub/Sub, Scrum Master

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring