Narasimhan MB, Developer in Bengaluru, Karnataka, India
Narasimhan is available for hire
Hire Narasimhan

Narasimhan MB

Verified Expert  in Engineering

Data Engineer and Python Developer

Bengaluru, Karnataka, India

Toptal member since December 5, 2024

Bio

Narasimhan has 8+ years of experience in the data domain, specializing in roles ranging from Python developer to data, ML, and analytics engineer. He excels in optimizing data pipelines, delivering insights, and aligning solutions with business objectives. Focused on automation, he emphasizes high data quality and scalable, reliable solutions by integrating software engineering practices into data engineering and ML. He's passionate about expanding his expertise in distributed systems and MLOps.

Portfolio

A Consulting Firm - Multiple clients
Spark, AWS IoT, Python, Foundry, Machine Learning, Performance Engineering...
Automotive Client
Palantir Foundry, PySpark, Data Science, Data Visualization, Real time Analytics

Experience

  • Python - 7 years
  • Data Analytics - 6 years
  • Spark - 4 years
  • AWS Cloud Architecture - 4 years
  • Data Warehouse Design - 4 years
  • Data Engineering - 4 years
  • Machine Learning - 2 years
  • Foundry - 2 years

Availability

Full-time

Preferred Environment

Spark, Python, Palantir, Machine Learning, Data Engineering, Data Analytics, AWS Big Data, Data Warehouse Design, Data Visualization, Business Communication

The most amazing...

...goal I've achieved is being able to deliver data and analytical solutions that are aligned with clients' business objectives.

Work Experience

Senior Data Engineer

2016 - PRESENT
A Consulting Firm - Multiple clients
  • Spearheaded initiatives to optimize data pipelines and AWS architecture, resulting in an almost threefold increase in cloud infrastructure efficiency and an approximately 55% reduction in cloud spending.
  • Directed two end-to-end projects, including one focused on near real-time data processing, from translating business requirements to implementing scalable data pipelines and actionable dashboards. These efforts resulted in a twofold increase in end user engagement and a nearly 20% faster time to market (TTM).
  • Initiated discussions on key performance indicators (KPIs) for customer segmentation and conducted funnel analysis and churn modeling, leading to a 1.3 times increase in customer growth quarter-on-quarter (QOQ).
  • Optimized ticket resolution by performing extensive data wrangling and applying advanced statistical techniques to address highly imbalanced data. This resulted in a 5% improvement in the accuracy of the multiclass classification model.
  • Developed a time series forecasting model for financial planners, seamlessly integrating it into the QlikSense dashboard and enabling the marketing team to optimize ad campaign spending based on accurate, data-driven insights.
  • Built a tool to read data from unstructured sources, which was then processed by a self-developed named-entity recognition (NER) model, achieving approximately 70% accuracy. This tool was used to identify and extract personally identifiable information (PII) as part of GDPR compliance initiatives.
Technologies: Spark, AWS IoT, Python, Foundry, Machine Learning, Performance Engineering, Data Science

Lead Data and Analytics Engineer

2021 - 2023
Automotive Client
  • Directed two end-to-end projects, including a near real-time, from translating business requirements to implementing scalable data pipelines and actionable dashboards, resulting in a ~2x increase in end user engagement and ~20% faster time to market.
  • Identified as an SME on Palantir Foundry. Worked on a range of applications on Palantir Foundry, such as Code Repositories, Monocle, Ontology, Workshop, Contour and Code Workbook, Quiver, and Pipeline Builder.
  • Developed code for image processing by understanding low-level concepts of the Palantir Foundry file system and internal architecture. This was recognized by the Palantir team as a novel solution, helping the team win the “Best Project Award”.
  • Implemented incremental processing on data pipelines for higher efficiency. Also implemented changelog, which acts as incremental indexing with Ontology.
  • Created a one-stop dashboard on Foundry's workshop application. This involved processing data from various sources, including MS Excel, and creating an incremental pipeline. Used object monitoring for Ontology updates.
Technologies: Palantir Foundry, PySpark, Data Science, Data Visualization, Real time Analytics

Experience

Automated Infrastructure Deployment

A large-scale infrastructure deployment on AWS that I designed and automated, starting with framework design, tool selection, and environment setup. I implemented best practices for scalable code architecture, incorporating mock tests for quality assurance and early error detection. I also managed cloud security aspects and streamlined infrastructure deployment through automation using GitHub Actions, with optional multiregional deployment support to ensure flexibility and resilience.

Near Real-time Analytics Dashboard App

As the lead data engineer, I spearheaded a project focused on real-time monitoring of customer shipments to prevent delays. We used Palantir Foundry's Pipeline Builder to build real-time pipelines. I oversaw the entire data lifecycle, from gathering business requirements to implementing scalable, near-real-time data pipelines.

I also developed visualization dashboards that aligned with key business objectives using Workshop and Quiver.

This solution enabled the business to proactively prevent shipment delays while enhancing manufacturing and logistics efficiency through just-in-time production. By reducing unnecessary production, the system optimized operations and improved overall supply chain effectiveness.

Much work during this time required us to extensively use Foundry's Contour, Object Explorer, and Ontology to build and analyze data models and relationships.

Improving BOM Efficiency

As a data engineer, I contributed to a project aimed at analyzing the efficiency of parts production in the automotive sector.

During the analysis, I identified an issue with a key metric used to monitor BOM efficiency: the weight of the parts produced compared to the weight of the theoretical BOM recipe. To investigate further, I focused on several materials with low production efficiency. Through exploratory analysis and reviewing the data distribution leveraging Foundry's Contour and Code workbook, I discovered that the weight of raw materials and components had been recorded in kilograms (kg) instead of grams (g) due to an error in the ERP system's data entry process.

This discovery prompted the implementation of comprehensive data quality checks using Palantir Foundry's Health checks application in the data pipeline and leveraging Foundry's Expectations API at the code level. These measures ensured accurate metrics on the dashboard, as these metrics directly impacted business decisions.

Education

2005 - 2009

Bachelor's Degree in Electronics and Communication

Visvesvaraya Technological University (VTU) - Karnataka, India

Certifications

SEPTEMBER 2024 - SEPTEMBER 2027

AWS Certified Data Engineer – Associate

Amazon Web Services

JANUARY 2023 - JANUARY 2025

Palantir Certified Foundry Data Engineer – Professional

Palantir Technologies

Skills

Libraries/APIs

PySpark

Tools

Terraform

Languages

Python

Frameworks

Spark

Paradigms

DevOps

Platforms

Apache Flink, Apache Kafka, AWS IoT, Palantir Foundry

Storage

Data Pipelines

Other

Palantir, Data Engineering, Data Analytics, AWS Big Data, Data Warehouse Design, Data Visualization, AWS Cloud Architecture, Foundry, Machine Learning, Business Communication, GitHub Actions, Streaming Data, Performance Engineering, Data Science, Programming, Operating Systems, Real time Analytics, Real time analytics

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring