Ashish Khachane, Developer in Pune, Maharashtra, India
Ashish is available for hire
Hire Ashish

Ashish Khachane

Verified Expert  in Engineering

Bio

Ashish is an experienced data engineer and scientist with 7+ years of experience in the automotive and product sectors. Skilled in Python, PySpark, SQL, and diverse data science technologies, he has directed significant initiatives like Sales Excellence Enabled by Data (SEED), innovating AI models for predictive analytics and recommendation engines. Ashish excels in refining data frameworks, deploying scalable pipelines, and collaborating closely with stakeholders to drive strategic insights.

Portfolio

Salesforce
Python 3, SQL, Amazon SageMaker, Snowflake, Apache Airflow, GitHub, Docker...
Faurecia
Python, PySpark, SQL, Data Analytics, ETL, Palantir, Big Data, Machine Learning...
Faurecia
Python 3, PySpark, SQL, Data Engineering, ETL, Data Mining, Palantir...

Experience

  • Data Analytics - 7 years
  • Python - 6 years
  • Machine Learning - 5 years
  • Palantir Foundry - 4 years
  • ETL - 4 years
  • SQL - 4 years
  • PySpark - 4 years
  • Snowflake - 2 years

Availability

Full-time

Preferred Environment

A/B Testing, Artificial Intelligence (AI), Apache Airflow, Data Analytics, GitHub, Impact Analysis, Machine Learning, PySpark, Python, SQL

The most amazing...

...thing I've developed is a seller performance predictor model that I utilized to build a recommendation model to improve seller performance.

Work Experience

Senior Data Engineer

2023 - 2024
Salesforce
  • Developed the seller performance predictor (SPP) model using historical data and machine learning, which improved sales target prediction accuracy by 20%.
  • Designed and implemented a recommendation engine that enhanced sales performance by 15% through optimized recommendations based on SPP model insights.
  • Installed and managed scalable data pipelines for real-time and batch processing, improving data processing efficiency by 30% using Python, Amazon SageMaker, Snowflake, and Airflow.
  • Created the Accountability Performance Matrix (APM) dashboard, which provided insights into enablement programs and improved decision-making effectiveness by 25% through data correlation and analysis.
  • Contributed to the end-to-end development of data frameworks, optimizing analysis and modeling processes, while effectively communicating data insights and their business impacts via interactive dashboards.
Technologies: Python 3, SQL, Amazon SageMaker, Snowflake, Apache Airflow, GitHub, Docker, Data Matching, Unit Testing, A/B Testing, Pandas, Machine Learning, Artificial Intelligence (AI), SnowSQL, Python, Optimization, Data Mining, Big Data, Data Analytics, Data Analysis, Large Language Models (LLMs), Regression, Classification, ETL, Impact Analysis, Amazon S3 (AWS S3), Windows, Statistical Analysis, Microsoft Excel

Data Engineer

2019 - 2023
Faurecia
  • Launched the Event and Alert Management System (EAMS), reducing production downtime caused by material shortages by 20%.
  • Created a system to alert material planners of potential shortages based on inventory levels and customer demand forecasts, enhancing inventory management efficiency.
  • Developed data pipelines using PySpark SQL to automate inventory monitoring and alert generation, leading to more timely responses and improved inventory management.
  • Leveraged Palantir Foundry tools, including Contour and Code Workbook, to analyze and tailor data, boosting operational efficiency by 20% through better alignment of insights with business needs.
  • Conducted data health checks, unit testing, and code optimization to enhance performance and stability, while developing and maintaining schedules for data pipeline execution within the data lineage framework.
Technologies: Python, PySpark, SQL, Data Analytics, ETL, Palantir, Big Data, Machine Learning, Data Engineering, Windows, Python 3, Pandas, Statistical Analysis, Microsoft Excel, Palantir Foundry

Associate Data Engineer

2018 - 2019
Faurecia
  • Leveraged structured, semi-structured, and unstructured data sources to enhance data retrieval efficiency by 25% and improve overall data accessibility.
  • Created business intelligence dashboards that accelerated decision-making processes by 30% and increased insight accuracy across various departments.
  • Processed datasets for prescriptive and predictive modeling, reducing model training times by 40% and enhancing prediction accuracy by 15%.
  • Built machine learning models that improved performance by 35% and operational efficiency by 20%, while implementing measurement solutions for 100% data-driven decision-making and a 25% boost in strategic effectiveness.
Technologies: Python 3, PySpark, SQL, Data Engineering, ETL, Data Mining, Palantir, Data Analytics, Windows, Pandas, Statistical Analysis, Microsoft Excel, Palantir Foundry

Data Analyst

2016 - 2017
Brose
  • Created a KPI dashboard to monitor and enhance key performance metrics, leading to a 15% improvement in process efficiency.
  • Developed Python-based automated data pipelines to streamline data collection and processing from manufacturing systems, reducing data processing time by 25%.
  • Performed trend analysis and process capability studies, identifying patterns and providing actionable insights that improved product quality and operational efficiency by 20%.
Technologies: Statistical Analysis, A/B Testing, Windows, Data Analytics, Microsoft Excel

Experience

Sales Excellence Enabled by Data (SEED)

Conceptualized and implemented an SPP, an AI model that helps predict the sales target through historical data analysis and machine learning. I engineered a recommendation engine by implementing various business logic to enhance sales performance driven by the SPP model predictions and provided strategic guidance on data collection, storage, compliance, and quality assurance. Also, I contributed to end-to-end development of data frameworks, analysis, provisioning, modeling, and optimization. My other responsibilities included:

• Installing, managing, and optimizing scalable data pipelines for real-time and batch processing using Python, Amazon SageMaker, Snowflake, and Airflow.
• Demonstrating proficiency in utilizing version control and CI/CD systems to maintain efficient development workflows.
• Constructing large-scale, high-performance data structures for analytics purposes.
• Collaborating effectively with both technical and business stakeholders to drive data-driven decisions.
• Managing data collection, cleaning, utilization, and storage processes.
• Communicating data insights and coherently articulating their business impact through various storytelling formats such as slides, charts, and dashboards.

Event and Alert Management System (EAMS)

Deployed the EAMS project, reducing production downtime due to material shortages by 20%. I established a system to notify material planners of a potential material shortage based on inventory levels and customer demand forecasts, streamlining inventory management processes, and improving overall efficiency. My other duties included:

• Building data pipelines using PySpark SQL to automate inventory monitoring and alert generation by comparing material inventory against business-set thresholds.
• Integrating customer demand forecasting to ensure accurate material requirements.
• Leveraging Palantir Foundry tools, including Contour and Code Workbook, to analyze and tailor data, aligning insights with business requirements and increasing operational efficiency by 20%.
• Carrying out data health checks, data expectations, unit testing, code optimization, and tuning strategies to enhance performance and minimize resource consumption, fortifying data pipelines and ensuring their stability.
• Developing and maintaining schedules for executing data pipelines in the data lineage framework.

Accountability Performance Matrix (APM) as a Service

Analyzed the effectiveness of enablement programs aligned with business outcomes. I also quantified the impact of various treatments—like plays, certifications, and planning processes—on outcomes such as annual contract value (ACV), pipeline generation, and customer renewal. Additionally, I:

• Provided insights into the impact of events like training sessions and marketing campaigns.
• Pioneered APM as a comprehensive tool correlating employee attributes with treatment outcomes.
• Crafted an accountability performance matrix (APM), a dashboard driven by a data-model engine.
• Facilitated informed decision-making by comparing and analyzing data.

Generalized A/B Testing as a Service

I designed and implemented a scalable A/B testing framework to assess the effectiveness of business interventions by comparing treated and control groups across various business units. I also conducted statistical analyses and hypothesis testing to confirm the significance of observed differences between treated and control groups, ensuring reliable and data-driven insights. Additionally, I quantified the impact of interventions, identifying a measurable improvement in outcomes (e.g., a 10% increase in sales) and isolating key drivers for success.

Education

2017 - 2019

Master's Degree in Mechanical Engineering

Savitribai Phule Pune University - Pune, India

2012 - 2016

Bachelor's Degree in Mechanical Engineering

University of Pune - Pune, India

Certifications

SEPTEMBER 2022 - PRESENT

Advanced Certification in Data Science and AI

Center for Continuing Education, IIT Madras

Skills

Libraries/APIs

Pandas, PySpark

Tools

Amazon SageMaker, GitHub, SnowSQL, Microsoft Excel, Microsoft Power BI, Apache Airflow, Tableau

Languages

Python 3, SQL, Snowflake, Python, R

Paradigms

Unit Testing, ETL, Mechanical Design

Platforms

Palantir Foundry, Docker, Windows

Storage

Amazon S3 (AWS S3)

Other

Mechanical Engineering, Data Matching, A/B Testing, Data Analytics, Statistical Analysis, Machine Learning, Data Mining, Optimization, Artificial Intelligence (AI), Impact Analysis, Product Design, Data Engineering, Palantir, Big Data, Data Analysis, Large Language Models (LLMs), Regression, Classification, Hypothesis Testing

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring