Rajesh Golani, Developer in Mumbai, Maharashtra, India
Rajesh is available for hire
Hire Rajesh

Rajesh Golani

Verified Expert  in Engineering

Web Scraping Developer

Mumbai, Maharashtra, India

Toptal member since February 29, 2020

Bio

With over 16 years of experience and background in system design, Rajesh is proficient in a variety of platforms, languages, and cloud-based frameworks. He can fulfill the role of back-end architect, data engineer, and full-stack developer. He has an interest in natural language processing and machine learning.

Portfolio

Fastgraphs (via Toptal)
JavaScript, HTML, Full-stack, MongoDB, Python, Django, Angular
Azim Premji Philanthropic Initiative
Amazon Web Services (AWS), Selenium, Scraping, Pandas, Python
Libtech India
Amazon Web Services (AWS), MySQL, Data Analytics, Angular, REST APIs, Django...

Experience

  • Python - 12 years
  • Web Scraping - 8 years
  • Flask - 8 years
  • Django - 8 years
  • Data Analysis - 8 years
  • Django REST Framework - 8 years
  • Angular - 3 years
  • Natural Language Processing (NLP) - 3 years

Availability

Part-time

Preferred Environment

Git, REST APIs, Python, Vim Text Editor, Ubuntu

The most amazing...

...project I've developed was a web-based lead-identification system for a client that increased closed sales by more than 20% in the first year.

Work Experience

Full-stack Developer

2020 - PRESENT
Fastgraphs (via Toptal)
  • Kicked off the re-development of the legacy financial app from AngularJS to Angular.
  • Helped design the API to enable communication between Angular front end and Django back end using Django Rest Framework.
  • Used HiCharts library for visualization and graphs.
  • Developed JSON Data Formats for efficient retrieval of large datasets from Mongo DB via Django API.
  • Used Bootstrap4 flex layout for a responsive web application.
Technologies: JavaScript, HTML, Full-stack, MongoDB, Python, Django, Angular

Data consultant

2019 - PRESENT
Azim Premji Philanthropic Initiative
  • Developed a crawling infrastructure using python for scraping Indian Government public data websites.
  • Extensively worked on Selenium, requests, regular expressions, Queues, Threading to effectively scrape data.
  • Worked on Python Pandas for data parsing and data processing making it available via the CSV.
  • Used Amazon S3, as primary storage engine, using boto3 python package to create, update and retrieve files from S3.
  • Developed a dashboard to access various reports in csv format using Angular as frontend and Django as backend.
Technologies: Amazon Web Services (AWS), Selenium, Scraping, Pandas, Python

Researcher and Data Engineer

2017 - PRESENT
Libtech India
  • Developed a complete backend in Django with powerful REST API to access government data.
  • Crawled millions of pages using requests and selenium from the government website.
  • Developed a chatbot application on top of the REST API to retrieve/search-relevant government data.
  • Developed an Angular-based data dashboard to access reports and analytics.
Technologies: Amazon Web Services (AWS), MySQL, Data Analytics, Angular, REST APIs, Django, Python

Full Stack Developer

2020 - 2020
Coast India Network
  • Developed a web application in association with Government of India to track migrant labourers during the ongoing Covid Epidemic.
  • Developed a complete backend in Django using Django Rest Framework for API, and MYSQL as database.
  • Integrated with Facebook and Google API to provide social media logins in Angular Frontend via Django Rest Framework.
  • Extensively used JSON fields in MYSQL to store and retrieve NOSQL like data.
  • Developed a admin dashboard using Angular8 with various filters and export data capability.
Technologies: JavaScript, HTML, Full-stack, Amazon Web Services (AWS), Python, Angular, Django REST Framework, Django

Data Engineering Contractor

2019 - 2020
Massive Data Institute George Town University
  • Scraped millions of rejected bank payment records for public beneficiaries from government website.
  • Extensively used Queuing and threading frameworks in python to accelerate scraping.
  • Analysed these records using Pandas/Apache spark and to find anomalies and trends for rejected transactions.
  • Joined various data tables using common fields to develop a full chain of events leading to rejected transactions.
  • Worked on building a model to predict payment rejection using Machine Learning and AI.
Technologies: MySQL, Pandas, Selenium, Scraping

Full Stack Web Development Consultant

2016 - 2019
Cache Consultants
  • Consulted on basic architecture for various web applications.
  • Developed full stack solutions for various clients primarily using Django, AWS and Angular.
  • Explored various python visualisation libraries like Altair and Plotly.
  • Created infrastructures for using various Google services such as Oauth and Google location and places API.
  • Used AWS services primarily S3, DyanamoDB, EBS for various applications.
Technologies: HTML, Full-stack, Flask, Amazon Web Services (AWS), Django REST Framework, Angular, Django, Python

Engineering Consultant

2013 - 2016
Program of Liberation Technologies Stanford University
  • Scraped millions of pages of public websites to gather data related to various public services in India.
  • Extracted, parsed and processed data from html in to csv tables.
  • Used mysql and Amazon S3 as primary storage services.
  • Developed dashboards for access to this information, using PHP/MYSQL and Django/Angular.
  • Developed a voice broadcast system based on cloud telephony services.
  • Co-Authored a paper on “Analysis of Payment Delays and Delay Compensation in MGNREGA” which has been used in PILs in supreme court of India.
Technologies: Cloud Telephony, PHP, Django, Python

Senior Field Application Engineer

2007 - 2013
Xilinx India Development Corporation
  • Helped customers architect their products around programmable chips.
  • Presented and conducted many technical workshops.
  • Led a team of 10+ engineers to form a robust technical support team.
Technologies: FPGA, Verilog, Python

Experience

Tracking Migrant Movement During Covid Crisis

I was part of the team which developed a web application to track each group of migrants and make coordination easier between different stakeholders. Currently, the app is being adopted by the Jharkhand state of the government of India.
My primary role was to develop a back end using Django, MYSQL, and an admin dashboard using Angular. I used JSON fields in JSON to store NoSQL kind of data. Enabled social media logins by integrating Facebook and Google API. I also used Google's location services to track the location of migrants and provide a Map view of the groups of migrants.
The tech stack is Django, Django Rest Framework, Angular, and Java.

Lead Identification system

Developed a web-based lead identification system, which increased the closed-sales by more than 20% in the first year. My primary task was to process, classify and tag the leads generated by marketing team, based on past sales and additional acquired metadata from the internet. These were then presented to the sales team via a web interface, which was powered by Angular front end and Flask back end.

Broadcast Management System

Developed a simple dashboard to enable scheduling of voice broadcasts and voice IVRs. The application interfaced with various cloud telephony services such us exotel, tringo and awaz De. There was a facility to manage phone numbers and create groups to manage subscribers. Implemented an efficient scheduling algorithms for maximum outreach and 100+ concurrent voice calls.
This system was serving over 50,000 subscribers and a million plus voice calls.
The tech stack of the project was PHP/Mysql and later migrated to Python/Django/MySQL and Angular.

Portal for Writings of Gandhi

A Django/Angular application to display works of Gandhi in easy to view format. All the writings of Gandhi are available in the public domain in the form of 98 PDF books. The primary task was to implement natural language processing on the extracted text from pdf, to identify dates, context, and entities from the articles. Each of the articles is extracted, cleaned, and indexed into the database for easy search and retrieval. A simple front end has been developed powered by Angular.

Public Data Portal

Data related to public welfare schemes is distributed on various government websites. The idea of the project was to download such data and generate analytics and actionable information that can be used for public policy discourse.

My primary task as a back-end engineer was data crawling, data modeling, and data analytics. The technology stack consisted of MySQL, Django web framework, Django Rest Framework, PySpark, AWS Services, along with various parsing and crawling libraries in python. A substantial amount of effort was put in Queuing frameworks to download data faster and accurately.

Education

1998 - 2002

Bachelor of Technology Degree in Electrical Engineering

Indian Institute of Technology Madras - Chennai, India

Skills

Libraries/APIs

Pandas, REST APIs, Beautiful Soup

Tools

Vim Text Editor, Git, Jupyter

Languages

HTML, JavaScript, Python, Verilog, PHP

Frameworks

Django, Flask, Angular, Bootstrap, Selenium, Django REST Framework

Paradigms

REST

Platforms

Ubuntu, Amazon Web Services (AWS), Firebase, Linode, Jupyter Notebook

Storage

MySQL, Cloud Firestore, MongoDB

Other

Data Analytics, Cloud Telephony, Scraping, Full-stack, APIs, Web Scraping, Data Visualization, Data Analysis, Data Modeling, Natural Language Processing (NLP), Machine Learning, Generative Pre-trained Transformers (GPT), FPGA, Chatbots

Collaboration That Works

How to Work with Toptal

Toptal matches you directly with global industry experts from our network in hours—not weeks or months.

1

Share your needs

Discuss your requirements and refine your scope in a call with a Toptal domain expert.
2

Choose your talent

Get a short list of expertly matched talent within 24 hours to review, interview, and choose from.
3

Start your risk-free talent trial

Work with your chosen talent on a trial basis for up to two weeks. Pay only if you decide to hire them.

Top talent is in high demand.

Start hiring