Scroll To View More
Sagar Sharma, Business Intelligence (BI) Developer in Toronto, ON, Canada
Sagar Sharma

Business Intelligence (BI) Developer in Toronto, ON, Canada

Member since July 11, 2016
Sagar is a seasoned data professional with more than seven years of work experience with relational databases and three years with big data—specializing in designing and scaling data systems and processes. He is a hardworking individual with a constant desire to learn new things and to make a positive impact in the organization. Sagar possesses excellent communication skills and is a motivated team player with the ability to work independently.
Sagar is now available for hire



  • Databases, 10 years
  • SQL, 10 years
  • Business Intelligence (BI), 8 years
  • Python, 7 years
  • SQL Server 2012, 7 years
  • MySQL, 6 years
  • Pentaho Data Integration (Kettle), 6 years
  • Hadoop, 5 years
Toronto, ON, Canada



Preferred Environment

macOS/Linux, Sublime Text, IntelliJ IDEA

The most amazing...

...project was building a data lake with Hadoop—this included a Hadoop ecosystem installation from scratch and building data pipelines to move data into the lake.


  • Data Engineer

    2017 - PRESENT
    • Designed, developed, and am now maintaining a reporting data warehouse built using PostgreSQL (AWS RDS).
    • Created an ETL framework using Python and PDI to load data into a data warehouse.
    • Connected to third-party APIs to import data on an incremental basis e.g., Salesforce, Sailthru, and CrowdTwist.
    • Managed PostgreSQL (RDS) and ETL servers (EC2) using AWS Console.
    • Created reports and dashboards in Looker to provide insights on the data.
    Technologies: Python, Amazon RDS for PostgreSQL, MySQL, Salesforce, MySQL
  • Senior Business Intelligence Engineer

    2016 - 2018
    Altus Group Limited
    • Built a reporting data warehouse using Pentaho, PostgreSQL, and Informatica.
    • Designed a database schema in PostgreSQL to represent the reporting use case.
    • Created ETL tasks in Informatica to move data from the production systems into PostgreSQL.
    • Built reports and dashboards using a Pentaho report designer and deployed them to the BI server.
    Technologies: PostgreSQL, MS SQL, Pentaho Reporting, Informatica
  • Data Engineer

    2014 - 2016
    Wave Accounting, Inc.
    • Designed, developed, and maintained big data and business intelligence solutions at Wave.
    • Designed and scheduled complex ETL workflows and jobs using Pentaho Data Integration (Kettle) to load data into the data systems.
    • Wrote custom Python scripts to access third party APIs and download data into the data systems.
    • Developed complex SQL queries including JOINS, subqueries, and common table expressions to address ad hoc business analytics and other requirements.
    • Coordinated with the product and executive teams to gather and understand business requirements.
    • Built an end-to-end relational data warehouse—including infrastructure, schema design, optimization, and administration.
    • Designed and developed a Hadoop Cluster using Horton Works HDP 2.0. Tasks include installing and configuring a Hadoop ecosystem and designing the HDFS.
    • Designed and scheduled Sqoop jobs to load data into the HDFS from the production systems.
    Technologies: Python, MS SQL, Pentaho Kettle, Sisense, PostgreSQL, Hadoop, MySQL, Hive, Sqoop, Ansible
  • Business Intelligence Developer

    2011 - 2014
    Eyereturn Marketing, Inc.
    • Designed real-time reporting solutions using a SQL server (SSIS, SSAS and SSRS) and Pentaho business intelligence tools (MySQL, Mondrian, and Pentaho).
    • Created custom automated/scheduled reports using Eclipse BIRT and Pentaho Report Designer.
    • Built custom ETL tasks to transform data for custom reports using Kettle (Pentaho Data Integration).
    • Designed and optimized database schemas to make reporting faster and efficient.
    • Created, maintained, and scheduled custom data processors to pull and manipulate data from HDFS using Pig, Sqoop, and Oozie (Cloudera Hadoop).
    Technologies: MS SQL, SSIS, SSAS, Pentaho, Hive, Hadoop, Pig, Sqoop, MySQL
  • Database Analyst

    2010 - 2011
    George Brown College
    • Handled and was responsible for the database administration in the organization using Blackbaud’s Raiser’s Edge.
    • Updated and maintained the alumni database using the MS SQL Server.
    • Conducted data validation and verification to ensure the accuracy and quality of the data.
    • Performed multiple queries at a complex level for the purposes of reports and provide information for divisional and marketing purposes.
    • Provided support to the project managers.
    Technologies: MS SQL Server, Raiser's Edge
  • Software Engineer

    2007 - 2009
    Tata Consultancy Services
    • Provided post-implementation support and training for an enterprise level banking application (TCS [email protected]) to 25,000+ corporate end-users.
    • Handled different modules of the banking operations such as routine banking, loans and mortgages, capital markets, and foreign exchange.
    • Analyzed client business needs and translated them into functional/operational requirements.
    • Communicated successfully with a variety of people including subject matter experts to establish a technical vision, business units, development teams, and support teams.
    Technologies: Java, HTML, SQL, Oracle


  • Data Lake Using Hadoop (Hortonworks HDP 2.0) (Development)

    I built a data lake using Hadoop.
    My tasks included:
    - Installing and configuring Hadoop ecosystem components on the RackSpace Cloud Big Data platform.
    - Automating the above process using Ansible.
    - Designing and scheduling Sqoop jobs to import data from MySQL and PostgreSQL production tables into HDFS.
    - Setting up Hive tables from HDFS files to enable SQL-like querying on HDFS data.

  • Sisense Rebuild Project (Development)

    I built a data warehouse for a global animal health company. The goal was to help business users by giving them a single source of truth about their mobile application.

    The project included the following tasks:
    1) Redesigning the data model in Sisense ElastiCube. I used a snowflake schema to establish database relationships.
    2) Creating a build schedule of the above created data model.
    3) Creating dashboards, based on the client's requirements.
    4) Setting-up email schedules for the dashboards to relevant stakeholders.


  • Languages

    SQL, Python, CSS, HTML, Transact-SQL, JavaScript, Java, Scala, SAS
  • Tools

    Looker, Pentaho Data Integration (Kettle), Informatica ETL, Pentaho Mondrian OLAP Engine, Sisense, Apache Sqoop
  • Paradigms

    ETL Implementation & Design, ETL, MapReduce
  • Storage

    SQL Server 2008, Databases, SQL Server 2012, SQL Server 2014, AWS RDS, SQL Server Integration Services (SSIS), HDFS, SQL Server Analysis Services (SSAS), Apache Hive, PostgreSQL, MySQL, MongoDB
  • Other

    ETL Tools, ETL Development, Pentaho Reports, RESTful APIs, Semantic UI, Data Analysis, Pentaho Dashboard, Informatica, Business Intelligence (BI), Dialectical Behavior Therapy
  • Frameworks

    Express.js, Bootstrap 3+, Materialize CSS, Foundation CSS, Flask, Hadoop
  • Libraries/APIs

    REST APIs, NumPy, Pandas, DreamFactory, React, Passport.js, Node.js
  • Platforms

    AWS EC2, Windows, MacOS, Linux, Apache Pig


  • Certificate in SAS Certified Base Programmer
    2011 - 2011
    SAS Institute Canada - Toronto, Canada
  • Postgraduate certificate in Strategic Relationship Marketing
    2010 - 2010
    George Brown College - Toronto, Canada
  • Bachelor's degree in Mechanical Engineering
    2003 - 2007
    YMCA University of Science and Technology - Faridabad, India
I really like this profile
Share it with others