Data Science Classroom & Online Demo on 28th September 2019 @11 AM in Mumbai for details contact : 9632156744. Register Now & Avail early bird price

ExcelR Jumbo PassExcelR Jumbo Pass Video   You May Have Heard About Offers, But Have You Heard Of ExcelR's JUMBO PASS? Well, Here's Your Chance To Avail The JUMBO PASS!! Watch The video

Data Analytics course Key Benefits

*Note: TCS Certification Cost Additional

Upcoming Batches

Select Your City



9:00 AM-1:00 PM Online ₹ 49000

Valid till 28 Sep 2019

Enroll Now

Filling Fast



10:00 AM-6:00 PM Online+Classroom ₹ 49000

Valid till 28 Sep 2019

Enroll Now

Filling Fast

Click Here

Course Description

ExcelR, as a leading institute of Data science courses in Mumbai , pay a lot of attention to the needs and requirements of the job market by keeping an eye out on the recent trending technologies. One such current trend that has swept the market away is Data Science for which there are myriad data science courses and here one can find the best Data science training. Today, almost every industry, domain and organization in them are using Data science as one of the major tools to help their business grow and adapt according to the demand and preferences of the customers. ExcelR’s curriculum, faculty and post training support is considered to be the best in the industry tailored to the needs of the Data Science job market. ExcelR’s Data Science course agenda has been meticulously designed with R Programming, Python , Machine Learning, Forecasting and Tableau addressing the complete Data life cycle.

What is Data Science & Data Science Course benefits?

In general terms, Data science is a pool of tools and techniques used to simplify the data so that it can be used to make business decisions. Owing to digitalization humongous data generation is in place which is waiting to be understood and visualized in a way that it can be useful by giving some insight into the subject.
Big companies tend to collect the data regarding the customer's behaviors, buying behavior, their likes and dislikes, reviews, etc. which will help them to evaluate their marketing plans, their products and services, all of which one can learn from our data science training in Mumbai. 
In data science, data is collected, cleaned and then visualized to find the hidden patterns and get a prescriptive or a predictive viewpoint. Data science is a mixture of statistics, mathematics, computers, algorithms and some business acumen which makes it approachable and widespread in many ways. We, as the best Data Science Institute in Mumbai, make it a point to train our students, regardless of their domain using the best methods possible, master the concepts and help to build a successful career in the space of Data Science. 
Some of the reasons why data science has become so significant in the organizations and why one should learn it from the best data science training institute are:

  • find new challenges of the market and try to get answers for them
  • helps in understanding the customers in a better way
  • development of products and services
  • find new trends and take actions accordingly
  • test the decisions and make them more refined
  • make quantifiable decisions

Data Science Course in demand 

Data science throughout the last decade has been showing phenomenal growth and that is why one can also experience immense career growth in it as well. This is one of the reasons why our data science face-to-face training and online training has become so popular among new professionals and students. Some of the points proving that data science is highly demanded today are: 

  • According to Glassdoor, the average salary of a data scientist in a company can range from 5-100 lac per annum which is higher than many IT based fields or any field in general. Most of the companies are paying higher salaries to their data scientists almost as comparable to the global standards.
  • As presented by NDTV, data science is a sector that Is facing manpower shortage and therefore making this the best time to grab the opportunity to become a data scientist and get placed in a respected position. Also, data science is supposed to see a growth of almost 8 times by 2025, making it one of the most pursued career choices.
  • According to India today, the average Indian data science position pays around 11 lacs per year. The highest number of jobs is in Bangalore, followed by Delhi, Mumbai, Pune, and others. The companies that hire the data scientists the most are Accenture, IBM, KPMG, Deloitte, Honeywell, Wells Fargo, Amazon, Dell, etc. Mostly finance and banking sectors hire data scientists followed by healthcare, energy, e-commerce, media, etc. 
  • According to Economic times, India has been churning jobs in the field of data science in the last few years. With almost a 400% rise in the job vacancies and requirements and around 1.5 million job openings, the demand is quite high for qualified data science specialists. 
  • According to the Times of India, in the last few years, it is seen that the salary of data scientists has increased by approximately 20%, especially in startups and new ventures. The demand for data scientists has risen to complete various tasks in the tech companies which requires detailing and analyzing large amounts of data for various projects and for developing new products.
  • PWC states that the number of positions for data science is quite high, but the requirement can be incomplete because of the lack of proper skills and knowledge. This is the reason why getting trained and skilled in data science and build a successful career and this why ExcelR is considered to be the  best Data Science Institute in Mumbai
  • Harvard has expressed that data science is a subject that everybody wants to talk about and being a data scientist is also the most lucrative job of the decade. The interested party can capitalize the raw data into something that can catalyze growth for many large-scale businesses.

What are the USPs of ExcelR Data Science Course? 

There are many points as to why students and professionals from all kinds of streams and sectors choose to join us at our data science institute. With our data science program, one can learn all of the details of data science and how to approach the subjects to achieve a promising job in the end. USPs that make us the best Data Science Institute in Mumbai are:

  • Attend as many lectures and batches you want to with the JUMBO PASS throughout the time span of an entire year.
  • Learn from the experienced industry experts with long-term teaching experience and data scientists with hands-on knowledge of the field and market passed down from IIT, IIM, and ISBs. 
  • Take advantage of the classroom programs, live sessions from the instructors and experts and recorded E-Learning videos that can be watched anytime according to one’s own comfort.
  • 2 capstone projects where participant’s will work on full length Data Science lifecycle
  • Practice and hone in on the skills with the help of more than 60 topic wise assignments.
  • Get trained in data science with more than 160 hours of lectures and sessions.
  • Get certified from Tata Consultancy Services (TCS-Ion)
  • Machine Learning and Artificial Intelligence concepts as part of Data Science Course will be provided
  • Get all the help needed after the completion of the course with mock interviews and resume building.
  •  Get full time placement assistance from our data science institute in several companies. 
  • ExcelR offers best Online training for data science Mumbai.

Who can take the Data Science Certification Course?

Data science has become such a pervasive field that almost everybody can be a part of this trend by learning new technologies and skills from our Data Science Institute in Mumbai. Anybody with a knowledge of mathematics, analysis, and business along with strong logical and analytical skills can be a part of this bandwagon of receiving the data science certification
Those who can be successful by joining the data science course are:

  • Business analysts
  • Market analysts
  • Software programmers
  • Statisticians
  • Mathematicians
  • Economists
  • Six Sigma consultants
  • Domain specialists
  • Freshers with good analytical skills

What is Covered in the Data Science Course Curriculum?

 We understand the requirements of the market and what is demanded by the industry today and that is why we do our best to keep our data science training and data science online training module as available as possible. We keep the curriculum updated at all times by adding new technologies and topics of the market. A few of the topics topics covered in the data science course training as part of the curriculum :

  • Python
  • R studio
  • SQL
  • Data Collection
  • Data Cleansing / Feature Engineering/ Exploratory Data Analysis
  • Statistical Analysis
  • Hypothesis Testing 
  • Regression-Linear Regression
  • Logistic Regression
  • Discrete Probability Distribution
  • Advanced Regression
  • Forecasting
  • Data Visualization with Tableau
  • Text Mining
  • Data Mining Supervised- Naïve Bayes
  • Machine Learning 
  • KNN
  • SVD
  • Decision Tree
  • Random Forest
  • Bagging And Boosting
  • Black-box technique-SVM
  • Neural Network
  • Data Mining Unsupervised-Clustering
  • Association Rule
  • Data decomposition Techniques- PCA
  • Natural Language Processing

Things You Will Learn


  • SQL Command
    • Data Query Language (DQL)
    • Data Manipulation Language (DML)
    • Data Definition Language (DDL)
    • Data Control Language (DCL)
  • SQL RDBMS Concept – Features & Advantages
    • Tables
    • Field
    • Record or a Row
    • Column
    • Constraints
  • RDBMS – Database Normalization
  • Primary and Foreign Keys
  • Index in RDBMS
  • SQL Data Types | Data Types in SQL Server
  • Clause in SQL 
    • WITH Clause 
    • SELECT Clause
    • FROM Clause
    • WHERE clause
    • GROUP BY clause
    • HAVING Clause
    • ORDER BY Clause
  • SQL Operators – Arithmetic, Comparison, & Logical
  • Operators in SQL – Alias, IN and Between
  • Create Database | SQL Drop & Select Database
  • SQL Join – Inner, Left, Right & Full Joins
  • SQL Index – Create, Unique, Composite Index
  • SQL Functions
    • Date
    • SUM
    • Count
  • Stored Procedure in SQL
  • Triggers in SQL

Introduction to R Programming

  • Introduction to R
  • Data Types in R

How To Install R & R Studio

Data Structures in R

  • Variable in R
  • R-Overview
    • Vector
    • Matrix
    • Array
    • List
    • Data-Frame
  • Operators in R
    • arithmetic
    • Relational
    • Logical
    • Assignment
    • Miscellaneous
  • Conditiional Statement
    • Decision Making<
      • IF Statement
      • IF-Else Statement
      • Nested IF-Else Statement
      • Switch Statement
    • Loops
      • While Loop
      • Repeat Loop
      • For Loop
    • Strings
    • Functions
      • User-defined Function
      • Calling a Function
      • Calling a Function without an Argument
      • Calling a Function with an Argument

Programming Statistical

  • Box Plots
  • Bar Charts
  • Histogram
  • Pareto Chart
  • Pie Chart
  • Line Chart
  • Scatterplot

How to Import Dataset in R

  • Read CSV Files
  • Read Excel Files
  • Read SAS Files
  • Read STATA Files
  • Read SPSS Files
  • Read JSON Files
  • Read Text Files


  • DpLyr
  • Hmisc or mise
  • Ggplot2
  • Caret
  • Data Table

How to Integrate R and SQL

How to Get Data From SQL to R

Description: You will get an introduction to the Python programming language and understand the importance of it. How to download and work with Python along with all the basics of Anaconda will be taught. You will also get a clear idea of downloading the various Python libraries and how to use them.


  • About ExcelR Solutions and Innodatatics
  • Do's and Don’ts as a participant
  • Introduction to Python
  • Installation of Anaconda Python
  • Difference between Python2 and Python3
  • Python Environment
  • Operators
  • Identifiers
  • Exception Handling (Error Handling)

Description: In this module, you will learn the basics such as assigning a variable, differences between dictionary, sets, tuple, and some decision making statements. Also, you will learn about working with different loops, data types and its usage.


  • Data Types
  • Conditional Statements
  • Functions
  • Loops

Description: This module helps you to learn and understand the different libraries used in Python. You will also get a clear idea about the NumPy library and how you can use it. NumPy is a Numeric Python library which helps in dealing with the numeric calculations with data frames.


  • NumPy Introduction
  • Arrays
  • Array Indexing
  • NumPy Data Types
  • Treating Missing and NA’s
  • Reshaping and combining Arrays

Description: In this module, you will learn how to download the Pandas package and syntax for the same. Pandas is also a library similar to Numpy which predominantly helps in working with series data and data frames. You will learn how to impute the data in the place of missing values called the missing value treatment done in the Pandas package itself.


  • Pandas Introduction
  • Basic Operations on Series
  • Dataframe
  • Working with Text Data
  • Working with Missing Data
  • Indexing and Selecting Data
  • Merge, Join and Concatenate

Description: In this module, you will learn where, how and when to use the Matplotlib library. This library is used to visualize the data. You will get an in-depth understanding of the importance of this library.


  • Introduction to Matplotlib
  • Matplotlib design and different visualizations

Description: This module will help you to understand the importance of Seaborn package and downloading the library Just like the Matplotlib library, the Seaborn library is also used in visualizing the data allowing high-level visualizations with categorical data.


  • Introduction to Seaborn Library
  • Visualizing the Distribution of the Datasets
  • Plotting the Categorical Data
  • Visualizing Linear Relationships
  • Visualizing Statistical Relationships

Description: In this module, you will understand the importance of both Scipy and Sklearn libraries which are predominantly used in building Machine Learning Algorithms working with Linear Equations. Sklearn also known as Scikit-learn, is a machine learning library for the Python programming language. You will get a clear idea of where you can use these libraries along with some examples.


  • Installing both SciPy and Sklearn Libraries
  • Introduction to SciPy (Mathematical Algorithms)
  • Introduction to Sklearn (Machine Learning Algorithms)

Description: Learn about High-level overview of Data Science project management methodology, Statistical Analysis using examples, understand Statistics and Statistics 101. Also, learn about exploratory data analysis, data cleansing, data preparation, feature engineering.


  • High-Level overview of Data Science / Machine Learning project management methodology
  • Videos for Data Collection - Surveys  and Design of Experiments will be provided
  • The various Data Types namely continuous, discrete, categorical, count, qualitative, quantitative and its identification and application. Further classification of data in terms of Nominal, Ordinal, Interval and Ratio types
  • Random Variable and its definition
  • Probability and Probability Distribution – Continuous probability distribution / Probability density function and Discrete probability distribution / Probability mass function

Description: Continue with the discussion on understanding Statistics, the various Moments of business decision and other Basic Statistics Concepts. Also, learn about some graphical techniques in Analytics.


  • Balanced vs Imbalanced datasets
  • Various sampling techniques for handling balanced vs imbalanced datasets
  • Videos for handling imbalanced data will be provided
  • What is Sampling Funnel, its application and its components
    • Population
    • Sampling frame
    • Simple random sampling
    • Sample
  • Measure of central tendency
    • Mean / Average
    • Median
    • Mode
  • Measure of Dispersion
    • Variance
    • Standard Deviation
    • Range
  • Expected value of probability distribution

Description: Learn about the other moments of business decision as part of Statistical Analysis. Learn more about Visual data representation and graphical techniques. Learn about Python, R programming with respect to Data Science and Machine Learning. Understand how to work with different Python IDE and Python programming examples.


  • Measure of Skewness
  • Measure of Kurtosis
  • Various graphical techniques to understand data
    • Bar plot
    • Histogram
    • Box plot
    • Scatter plot
  • Introduction to R and RStudio  
  • Installation of Python IDE
  • Anaconda and Spyder
  • Working with Python and R with some basic commands

Description: Learn about Normal Distribution and Standard Normal Distribution. Rules and Principles of Normal distribution. And how to check for normality by QQ normal distribution Plot.


  • Normal Distribution
  • Standard Normal Distribution / Z distribution
  • Z scores and Z table
  • QQ Plot / Quantile-Quantile plot

Description: Under this last topic on Basics of statistics, learn some higher statistical concepts and gain understanding on interval estimates.


  • Sampling Variation
  • Central Limit Theorem
  • Sample size calculator
  • T-distribution / Student's-t distribution
  • Confidence interval
    • Population parameter - Standard deviation known
    • Population parameter - Standard deviation unknown

Description: Get introduced to Hypothesis testing, various Hypothesis testing Statistics, understand what is Null Hypothesis, Alternative hypothesis and types of hypothesis testing.


  • Parametric vs Non-parametric tests
  • Formulating a Hypothesis
  • Choosing Null and Alternative hypothesis
  • Type I and Type II errors
  • Comparative study of sample proportions using Hypothesis testing
  • 2 sample t test

Description: Learn about the various types of tests in Hypothesis testing. Get introduced to the prerequisites and conditions needed to proceed with a Hypothesis test. Understand the interpretation of the results of a Hypothesis testing and probabilities of Alpha error.


  • 1 sample t test
  • 1 sample z test
  • 2 Proportion test
  • Chi-Square test
  • Non-Parametric test

Description: Continuing the discussion on Hypothesis testing, learn more about non-parametric tests. Perform tests using R and interpret the results.


  • Non-Parametric test continued
  • Hypothesis testing using Python and R

Description: Learn about Linear Regression, components of Linear Regression viz regression line, Linear Regression calculator, Linear Regression equation. Get introduced to Linear Regression analysis, Multiple Linear Regression and Linear Regression examples.


  • Scatter diagram
    • Correlation Analysis
    • Correlation coefficient
  • Ordinary least squares
  • Principles of regression
  • Splitting the data into training, validation and testing datasets
  • Understanding Overfitting (Variance) vs Underfitting (Bias)
  • Generalization error and Regularization techniques
  • Introduction to Simple Linear Regression
  • Heteroscedasticity / Equal Variance

Description: In the second part of the tutorial, you will learn about the Models and Assumptions for building Linear Regression Models, build Multiple Linear Regression Models and evaluate the results of the Linear Regression Analysis.     


  • LINE assumption
    • Collinearity (Variance Inflation Factor)
    • Linearity
    • Normality
  • Multiple Linear Regression
  • Model Quality metrics
  • Deletion diagnostics

Description: Learn to analyse Attribute Data, understand the principles of Logistic Regression, Logit Model. Learn about Regression Statistics and Logistic Regression Analysis.


  • Principles of Logistic Regression
  • Types of Logistic Regression
  • Assumption and Steps in Logistic Regression
  • Analysis of Simple Logistic Regression result

Description: Learn about the Multiple Logistic Regression and understand the Regression Analysis, Probability measures and its interpretation. Know what is a confusion matrix and its elements. Get introduced to “Cut off value” estimation using ROC curve. Work with gain chart and lift chart.     


  • Multiple Logistic Regression
  • Confusion matrix
    • False Positive, False Negative
    • True Positive, True Negative
    • Sensitivity, Recall, Specificity, F1
  • Receiver operating characteristics curve (ROC curve)
  • Lift charts and Gain charts 

Description: Learn about the Discrete probability distribution. Types of Discrete probability distribution viz Binomial distribution, Poisson distribution and working with the probability distribution formula.


  • Binomial Distribution
  • Negative Binomial Distribution
  • Poisson Distribution

Description: Get introduced to various advanced regression techniques, especially regression analysis of count data namely Poisson Regression, Negative binomial regression. Learn when to use Poisson regression and Negative binomial regression for predicting count data.


  • Poisson Regression
  • Poisson Regression with Offset
  • Negative Binomial regression
  • Treatment of data with excessive zeros
    • Zero-inflated Poisson
    • Zero-inflated Negative Binomial
    • Hurdle model

Description: Get introduced to Multinomial regression, or otherwise known as multinomial logistic regression, learn about multinomial logit models and multinomial logistic regression examples.


  • Logit and Log Likelihood
  • Category Baselining
  • Modeling Nominal categorical data
  • Additional videos are provided on Lasso / Ridge regression for identifying the most significant variables

Description: As part of Data Mining Unsupervised get introduced to various clustering algorithms, learn about Hierarchal clustering, K means clustering using clustering examples and know what clustering machine learning is all about.



  • Supervised vs Unsupervised learning
  • Data Mining Process
  • Measure of distance
    • Numeric - Euclidean, Manhattan, Mahalanobis
    • Categorical - Binary Euclidean, Simple Matching Coefficient, Jaquard’s Coefficient
    • Mixed - Gower’s General Dissimilarity Coefficient
  • Types of Linkages
    • Single Linkage / Nearest Neighbour
    • Complete Linkage / Farthest Neighbour
    • Average Linkage
    • Centroid Linkage
  • Hierarchical Clustering / Agglomerative Clustering

Description: In this continuation lecture learn about K means Clustering, Clustering ratio and various clustering metrics. Get introduced to methods of making optimum clusters.


  • Non-clustering
    • K-Means Clustering
    • Measurement metrics of clustering - Within Sum of Squares, Between Sum of Squares, Total Sum of Squares
    • Choosing the ideal K value using Scree plot / Elbow Curve
  • Additional videos are provided to understand K-Medians, K-Medoids, K-Modes, Clustering Large Applications (CLARA), Partitioning Around Medoids (PAM), Density-Based Spatial Clustering of Applications with Noise (DBSCAN) and Ordering Points To Identify the Clustering Structure (OPTICS)

Description: Learn to apply data reduction in data mining using dimensionality reduction techniques. Gain knowledge about the advantages of dimensionality reduction using PCA and SVD.


  • Why dimension reduction
  • Advantages of PCA
  • Calculation of PCA weights
  • 2D Visualization using Principal components
  • Basics of Matrix algebra
  • SVD – Decomposition of matrix data

Description: Under data mining unsupervised techniques, learn about Network Analytics and the measures used. Get introduced to Network Analysis tools like NodeXL.


  • Definition of a network (the LinkedIn analogy)
  • Measure of Node strength in a Network
    • Degree centrality
    • Closeness centrality
    • Eigenvector centrality
    • Adjacency matrix
    • Betweenness centrality
    • Cluster coefficient
  • Introduction to Google Page Ranking

Description: Learn one of the most important topic Association rules in data mining. Understand how the Apriori algorithm works, and the association rule mining algorithm.


  • What is Market Basket / Affinity Analysis
  • Measure of association
    • Support
    • Confidence
    • Lift Ratio
  • Apriori Algorithm
  • Sequential Pattern Mining

Description: Learn how online recommendations are made. Get insights about online Recommender System, Content-Based Recommender Systems, Content-Based Filtering and various recommendation engine algorithms. Get to know about people to people collaborative filtering and Item to item collaborative filtering.


  • User-based collaborative filtering
  • Measure of distance / similarity between users
  • Driver for recommendation
  • Computation reduction techniques
  • Search based methods / Item to item collaborative filtering
  • SVD in recommendation
  • Vulnerability of recommender systems

Description: Learn about Machine Learning modeling using KNN, the K nearest neighbour algorithm using KNN algorithm examples. The KNN classifier is one of the most popular classifier algorithms.


  • Deciding the K value
  • Building a KNN model by splitting the data
  • Understanding the various generalization and regulation techniques to avoid overfitting and underfitting

Description: The aim of this course is to understand what Data Visualization is all about. You will understand what are the best practices of Data Visualization, creating data visualization charts and understanding which visualization tools can be considered. Further, you will look into why you need to consider Tableau. You will also get an understanding of products in Tableau You will get an understanding as to what Data Visualization Principles are. Our course content is designed as per Tableau Certification. Edward Tufte considered as the father of Data Visualization came up with Data Integrity rules that need to be followed to get beautiful pieces of evidence.


  • Why visualization came into Picture
  • Importance of Visualizing Data
  • How Data is getting generated
  • Poor Visualizations Vs.Perfect Visualizations
  • Principles of Visualizations
  • Examples of Perfect Visualization
  • Tufte’s Graphical Integrity Rule
  • Tufte’s Principles for Analytical Design Visual

Description: You will get to know about Tableau Desktop, Tableau Server, Tableau Online, Tableau Prep and get an understanding on 14-day trial option i.e. the free version of Tableau - Tableau Public and how Tableau Public login works. You will also look into Tableau Public vs Tableau Desktop and glance on Tableau free download for Students. Get an understanding of Tableau Interactive Dashboards. You will also see what Tableau Reader is all about. You will see different types of Data. You will see what Tableau Architecture is and how it works. What Tableau Data source page is and how to customize the data is learnt here. You will understand what Discrete data and Continuous data are and their differences. You will also see how Data Interpretation works and what exactly happens to the data after interpretation. You will also get to understand the user interface on Tableau.


  • Products of Tableau
  • Tableau Public in detail
  • About Viz of the day, Viz of the week
  • Start Page on Tableau Desktop Professional
  • Tableau Architecture
  • Connecting to Data Source
  • Understanding on Data Source Page
  • Pivot Tables
  • Difference between Discrete data and Continuous data
  • Data Interpretation
  • Tableau User Interface

Description:  You will get a know around of the Charts in Tableau from the Show me Panel that is available in the Tableau work area. Text Table, hands-on understanding is given. Get an understanding of how Heat map for websites works. Check the Highlight tables in Tableau. You will see how Pie Charts are created. Also, see some Pie chart example like which do we use these pie chart. You will also see how the Bar charts are created. You will also see what are the other bar charts. Eg bar chart stacked, Side by side bar chart.


  • Text Tables, Totals
  • Highlight Tables
  • Heat Maps
  • Copy and Exporting the Data
  • Pie Chart
  • Bar Chart
  • Arbitrary Formatting of Colors
  • Conditional Formatting
  • Stacked Bars
  • Side by Sidebars
  • Tree Chart
  • Circle Chart
  • Side by side Circle chart

Description: If say the data is present in different sheets or different Data sources, then the need to learn how joins, Unions, Cross database joins tableau and data blending in tableau help us in connecting them.


  • Joins
  • Cross-Database Connections
  • Unions
  • Data Blending

Description: Learn how to create Filters in Tableau. You will find out the types of filters in Tableau and understand the hierarchy of filters, other filters like quick filters and context filters in Tableau. Get an exposure to how Extracts, Extracts Filters and Live data works in Tableau.


  • Extract Filters
  • Extract and Live Connections
  • Data Source Filters
  • Dimension Filters
  • Measures Filters
  • Date Filters
  • Various Options on Filters

Description: Understand what are Sets and Groups in Tableau. You will also get an understanding of sets vs groups in Tableau. Understand about folders and Tableau hierarchy. Get hands-on exposure on creating Folders, Groups, Sets Hierarchy in Tableau. You will be able to understand the parameters in Tableau that makes the visualization dynamic. Hands-on exposure to how parameters in tableau help come up with dynamic or interactive Dashboards.


  • Hierarchy
  • Folders
  • Grouping
  • Sets
  • Parameters

Description: Time series charts can also be created. Listed below are some of the Intermediate level charts that can be analysed. This helps you to work with multiple dimensions and multiple measures on the view area.


  • Time Series Charts (Line Chart)
  • Area Chart
  • Dual Line Chart
  • Dual Combination
  • Combination Chart

Description: You will understand about maps in Tableau. Also, see what is tableau map layers are and how to see latitude and longitude on google maps and customize geocoding. You will also understand what symbol maps and filled maps are in Tableau.


  • Symbol Maps
  • Filled Maps
  • Background images
  • Polygon Maps
  • Connecting to WMS Server
  • Lasso, Radial and Rectangular selection

Description: You will learn about the box and whisker plot. We generally refer this as box plots. You will learn about scatter plots and then see what is trend analysis and different models. You will understand what is predictive analysis and see how you can use predictive analysis in forecasting in Tableau. Learn how to create Histogram. Get an understanding of Funnel chart. We will learn about donut chart and how to create donut chart in Tableau. You will also learn about the waterfall chart. The other name of the waterfall model is the Gantt Chart. You will get a look into how Pareto chart is created, get an understanding of what Pareto analysis is before you get into the working. Understand the concepts of bullet charts.


  • Scatter Plot
  • Clustering
  • Trendlines
  • Box Plot
  • Histogram
  • Bullet Chart
  • Forecasting
  • Packed Bubble
  • Funnel Chart
  • Donut Chart
  • Waterfall Chart / Gantt Chart
  • Pareto Chart

Description: Look at how the calculations can be done using “Create Calculated Fields” option. Understand of how various Logical, String, Numerical, Ad-hoc Calculations and Quick Table Calculations can be done here. You will work on LoD in Tableau. Which means Level of Detail. This helps in build little more advanced calculations.


  • Logical Calculations
  • String Calculations
  • Numerical Calculations
  • Quick Table Calculations
  • Ad-hoc Calculation
  • LOD Expressions

Description: You will learn about Actions in Tableau and different actions like Filters in Tableau Dashboards. You will get an understanding of the Tableau server. You will also learn the concepts done on R tool implement in Tableau and see how the integration between these tools take place.


  • Integration between R and Tableau
  • Integration between Hadoop and Tableau
  • Dashboards and Actions
  • Story
  • Connecting Data to Tableau Server

Description: The key to any successful project accomplishment including analytics consulting projects would be to understand the business problem. Also, you should understand the initial activities to be performed in Data Science projects for solving business problems using Data Analytics.


  • Business Objectives
  • Business Constraints
  • Creating a Business Case
  • Components of Business Case
  • Creating Project Charter
  • Components of Business Case

Description: Understanding the various forms of collecting data and collecting the right data is of paramount importance for developing interesting insights in solving analytics problems. Deciding on the various market research techniques and ways of collecting data is pivotal to the success of Data Science projects.


  • Market Research using Secondary Data Sciences
  • Data Collection from Primary Data Sources
  • Performing Surveys and Questionnaire
  • Performing Experiments
  • Validating Data Quality

Description: Gathering the data alone is not sufficient, Data Scientists need to ensure that it is in a clean format. Exploring the data while performing data cleansing consumes a significant amount of time and allocating the right amount of effort towards these activities is very important.


  • Data cleansing including Outlier Analysis, Imputation, etc.
  • EDA to bring interesting Descriptive Analytics for actioning
  • Feature Engineering to get Derived Variables
  • Applying Domain Knowledge
  • Getting the final data for Predictive Modeling

Description: Determine whether Data Mining supervised learning or unsupervised learning is applicable for solving the business problem or do you need to implement a combination of both to solve the problem. Understanding what process has to be followed from selecting the right variables and algorithms required for solving a problem is learnt in this module.


  • Decide statistically on what are the most important variables
  • How to decide on which is the right technique / algorithm
  • Deciding on how to deal with balanced / imbalanced dataset
  • Deciding on highest accuracy model and high-performance model

Description: Learn how to close a Data Science project or Artificial Intelligence project and determine whether the purpose of the project success criteria is met or not. Deciding on how to deploy the solution at the client side is very important because all the hard work will be meaningless if customers do not get an easy way of viewing the solution and results.


  • Decide on the model deployment strategy - Web / Mobile / Etc.
  • How to gauge the project closure criteria
  • Performing Review and Retrospection
  • Deciding upon model maintenance and upgradation strategy

Description: Learn about how data is playing a key role in an organization. Data is the new oil that is the driving force for all industry, sectors and domains. With big data in the current world, organizations need to take leverage from Data to gain a competitive edge in real-time. Understand the need for Big Data tools, various components of Big Data, the architecture and the Big Data tools for processing.


  • Introduction to Big Data
  • Data, Data, Data Everywhere
  • 3 V’s of Big Data (Volume, Variety and Velocity)
  • Challenges with Big data
  • Need and significance of innovative technologies
  • What is Hadoop
  • History of Hadoop and its uses
  • Different components of Hadoop
  • Various Hadoop Distributions

Description: Learn about the three main components of Big Data Hadoop. Understand the Master / Slave architecture of Hadoop. Learn about the Demons of Storage component – HDFS and Processing component – MapReduce and finally learn about the resource manager which manages all the operations in the Hadoop Cluster.


  • Significance of HDFS in Hadoop
  • HDFS Features
  • Daemons of Hadoop and functionalities
  • Data Storage in HDFS
  • Accessing HDFS
  • Data Flow
  • HDFS commands hands-on
  • Introduction to MapReduce
  • MapReduce Architecture
  • Data Types
  • Input Splits and Records
  • Basic MapReduce Program
  • The MapReduce Web UI

Description: Learn about the first multi-user operating system - Linux and file system of Linux OS, Kernel, Interactive Shells, etc. Understand the usage of the Terminal and its commands. Learn about virtualization softwares like VMware and VirtualBox. Creation of a virtual Linux machine for Pseudo Hadoop Cluster setup


  • Virtualization
  • VMware Workstation
  • VirtualBox
  • Setup of Linux Virtual Machine
  • What is Linux OS
  • Flavours of Linux Os
  • Linux File System
  • Advantages of Linux Os
  • Hands-on Linux Terminal Commands

Description: Introduction to SQL like programming language on Big Data Hadoop over MapReduce.  Components of the Hive execution engine and the flow of the execution. Learn how different Data Warehousing tool - Apache Hive is with respect to SQL language.


  • Hive Engine and its Components
  • RDBMS Hive Metastore
  • Comparison with Traditional Databases
  • HiveQL
  • Hive Tables
  • Querying Data
  • User-Defined Functions

Description: Introduction to traditional Database system – RDBMS and its SQL programming language. Learn about NoSQL database (HBase) and its advantages. Learn how to move data from traditional Database to Big Data Hadoop system and vice versa using Apache Sqoop.


  • Introduction to MySQL
  • Basics of traditional RDBMS concepts
  • Difference between SQL and NoSQL (HBase)
  • Introduction to Sqoop
  • Benefits of Sqoop
  • Sqoop Architecture and Internals
  • MySQL client and server installation
  • How to connect to the relational database using Sqoop
  • Sqoop Commands

Description: Introduction to super-fast, memory based, cluster computing framework - Apache Spark. Components of the Unified Stack Apache Spark. Learn how Spark attains super speed over the Big Data residing in HDFS. Comparison between distributed frameworks - Hadoop and Spark. Learn what is RDD and its creation. Difference between Dataframe, Datasets and RDD in Apache Spark 2.X and their applications. Start writing Spark functions using multiple programming languages.


  • Introduction to Apache Spark
  • Apache Spark vs Hadoop
  • Spark Architecture
  • Spark Execution Environment - SparkContext, SQLContext, SparkSession
  • RDD and Operations on RDD’s
  • Spark Unified Stack
  • Spark Core
  • Spark SQL
  • Spark MLlib
  • Spark Streaming
  • PySpark (Spark using Python)
    • Introduction
    • SparkContext
    • RDDs
    • Broadcast & Accumulator
    • SparkConf
    • SparkFiles
    • StorageLevel
    • MLlib
    • Serializers


  • Understand the core Azure architectural components
    • describe Regions
    • describe Availability Zones
    • describe Resource Groups
    • describe Azure Resource manager
    • describe the benefits and usage of core Azure architectural components
  • Core products available in Azure
    • products available for Compute such as Virtual Machines, Virtual Machine Scale Sets, App Service and Functions
    • products available for Storage such as Blob Storage, Disk Storage, File Storage, and Archive Storage
    • products available for Databases such as CosmosDB, Azure SQL Database, Azure Database Migration service, and Azure SQL Data Warehouse
  • solutions available on Azure
    • Big Data and Analytics and products that are available for Big Data and Analytics such as SQL Data Warehouse, HDInsight and Data Lake Analytics
    • Artificial Intelligence (AI) and products that are available for AI such as Azure Machine Learning Service and Studio
  • Azure management tools
    • Azure CLI, PowerShell, and the Azure Portal
  • Overview of Azure Machine Learning studio

Description: Under the Naive Bayes classifier tutorial, learn how the classification modeling is done using Bayesian classification, understand the same using Naive Bayes example. Learn about Naive Bayes through the example of text mining.


  • Probability – Recap    
  • Bayes Rule
  • Naive Bayes Classifier
  • Text Classification using Naive Bayes

Description: Bagging and Boosting is an ensemble technique which is a part of the random forest algorithm. Learn about Bagging and Boosting examples under this tutorial.


  • Boosting / Bootstrap Aggregating
  • AdaBoost / Adaptive Boosting
  • Stacking
  • Gradient Boosting
  • Extreme Gradient Boosting (XGB)

Description: Decision Tree and Random Forest are one of the most powerful classifier algorithms today. Under this tutorial, learn about Decision Tree Analysis, Decision Tree examples and Random Forest algorithms.


  • Elements of Classification Tree - Root node, Child Node, Leaf Node, etc.
  • Greedy algorithm
  • Measure of Entropy
  • Attribute selection using Information Gain
  • Ensemble techniques
  • Decision Tree C5.0 and understanding various arguments
  • Random Forest and understanding various arguments

Description: Artificial Neural Network and Support Vector Machines are the two powerful Deep learning algorithms. Get introduced to Neural Net, Convolutional Neural Network, Recurrent Neural Network. Learn how to work with Support Vector Machine, SVM classifiers and SVM regression.


  • Artificial Neural Network
  • Biological Neuron vs Artificial Neuron
  • ANN structure
  • Activation function
  • Network Topology
  • Support Vector Machines
  • Classification Hyperplanes
  • Best fit “boundary”
  • Kernel Trick

Description: Text mining or Text data mining is one of the wide spectrum of tools for analyzing unstructured data. As a part of this course, learn about Text analytics, the various text mining techniques, its application, text mining algorithms and sentiment analysis.


  • Sources of data
  • Bag of words
  • Pre-processing, corpus Document-Term Matrix (DTM) and TDM
  • Word Clouds
  • Corpus level word clouds
    • Sentiment Analysis
    • Positive Word clouds
    • Negative word clouds
    • Unigram, Bigram, Trigram
  • Semantic network
  • Clustering

Description: Learn how to extract data from Social Media, download user reviews from E-commerce and Travel websites. Generate various visualizations using the downloaded data.     


  • Extract Tweets from Twitter
  • Extract user reviews of the products from Amazon, Snapdeal and TripAdvisor

Description: Learn how to perform text analytics using Python and work with various libraries that aid in data extraction, text mining, sentiment analysis and  


  • Install Libraries from Shell
  • Extraction and text analytics in Python

Description: Natural language processing applications are in great demand now and various natural language processing projects are being taken up. As part of this tutorial, learn about Natural language and ‘Natural language understanding’.


  • LDA
  • Topic Modeling 
  • Sentiment Extraction
  • Lexicons and Emotion Mining

Description: Forecasting or Time Series Analysis is an important component in analytics. Here, get to know the various forecasting methods, forecasting techniques and business forecasting techniques. Get introduced to the time series components and the various time series analysis using time series examples.


  • Introduction to time series data
  • Steps of forecasting
  • Components of time series data
  • Scatter plot and Time Plot
  • Lag Plot
  • ACF - Auto-Correlation Function / Correlogram
  • Visualization principles
  • Naive forecast methods
  • Errors in forecast and its metrics
  • Model Based approaches
    • Linear Model
    • Exponential Model
    • Quadratic Model
    • Additive Seasonality
    • Multiplicative Seasonality
  • Model-Based approaches
  • AR (Auto-Regressive) model for errors
  • Random walk
  • ARMA (Auto-Regressive Moving Average), Order p and q
  • ARIMA (Auto-Regressive Integrated Moving Average), Order p, d and q
  • Data-driven approach to forecasting
  • Smoothing techniques
    • Moving Average
    • Exponential Smoothing
    • Holts / Double Exponential Smoothing
    • Winters / HoltWinters
  • De-seasoning and de-trending
  • Econometric Models
  • Forecasting Best Practices
  • Forecasting using Python
  • Forecasting using R


  • SLR
  • MLR
  • SVM
  • PCA
  • KNN
    • Amazon Review Extraction
    • TripAdvisor
    • IMDB Review Extraction
    • Snapdeal Review Extraction

Project 1: How To Identify Fraudulent And Ilegal Transactions Due To Insider Trading
Project Related To:
Finance Service Insurance
Problem Description: In spite of the mature regulatory norms, the act of insider trading is on the rise. More robust the regulatory norms become, more intelligent the insider traders become. This is forcing the firms to always be on toes and keep developing better ways of identifying the fraud. Insider trading gives away the secrets of the organizations, which are strictly not to be disclosed outside the boardroom. The ugly politics of companies, which cannot get head-on with the ethical businesses are heavily resorting to these ways of mending the rules to make this way to success. How do you identify the sheep in wolves’ clothing?


Project 2: Learn On How To Predict The Deposits Churn And Reduce The Risk Of Losing Customers
Project Related To:
Finance Service
Problem Description: Considerably, alongside growing the customer base, not maintaining sufficient funds as deposit amounts could lead to levying penalty and this could, in turn, lead to customer churn.
a) How to devise strategies in retaining customers and also ensuring that they maintain required funds in deposits or increase the funds in deposits
b) How to predict on who is the most probable customer to churn
c) How to find out about customers who will continue to stay despite levying penalty for maintaining an amount below the par in the deposit accounts
d) How to segment customers and devise business strategies for each of these segments
e) These are the challenges for which banks need an immediate solution.


Project 3: Want To Know On How Sentiment Analysis Is Performed From Twitter’s Unstructured Data
Project Related To: Social Media Analytics
Project Description: With the increase in digitization, the amount of accessibility to social media for a common person has increased manifold. The Advent of technology not only comes with the advantages but also the disadvantages. Many people who have access to the internet do not restrain from giving to-the-point feedbacks and are not at all shying away. But sometimes, these reviews and feedbacks are given only because of the unhealthy competition.

At times, this is creating a lot of trouble to the genuine products and manufacturers, risking them to drop the plans of manufacturing those products. It also results in dropping of rating of those products.


Project 4: How To Increase The Probability Of ‘Click-Through Rate’ Of Ads Posted On Social Media
Project Related To: Social Media Analytics
Problem Description: The world is now experiencing the highest internet penetration ever. Companies without proper online presence hardly survive. In this context, increasing online visibility, especially when netizens perform a search on search engines is at its prime. There is fierce competition among companies to feature on the first page and being on the top of the search results; this is because people hardly ever move to pages beyond the first page to explore the results. Both top-line and bottom-line of companies are now greatly dependent on Social Media Presence.
a) How prominently your website appears in search results.
b) What should be done to be on the first page


Project 5: Analytics On Political Party Representatives
Project Related To: Social Media Analytics
Project Description: Citizens are resorting to posting messages on social media and the web to vent out the frustration or happiness associated with the daily activities going around. There is no transparency on how many promises were done by political party members at the time of the election. Lack of clarity on the performance of the elected representatives leading to some sections misguiding the society with false claims.


Project Related To: Retail Organizations
Project Description: There is an ever-increasing focus on effective recruitment. An organization invests a lot of its time and resources in search of the potential candidates. The investment become loses if the selected candidates do not join the organization in the end.


Project 7: Warranty Claims
Project Related To: Retail sector
Project Description: Analysis to predict an item when sold, what is the probability that a customer would file for warranty and to understand important factors associated with them.


Project 8: Performance Prediction For Teachers and Students
Project Related To: Retail Sector
Project Description: Educational Data Mining (EDM) aims at knowledge discovery by applying mining techniques to identify hidden knowledge and patterns about students and teachers performance. The idea is to help improve performance by taking appropriate action based on the prediction. Early prediction helps in devising appropriate solutions to draw better results for both students and teachers.


Project 9: Students School Dropouts
Project Related To: Retail Sectors
Project Description: Educational Data Mining (EDM) aims at knowledge discovery by applying mining techniques to identify hidden knowledge and patterns about students dropouts from primary schools. The idea is to help improve the overall quality of primary education by taking appropriate action based on the prediction in school dropouts. Early prediction helps in devising appropriate solutions to help schools address students dropout.


Project 10: Chat Bot
Project Related To: Retail Sector
Project Description: Digitization is penetrating into the remote parts of even the third world in recent times. With the advent of advanced technology and digitization, the data that is being generated is very huge and the number of hands asking for queries on customer product and services is increasing at a rapid pace. Keeping the current and future demand in mind, it will and is becoming a challenging task for the clients to satisfy their customers in responding to their queries.


Project 11: How To Bring Data From Varied Sources To Generate Reports For Businesses To Draw Insights To Devise Strategies
Project Related To: Business Intelligence and Reporting
Project Description:

  • Analytical capability of the reporting tool (Tableau) helps in drawing significant insights to make swift decisions
  • Aesthetic visual pop coupled with analytics feature helps in knowing the potential of the data
  • The extremely easy-to-integrate feature of reporting and analytical tool helps in collaborating data from varied sources, giving scope for robust visualization.
  • The development of visually attractive reports of dashboard combines many sheets in single place giving room for faster analysis


Project 12: How To Generate A Single Report Personalized To Various Departments Using View Security Settings On Server End
Project Related To: Business Intelligence and Reporting
Project Description:

  • Increased efficiency among departments
  • Reduced Data leakages resulted in huge cost savings
  • Parallel reports enhanced the resolution capability at a low time
  • Actionable Insights are derived at a faster rate, resulting in profit generation


Project 13: How To Connect Big Data Source Engines To Tableau And Establish Dashboard Reporting Through Streaming Data
Project Related To: Business Intelligence and Reporting
Project Description: Want to know how to connect big data source engines to Tableau and establish dashboard reporting through streaming data.

Contact Our Team of Experts

Why ExcelR?

Participants Placed Through ExcelR



  • We here at ExcelR, as one of the most leading data science institutes, have come up with an idea where a student can take part in as many as classes and lectures from different trainers that they want to in a span of a year.
  • Upon the completion of the data science training, one will be counted as an important part of the decision-making team. Some of the positions that one can target are:
    • Data analysts
    • Big data specialists
    • Data scientists
    • Business analysts
    • Market analysts
    • Research analysts
    • Business analyst consultants
  • Anybody can receive the data science certification or data science online certification but some of the desired skills are:
    • Analytical skills
    • Problem-solving mindset
    • Mathematical affinity
    • Business knowledge
    • Communication skills
  • The expected salary brackets after the data science training depends on experience and skills-building. What one can expect to earn is:
    • Base level/fresher: up to 5 lakhs
    • Intermediate stage: up to 8 lakhs
    • Experience data scientists: up to 12.5 lakhs
    • Advanced stage: up to 18 lakhs
    • Data science expert: up to 25 lakhs
  • Here at ExcelR’s data science certification and data science online certification course, one can be a part of the most fluid and easy build sessions. One can attend the classroom lectures, they can also be a part of the live online sessions and later can watch the recorded sessions for a better understanding of the data science courses.
  • In the live session of the data science online certification, one will have to log in at the recommended time and the instructor will also do the same to give a live lecture for the data science course. One can also interact and ask questions during the live session of the data science course.
  • One can attend the subsequent batches of their choice or can watch the recorded lecture of the live Data Science course sessions by our subject expert from the learning management system (LMS) of our website for data science courses.
  • Every participant will be part of the whatsapp group along with the trainers. You would be just whatsapp away to clarify your doubts. One can clear their doubts about data science training classes and lectures through the website by chatting with the customer executives, or by sending an email to [email protected] or also by calling the toll free number 1-800-212-2120 and 800 800 9706 (India), 203-514-6638 (United Kingdom), 608-218-3798 (USA), 128-520-3240 (Australia) to contact our best data science institutes.
  • Some of the payment modes available for the data science training module are:
    • cash
    • credit/debit card
    • internet banking
    • checks
    • wallets, etc.
  • Yes, one will be provided with a data science course certificate from ExcelR at the end of the data science training.

Global Presence

ExcelR is a training and consulting firm with its global headquarters in Houston, Texas, USA. Alongside to catering to the tailored needs of students, professionals, corporates and educational institutions across multiple locations, ExcelR opened its offices in multiple strategic locations such as Australia, Malaysia for the ASEAN market, Canada, UK, Romania taking into account the Eastern Europe and South Africa. In addition to these offices, ExcelR believes in building and nurturing future entrepreneurs through its Franchise verticals and hence has awarded in excess of 30 franchises across the globe. This ensures that our quality education and related services reach out to all corners of the world. Furthermore, this resonates with our global strategy of catering to the needs of bridging the gap between the industry and academia globally.

ExcelR's Global Presence


Data Science Certification Training locations in Mumbai : A I staff colony [400029], Aareymilk Colony [400065], Agripada [400011], Airport [400099], Ambewadi [400004], Andheri [400053], Andheri East [400069], Andheri Railway station [400058], Antop Hill [400037], Asvini [400005], Azad Nagar [400053], B P t colony [400003], B.N. bhavan [400051], B.P.lane [400003], Bandra West [400050], Bandra(east) [400051], Bangur Nagar [400090], Bazargate [400001], Best Staff colony [400012], Bharat Nagar [400007], Bhawani Shankar [400028], Borivali [400091], Borivali East [400066], Borvali West [400092], C G s colony [400013], Central Building [400020], Century Mill Chakala Midc [400093], Chamarbaug [400012], Charkop [400067], Chaupati [400004], Chinchbunder [400009], Chinchpokli [400011], Churchgate [400020], Colaba [400005], Cotton Exchange [400033], Cumballa Hill [400026], Dadar [400014], Dahisar [400068], Danda [400052], Daulat Nagar [400066], Delisle Road [400013], Dharavi [400017], Dockyard Road [400010], Dr Deshmukh marg [400026], Falkland Road [400008], Girgaon [400004], Gokhale Road [400028], Goregaon [400062], Goregaon East [400063], Government Colony [400051], Gowalia Tank [400026], Grant Road [400007], H.M.p. school [400058], Haffkin Institute [400012], Haines Road [400011], Hajiali [400034], Hanuman Road [400057], High Court bulding [400032], Holiday Camp [400005], Irla [400056], Ins Hamla [400095], International Airport [400099], J.B. nagar [400059], [400008], Jacob Circle [400011], Jogeshwari East [400060], Jogeshwari West [400102], Juhu [400049], Kalachowki [400033], Kalbadevi [400002], Kamathipura [400008], Kandivali East [400101], Kandivali West [400067], Kapad Bazar [400016], Ketkipada [400068], Khar Colony [400052], Kharodi [400095], Kherwadi [400051], Kidwai Nagar [400031], L B s n e collage [400033], Lal Baug [400012], Liberty Garden [400064], M A marg [400008], M.P.t. [400001], Madh [400061], Madhavbaug [400004], Magthane [400066], Mahim [400016], Malabar Hill [400006], Malad [400064], Malad East [400097], Malad West dely [400064], Mandapeshwar [400103], Mandvi [400003], Mantralaya [400032], Marine Lines [400020], Marol Bazar [400059], Masjid [400003], Matunga Railway workshop [400019], Mazgaon [400010], Mori Road [400016], Motilal Nagar [400104], Mumbai Central [400008], Mumbai[400001], N . s.patkar [400007], Nagardas Road [400069], Nagari Niwara [400065], Naigaon [400014], Nariman Point [400021], New Prabhadevi road [400025], New Yogakshema [400021], Noor Baug [400003], Null Bazar [400003], Opera House [400004], Orlem [400064], Oshiwara [400102], Parel [400012], Parel Rly work shop [400003], Prabhadevi [400025], Princess Dock [400009], Rajbhavan [400035], Rajendra Nagar [400066], Ramwadi [400002], Ranade Road [400028], Rani Sati marg [400097], Reay Road [400033], S R p f camp [400060], S Savarkar marg [400028], S V marg [400007], S. c. court [400002], S. k.nagar [400066], Sahar P & t colony [400099], Santacruz Central [400054], Santacruz P&t colony [400029], Santacruz(east) [400055], Santacruz(west) [400054], Secretariate [400032], Seepz [400096], Sewri [400015], Sharma Estate [400063], Shivaji Park [400028], Shroff Mahajan [400002], Stock Exchange [400001], Tank Road [400033], Tardeo [400007], Thakurdwar [400002], Tulsiwadi [400034], V J b udyan [400027], V K bhavan [400010], V.P. road [400052], V.W.t.c. [400005], Vakola [400055], Vesava [400061], Vidyanagari [400098], Vileeparle (east) [400057], Vileparle(west) [400056], Wadala [400031], Worli [400018], Worli Colony [400030].

Call Us whatsapp