Associate Analytics M1 Faculty Guide
Associate Analytics M1 Faculty Guide
ASSOCIATE ANALYTICS
FACILITATORS GUIDE
MODULE 1
Page 2 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Facilitator’s Guide
Associate - Analytics
Powered by:
Page 3 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Copyright © 2014
NASSCOM
4E-Vandana Building (4th Floor)
11, Tolstoy Marg, Connaught Place
New Delhi 110 001, India
T 91 11 4151 9230; F 91 11 4151 9240
E [email protected]
W www.nasscom.in
Published by
Disclaimer
The information contained herein has been obtained from sources reliable to
NASSCOM. NASSCOM disclaims all warranties as to the accuracy, completeness or
adequacy of such information. NASSCOM shall have no liability for errors, omissions,
or inadequacies, in the information contained herein, or for interpretations thereof.
Every effort has been made to trace the owners of the copyright material included in
the book. The publishers would be grateful for any omissions brought to their notice
for acknowledgements in future editions of the book.
No entity in NASSCOM shall be responsible for any loss whatsoever, sustained by any
person who relies on this material. The material in this publication is copyrighted. No
parts of this report can be reproduced either on paper or electronic media, unless
authorized by NASSCOM.
Page 4 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Foreword
Page 5 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Acknowledgements
Page 6 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Prologue
Page 7 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
CORE CONTENT
Page 8 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Introduction
Qualifications Pack-Associate –Associate Analytics SSC/Q2101
SECTOR: IT-ITeS
OCCUPATION: Analytics
Brief Job Description: Individuals at this job are responsible for building analytical packages
using Databases, Excel or other Business Intelligence (BI) tools
Personal Attributes: This job requires the individual to follow detailed instructions and
procedures with an eye for detail. The individual should be analytical and result oriented and
should demonstrate logical thinking.
Page 9 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 10 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 11 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 12 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Associate - Analytics
Job Details
Job Role This job role is applicable in both national and international
scenarios
Credits(NVEQF/NVQF/NSQF) Version number 0.1
Sector IT-ITeS Drafted on 30/04/13
Business Process 30/04/13
Sub-sector Last reviewed on
Management
30/06/14
Occupation Analytics Next review date
Page 13 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Session Overview
In the Associate Analytics “Working with Documents”, the participant will learn about the most
prominently used documentation techniques in corporate organizations. The Documentation
types covered would include case studies, best practices, project artifacts, reports, minutes,
policies, procedures, work instructions etc.
This session is NOT intended to cover technical documents or documents to support the
deployment and use of products/applications, which are dealt with in different standards.
Session Goal
Participants should be able to have a good hands on understanding of MS Word and MS Visio,
where there will be required to draft various documents/reports. The goal of the session is for the
participant to be aware of the various documentation techniques which are used prominently in
organizations.
Session Objectives
Upon completion of both parts of this course, the participants will be able to:
PC1. establish with appropriate people the purpose, scope, formats and target audience for
the documents
PC2. access existing documents, language standards, templates and documentation tools from
your organization’s knowledge base
PC3. liaise with appropriate people to obtain and verify the information required for the
documents
PC4. confirm the content and structure of the documents with appropriate people
PC5. create documents using standard templates and agreed language standards
PC6. review documents with appropriate people and incorporate their inputs
PC7. submit documents for approval by appropriate people
PC8. publish documents in agreed formats
PC9. update your organization’s knowledge base with the documents
PC10. comply with your organization’s policies, procedures and guidelines when creating
documents for knowledge sharing
Note: The material for this NOS has been covered in the Associate Analytics Module 3 Book
(book 3) in Unit 5
Page 14 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
In the Associate Analytics Carry out rule based statistical analysis, the participants will go
through Business Analytics using R tool. The participants will also learn Applied Statistical
concepts like Descriptive Statistics and find their usage along with R. Furthermore, they willalso
have an overview of Big Data tools and their basic functioning.
Then they will learn about Machine Learning algorithm and their use in Data Mining and
Predictive Analytics. Finally the participants will learn about Data Visualization and gather
knowledge on Graphical representation of Data as well as results and reports.
Session Goal
The primary goal of the session is for the participants to learn the R tools and its various
functions and features. Then also learn about Big Data tools and Big Data Analytics. Students
will also learn about basic applied statistical concepts.
Session Objectives
To be competent, participants must be able to:
Note: The material for this NOS has been covered in all the three Modules of Associate
Analytics
Page 15 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Session Goal
The primary goal of the session is for the participants to learn and manage time to be able to
complete their work as required. The requirements of a work unit may be further classified
into; activities, deliverable, quantity, standards and timelines. The session makes participants
to be aware of defining requirements of every work unit and then ensuring delivery.
Additionally, this session discusses practical application of planning and execution of work
plans to enable the participants to effectively deal with the failure points, minimize the
impact, if any. Equally critical is the escalation plan and root cause analysis of exceptions.
Successful candidates will be able to understand the inter-relationship of time, effort,
impact and cost.
Session Objectives
Upon completion of both parts of this course, the participants will be able to:
PC1. Establish and agree your work requirements with appropriate people
PC2. Keep your immediate work area clean and tidy
PC3. Utilize your time effectively
PC4. Use resources correctly and efficiently
PC5. Treat confidential information correctly
PC6. Work in line with your organization’s policies and procedures
PC7. Work within the limits of your job role
PC8. Obtain guidance from appropriate people, where necessary
PC9. Ensure your work meets the agreed requirements
Note: The material for this NOS has been covered in Unit 1 of Module 1. Much of the material
herein is going to be self-study for the participants
Page 16 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Session Goal
The primary goal of the session is for the participants to understand the importance of
professional relationships with colleagues.Additionally, this session discusses importance of
personal grooming.
Successful candidates will be able to understand the inter-relationship of professionalism and
team-work.
Session Objectives
Upon completion of both parts of this course, the participants will be able to:
PC1. Communicate with colleagues clearly, concisely and accurately.
PC2. Work with colleagues to integrate your work effectively with theirs.
PC3. Pass on essential information to colleagues in line with organizational requirements.
PC4. Work in ways that show respect for colleagues.
PC5. Carry out commitments you have made to colleagues.
PC6. Let colleagues know in good time if you cannot carry out your commitments, explaining
the reasons.
PC7. Identify any problems you have working with colleagues and take the initiative to solve
these problems.
PC8. Follow the organization’s policies and procedures for working with colleagues.
Note: The material for this NOS has been covered in Unit 2 of Module 1. Much of the material
herein is going to be self-study for the participants
Page 17 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Session Goal
The primary goal of the session is for the participants to be aware about the various hazards
that they may come across at workplace and what are the defined health, safety and security
measures that should be followed at the time of occurrence of such unpredictable events.
Additionally, this session discusses practical application of the health and safety procedures to
enable the participants to effectively deal with the hazardous events to minimize the impact, if
any.
Session Objectives
Upon completion of both parts of this course, the participants will be able to:
PC1. Comply with your organization’s current health, safety and security policies and procedures
PC2. Report any identified breaches in health, safety, and security policies and procedures to the
designated person
PC3. Identify and correct any hazards that you can deal with safely, competently and within the
limits of your authority
PC4. Report any hazards that you are not competent to deal with to the relevant person in line
with organizational procedures and warn other people who may be affected
PC5. Follow your organization’s emergency procedures promptly, calmly, and efficiently
PC6. Identify and recommend opportunities for improving health, safety, and security to the
designated person
PC7. Complete any health and safety records legibly and accurately
Note: The material for this NOS has been covered in Unit 2 of Module 1. Much of the material
herein is going to be self-study for the participants
Page 18 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Session Goal
The primary goal of the session is for the participants to analyze data and present it in a
suitable format, as is suitable for the given process or organization.
Successful candidates will be able to understand the process of standardized reporting and the
nuances of a publishing a report with a specified end objective in mind.
Session Objectives
Upon completion of both parts of this course, the participants will be able to:
PC1. establish and agree with appropriate people the data/information you need to provide,
the formats in which you need to provide it, and when you need to provide it
PC2. obtain the data/information from reliable sources
PC3. check that the data/information is accurate, complete and up-to-date
PC4. obtain advice or guidance from appropriate people where there are problems with the
data/information
PC5. carry out rule-based analysis of the data/information, if required
PC6. insert the data/information into the agreed formats
PC7. check the accuracy of your work, involving colleagues where required
PC8. report any unresolved anomalies in the data/information to appropriate people
PC9. provide complete, accurate and up-to-date data/information to the appropriate people in
the required formats on time
Page 19 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Session Goal
The pimary goal of the session is to give a overview on how skills and competency can be
enhanced in a professional environment. It gives knowledge on organizational context, technical
knowledge, core skills/geneic skills, professional skills and technical skills.The session makes
participants to understand the need of skills improvement for personal and organizational
growth.
Successful candidates will be able ro understand the relationship between skill enhancement and
growth.
Session Objectives
Upon completion of both parts of this course, the participants will be able to:
PC1. obtain advice and guidance from appropriate people to develop their knowledge, skills
and competence
PC2. identify accurately the knowledge and skills you need for their job role
PC3. identify accurately their current level of knowledge, skills and competence and any
learning and development needs
PC4. agree with appropriate people a plan of learning and development activities to address
their learning needs
PC5. undertake learning and development activities in line with their plan
PC6. apply their new knowledge and skills in the workplace, under supervision
PC7. obtain feedback from appropriate people on their knowledge and skills and how
effectively they apply them
PC8. review their knowledge, skills and competence regularly and take appropriate action
Page 20 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Module 2 – Book 2
Subject II / SSC NASSCOM - NOS- 2010, 9003, 9004 NOS Hours Minutes
NOS
Unit – 1 2101/9003
Data Management 7 420
Maintain Healthy, Safe & Secure Working environment 4 240
NOS
Unit – 2 2101/9004
Big Data Tools 7 420
Provide Data/Information in Standard formats 4 240
Unit – 3 NOS 2101
Big Data Analytics 8 480
Unit – 4 NOS 2101
Machine Learning Algorithms 8 480
Unit – 5 NOS 2101
Data Visualization 6 360
Product Implementation 6 360
Total Hrs/Minutes 50 3000
Page 21 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Module 3 – Book 3
Subject III / SSC NASSCOM - NOS - 0703, 2101, 9005 NOS Hours Minutes
Unit – 1 NOS 2101
Introduction to Predictive Analytics 6 360
Linear Regression 6 360
Unit – 2 NOS 2101
Logistics Regression 9 540
NOS
Unit – 3 2101/9005
Objective Segmentation 6 360
Develop Knowledge Skill and competences 3 180
Unit – 4 NOS 2101
Time Series Methods/Forecasting, Feature Extraction 5 300
Project 5 300
Unit – 5 NOS 0703
Working with documents 10 600
Total Hrs/Minutes 50 3000
Page 22 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Glossary of Terms
businesses and interests. It may also be defined asa distinct subset of the
economy whose components share similar characteristics and interests.
Sub-sector Sub-sector is derived from a further breakdown based on the characteristics and
interests of its components.
Vertical Vertical may exist within a sub-sector representing different domain areas or the
client industries served by the industry.
Occupation Occupation is a set of job roles, which perform similar/related set of functions in
an industry.
Function Function is an activity necessary for achieving the key purpose of the sector,
occupation, or area of work, which can be carried out by a person or a group of
persons. Functions are identified through functional analysis and form the basis
of OS.
Sub-functions Sub-functions are sub-activities essential to fulfill the achieving the objectives of
the function.
Job role Job role defines a unique set of functions that together form a unique
employment opportunity in an organisation.
Occupational OS specify the standards of performance an individual must achieve when
Standards (OS) carrying out a function in the workplace, together with the knowledge and
understanding they need to meet that standard consistently. Occupational
Standards are applicable both in the Indian and global contexts.
Performance
Performance Criteria are statements that together specify the standard of
Criteria
performance required when carrying out a task.
National
Occupational NOS are Occupational Standards which apply uniquely in the Indian context.
Standards (NOS)
Qualifications Pack Qualifications Pack Code is a unique reference code that identifies a qualifications
Code pack.
Qualifications Qualifications Pack comprises the set of OS, together with the educational,
Pack(QP) training and other criteria required to perform a job role. A Qualifications Pack is
assigned a unique qualification pack code.
Unit Code Unit Code is a unique identifier for an OS unit, which can be denoted with either
an ‘O’ or an ‘N’.
Unit Title Unit Title gives a clear overall statement about what the incumbent should be
able to do.
Page 23 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Description Description gives a short summary of the unit content. This would be helpful to
anyone searching on a database to verify that this is the appropriate OS they are
looking for.
Scope Scope is the set of statements specifying the range of variables that an individual
may have to deal with in carrying out the function which have a critical impact on
the quality of performance required.
Knowledge and Knowledge and Understanding are statements which together specify the
Understanding technical, generic, professional and organisational specific knowledge that an
individual needs in order to perform to the required standard.
Organisational Organisational Context includes the way the organisation is structured and how it
Context operates, including the extent of operative knowledge managers have of their
relevant areas of responsibility.
Technical
Technical Knowledge is thespecificknowledgeneededto accomplish specific
Knowledge
designated responsibilities.
Core Skills/Generic Core Skills or Generic Skills are a group of skills that are key to learning and
Skills working in today's world. These skills are typically needed in any work
environment. In the context of the OS, these include communication related skills
that are applicable to most job roles.
Helpdesk Helpdesk is an entity to which the customers will report their IT problems. IT
Service Helpdesk Attendant is responsible for managing the helpdesk.
Keywords /Terms Description
IT-ITeS Information Technology - Information Technology enabled Services
BPM Business Process Management
BPO Business Process Outsourcing
KPO Knowledge Process Outsourcing
LPO Legal Process Outsourcing
IPO Information Process Outsourcing
BCA Bachelor of Computer Applications
B.Sc. Bachelor of Science
OS Occupational Standard(s)
NOS National Occupational Standard(s)
QP Qualifications Pack
UGC University Grants Commission
MHRD Ministry of Human Resource Development
MoLE Ministry of Labour and Employment
NVEQF National Vocational Education Qualifications Framework
NVQF National Vocational Qualifications Framework
NSQF National Skill Qualification Framework
Page 24 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Occupation (2 numbers)
It is important to note that an OS unit can be denoted with either an ‘O’ or an ‘N’.
If an OS unit denotes ‘O’, it is an OS unit that is an international standard. An example of OS unit
denoting ‘O’ is SSC/O0101.
If an OS unit denotes ‘N’, it is an OS unit that is a national standard and is applicable only for the
Indian IT-ITeS industry. An example of OS unit denoting ‘N’ is SSC/N0101
Page 25 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 26 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Introduction to Analytics or R By the end of this session, you will be able to:
programming 1. Understand R
2. Use functions of R
Session Plan:
Activity Location
Knowing language R Classroom
Summary Classroom
Page 27 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Look at R!
Using R as calculator
R can be used as a calculator. For example, if we have to know what is 2+2 then-
> 2+2
Press enter and we get the answer as
[1] 4
R-Studio Interface
Similarly we can calculate anything as if done on calculator.
Page 28 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
1. Log of 2
2. 23 X 32
3. e3
Understanding components of R
1. Data Type:
There are two types of data classified on very broad level. They are Numeric and Character data.
Numeric Data: - It includes 0~9, “.” and “- ve” sign.
Character Data: - Everything except Numeric data type is Character. For Example, Names, Gender
etc.
For Example, “1,2,3…” are Quantitative Data while “Good”, “Bad” etc. are Quantitative Data.
Although we can convert Qualitative Data into Quantitative Data using Ordinal Values.
For Example, “Good” can be rated as 9 while “Average” can be rated as 5 and “Bad” can be rated as 0.
2. Data Frame:
A data frame is used for storing data tables. It is a list of vectors of equal length.
The top line of the table, called the header, contains the column names. Each horizontal line afterward
denotes a data row, which begins with the name of the row, and then followed by the actual data. Each data
member of a row is called a cell. To retrieve data in a cell, we would enter its row and column coordinates in
the single square bracket "[]" operator. The two coordinates are separated by a comma. In other words, the
coordinates begins with row position, then followed by a comma, and ends with the column position. The
order is important.
For Example,
Here is the cell value from the first row, second column of mtcars.
Page 29 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
We have two different options for constructing matrices or arrays. Either we use the creator
functions matrix () and
Array (), or you simply change the dimensions using the dim () function.
For example, you make an array with four columns, three rows, and two “tables” like this:
In the above example, “my.array” is the name of the array we have given. And “←” is the assignment
operator.
There are 24 units in this array mentioned as “1:24” and are divided in three dimensions “(3, 4, 2)”.
Note: - Although the rows are given as the first dimension, the tables are filled column-wise. So, for arrays,
R fills the columns, then the rows, and then the rest.
Alternatively, you could just add the dimensions using the dim ( ) function. This is a little hack that goes a
bit faster than using the array ( ) function; it’s especially useful if you have your data already in a vector.
(This little trick also works for creating matrices, by the way, because a matrix is nothing more than an array
with only two dimensions.)
Say you already have a vector with the numbers 1 through 24, like this:
You can easily convert that vector to an array exactly like my.array simply by assigning the dimensions, like
this:
You can check whether two objects are identical by using theidentical () function.
Page 30 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
We can import Datasets from various sources having various files types for example,
.csv format
Big data tool – Impala
CSV File
The sample data can also be in comma separated values (CSV) format. Each cell inside such data file is
separated by a special character, which usually is a comma, although other characters can be used as well.
The first row of the data file should contain the column names instead of the actual data. Here is a sample of
the expected format.
Col1,Col2,Col3
100,a1,b1
200,a2,b2
300,a3,b3
After we copy and paste the data above in a file named "mydata.csv" with a text editor, we can read the data
with the function read.csv.
In various European locales, as the comma character serves as the decimal point, the
functionread.csv2 should be used instead. For further detail of the read.csv and read.csv2 functions, please
consult the R documentation.
> help(read.csv)
Big data tool – Impala
Cloudera 'Impala', which is a massively parallel processing (MPP) SQL query engine runs natively in
Apache Hadoop.
Page 31 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
RImpala enables querying the data residing in HDFS and Apache HBase from R, which can be further
processed as an R object using R functions. RImpala is now available for download from the Comprehensive
R Archive Network (CRAN) under GNU General Public License (GPL3).
To install RImpala :
We use following code to install RImpala package.
>install. packages("RImpala")
Page 32 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
“Getwd” means get the working directory (wd) and “setwd” is used to set the working directory.
Shaan, 21, M
Ritu, 24, F
Raj, 31, M
Working on Variables
Before learning about creating and modifying variables in R we will know the various operators in R.
Page 33 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Inputting missing data using standard methods and algorithmic approaches (mice package R):
In R, missing values are represented by the symbol NA (not available).
Impossible values (e.g., dividing by zero) are represented by the symbol NaN (not a number).
Unlike SAS, R uses the same symbol for character and numeric data.
For Example,
We have defined “y” and then checked if there is any missing value. T or True means that there is a missing
value.
y <- c(1,2,3,NA)
is.na(y)
# returns a vector (F FF T)
Page 34 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
For Example,
We can create new dataset without missing data as below: -
newdata<- na.omit(mydata)
Or, we can also use “na.rm=TRUE” in argument of the operator. From above example we use na.rm and get
desired result.
x <- c(1,2,NA,3)
mean(x, na.rm=TRUE)
# returns 2
PMM-> Predictive Mean Matching (PMM) is a semi-parametric imputation approach. It is similar to the
regression method except that for each missing value, it fills in a value randomly from among the a observed
donor values from an observation whose regression-predicted values are closest to the regression-predicted
value for the missing value from the simulated regression model.
Outliers:
Outlier is a point or an observation that deviates significantly from the other observations.
●Due to experimental errors or “special circumstances”
●Outlier detection tests to check for outliers
●Outlier treatment –
Retention
Exclusion
Other treatment methods
We have “OUTLIER” package in R to detect and treat outliers in Data.
Normally we use BOX Plot and Scatter plot to find outliers from graphical representation.
Page 35 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Scatter Plot
Box Plot
To merge two data frames (datasets) horizontally, use the merge function. In most cases, you join
two data frames by one or more common key variables (i.e., an inner join).
For example,
To merge two data frames by ID:
total <- merge(data frameA,data frameB,by="ID")
To merge on more than one criteria we pass the argument as follows
To merge two data frames by ID and Country:
total <- merge(data frameA,data frameB,by=c("ID","Country"))
To join two data frames (datasets) vertically, use the rbind function. The two data frames must have
the same variables, but they do not have to be in the same order.
Page 36 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
For example,
total <- rbind(data frameA, data frameB)
Note:-
If data frameA has variables that data frameB does not, then either:
1. Delete the extra variables in data frameA or
2. Create the additional variables in data frameB and set them to NA (missing)
before joining them with rbind( ).
We use cbind() function to combine data by column the syntax is same as rbind().
We use rbind.fill() in plyr package in R. It binds or combines a list of data frames filling missing columns
with NA.
For example,
rbind.fill(mtcars[c("mpg", "wt")], mtcars[c("wt", "cyl")])
“FOR” Loop:-
To repeat an action for every value in a vector by using a “for” loop.
We construct a “for” loop in R as follows:
for(i in values){
... do something ...
}
Page 37 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
“IFELSE” Function:-
When using R, sometimes we need our function to do something if a condition is true and something else if
it is not.
You could do this with two if statements, but there’s an easier way in R: an if...else statement.
An if…else statement contains the same elements as an if statement), and then some extra:
The keyword else, placed after the first code block
A second block of code, contained within braces, that has to be carried out if and only if the result of the
condition in the if() statement is FALSE
For example,
if(hours > 100) net.price<- net.price * 0.9
if(public) {tot.price<- net.price * 1.06 } else
{tot.price<- net.price * 1.12} round(tot.price)}
Page 38 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Case Study:
Segregate Sepal length based on the “Sepal Length” and “Sepal Width”. Use the dataset named “IRIS”.
Data Set
Summary
Page 39 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 40 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Facilitator Preparation
Responsibilities
Review examples provided: reflect on your own experiences and determine when to share
them.
Review all material – Facilitator Guide, Presentation, Guides and Handouts (if any)
Conduct a run through of the content. Conduct a dress rehearsal of the session as you move
through the content. Make sure you are comfortable with the tools and interactions
recommended in the facilitator guide.
Note that all examples are in italics to emphasize key learning points; however, you may use
your own professional experience to enhance the learning.
Page 41 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Principles of Facilitating
Personal Experiences
As a facilitator, you lead participants through prepared scenarios and discussions. During
this process, relate your own professional experience to add realism. Often, personal
experiences on how you helped a colleague through the career ownership process and guided
them to achieving work satisfaction are more memorable than step-by-step instructions on
following the career ownership process. Sharing experiences helps participants understand
how professionals work and think, and gives them the opportunity to apply those lessons to
their own work processes. Also, participants are more likely to remember answers if they
have to think and explore on their own. Your goal is to foster independent thinking and
action rather than having participants depend on your experience.
Experiential Learning
This workshop includes exercises designed to help participants discover the principles of
guiding the participants through the career ownership process and career satisfaction.
Encourage a free-wheeling discussion and call out important trends and insights. Make
liberal use of the whiteboard to capture and display critical participant insights.
Socratic Questions
Your goal throughout the session is to guide participants towards thinking through the
scenarios and discussion questions independently, rather than providing answer. For
example:
The Reality Check worksheet provides valuable What information can you gather from
information about how time is currently spent and the Reality Check worksheet and how
what it would look like in the best case scenario. can the information be used to move
towards career satisfaction?
Page 42 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Time Management
Welcome the participants to the course and move to the introductions.
Introductions
I am <Facilitator’s Name> and I am your facilitator today.”
Briefly review the roles of the Lead Facilitator and Support Facilitator, if any.
Give a brief of your own experience and background.
“Regardless of why you’re here today, we’re all going to walk away with some key benefits – let’s discuss
those briefly.”
Suggested Responses/Benefits to Debrief:
The benefits of this course include:
Efficient and Effective time management
Efficient – Meeting timelines
Effective – Meeting requirement for desired output
Awareness of the SSC environment and time zone understanding
Awareness of the SSC environment and importance of meeting timelines to handoffs
Page 43 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 44 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Provide a brief overview of the session. Discuss the importance of better utilization of time as the
only tool to prevention of slippages on timelines.
Open up the discussion for the session and ask participants to share their thoughts on
“time management”?
Page 45 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Share the SSC model and how working along several time zones is important for the Shared Services
Center.
Activity Description:
Suggested Responses:
Time management has to be looked at an organizational level and not just individual level
These Aspects teach us how to build the blocks of time management.
Prompt participants to come up with some aspects and relate them back to here.
The first 4 Interconnect and Interact to give the 5th one – Results
Page 46 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 47 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Suggested Responses:
False – Time once lost cannot be gotten back – hence important to plan time utilization properly
Suggested Responses:
True – Time lost is lost for every – lost moments cannot be gotten back
Suggested Responses:
True – plan for activities organizational level and also at individual level
Suggested Responses:
True – prioritization should be based on 2x2 matrix of urgency and importance
Page 48 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Team Exercise
List the items and ask participants to classify them as per the quadrant.
Ask the participants to pick up the items listed below and place them in the
Urgent/Important quadrant. Discuss the rationale of their thoughts and categorization.
Activity Description:
1. Refer to the Time Management Quadrant on the display / in the Student
Workbook and categorize the below items.
2. Refer to the Vocabulary Words table if you do not understand the
meaning of a word/term.
Page 49 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Suggested Answers:
Depends on rationale shared
Page 50 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Summary
Page 51 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Work Management
Six steps for expectation setting with the stakeholders
1. Describe the jobs in terms of major outcomes and link to the organization’s need
The first step in expectation setting is to describe the job to the employees. Employees need to
feel there is a greater value to what they do. We need to feel out individual performance has an
impact on the organization’s mission.
Answer this question: My work is the key to ensuring the organization’s success because…
While completing the answer link it to
- Job Description
- Team and Organization’s need
- Performance Criteria
Page 52 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
3. Maximize Performance - Identify what is required to complete the work: Supervisor needs /
Employee needs. Set input as well as output expectations
In order to ensure employees are performing at their best, the supervisor needs to provide not
only the resource (time, infrastructure, desk, recognition etc.) but also the right levels of
direction (telling how to do the task) and support (engaging with employees about the task).
Page 53 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Suggested Responses:
False, work expectations have to be set from day 1 – so that roles & responsibilities are clear
2. True or False? Do not provide too many details when setting expectations.
a. True
b. False
Suggested Responses:
False, as much details with examples can help clarify all the expectations
Suggested Responses:
True, asking the person to rearticulate the understanding of expectations is best way to ensure there is clear
understanding on both sides
4. True or False? Try not to ask too many questions while setting expectations.
a. True
b. False
Suggested Responses:
False, questions always to be encouraged to ensure any clarifications are responded
Page 54 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
5. True or False? Employees need to know what tasks to do and how to communicate,
appreciating work styles.
a. True
b. False
Suggested Responses:
True, provides clarity and enables response based on work styles
6. True or False? Employees do not need to know how their work contributes to
organizational results.
a. True
b. False
Suggested Responses:
False, linking efforts with common goals is very motivating and develops team effort
7. True or False? Employees need to know what their team members’ performance
problems are.
a. True
b. False
Suggested Responses:
True, knowing common problems bring teams focused towards solutions.
8. True or False? Employees how have work style different from the Boss/Peers need
to change.
a. True
b. False
Suggested Responses:
False, they need to adapt and respond based on the partners work style – understanding the work styles is very critical to
enhance team operating performance.
Page 55 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Summary
What
How
Whom to serve
Page 56 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
S
Specific M
Work activities
should be
specific. Why Measurable A
R
and How to be The output
defined metrics and
yardsticks should Achievable
T
be defined. Should be
assigned to
those Realistic
responsible for Should be
achieving it. challenging yet
attainable. Have Time bound
a motivational Time period for
effect achievement is
clearly stated
Page 57 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Efficiency vs Effectiveness
In efficient Efficient
Goals / Doing the Right
Pursuing right goals but in efficient Pursuing right goals and efficient
Pursuit of Appropriate
Effective
Things
Pursuing wrong goals and inefficient. Pursuing Wrong goals but is Efficient
Effective
In
In efficient Efficient
Service Level Agreement (SLA) is a contract between a service provider and its internal or external
customers that documents what services the provider will furnish
SLA measures the service provider’s performance and quality in a number of ways.
Availability and uptime – the percentage of the time services will be available
The no of users being served, the bandwidth or volume being addressed or the quantum f work
being performed in work units
Specify performance benchmarks to which actual performance will be periodically compared
Turnaround time
In addition to establishing performance metrics, an SLA may include a plan for addressing downtime and
documentation for how the service provider will compensate customers in the event of a contract breach.
SLAs, once established, should be periodically reviewed and updated to reflect changes in technology and
the impact of any new regulatory directives
Page 58 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Summary
Every activity must have defined goals and objectives. These goals and objectives
should be SMART complaint (Specific, Measurable, Achievable, Realistic and
Time-bound).
One must balance the efficiency and effectiveness while performing the tasks to
achieve the desired objectives.
The Service Level Agreements should be clearly laid out to measure the quality
and performance.
Course Conclusion
Course Conclusion
“We’ve almost reached the end of the course! Before we wrap up, let’s review what
we’ve learned today”
Ask the participants to recall key learning points from the session and map these
learning points to the course objectives.
Page 59 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Summarizing Data, and By the end of this session, you will be able
to:
Revisiting Probability
1. Summarize Data
2. Work on Probability.
Session Plan:
Activity Location
Summary statistics- summarizing data with R Classroom
Probability Classroom
Summary Classroom
Page 60 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
•summary(data_frame)
•summary(iris)
•Output : Mean, Median, Minimum, Maximum, 1st and 3rd quartile
>summary(dataset)
For example ,
Probability
A probability distribution describes how the values of a random variable are distributed.
For example, the collection of all possible outcomes of a sequence of coin tossing is known to follow
the binomial distribution. Whereas the means of sufficiently large samples of a data population are known
to resemble the normal distribution. Since the characteristics of these theoretical distributions are well
understood, they can be used to make
Page 61 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Now, “At Random” means that there is no biased treatment with any card and the result will be totally at
random.
So, No. of Ace of Diamond in a pack = S = 1
Total no of possible outcomes = Total no. of cards in pack = 52
Probability of positive outcome = S/P = 1/52
That is we have 1.92% chance that we will get positive outcome.
Expected value
The expected value of a random variable is intuitively the long-run average value of repetitions of the
experiment it represents.
For example, the expected value of a dice roll is 3.5 because, roughly speaking, the average of an
extremely large number of dice rolls is practically always nearly equal to 3.5.
Less roughly, the law of large numbers guarantees that the arithmetic mean of the values almost
surely converges to the expected value as the number of repetitions goes to infinity.
The expected value is also known as the expectation, mathematical expectation, EV, mean, or first
moment.
More practically, the expected value of a discrete random variable is the probability-weighted average of all
possible values. In other words, each possible value the random variable can assume is multiplied by its
probability of occurring, and the resulting products are summed to produce the expected value. The same
works for continuous random variables, except the sum is replaced by an integral and the probabilities
by probability densities. The formal definition subsumes both of these and also works for distributions
which are neither discrete nor continuous: the expected value of a random variable is the integral of the
random variable with respect to its probability measure.The expected value is a key aspect of how one
characterizes a probability distribution; it is a location parameter.
Page 62 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Random Variable:
A random variable, aleatory variable or stochastic variable is a variable whose value is subject to
variations due to chance (i.e. randomness, in a mathematical sense). A random variable can take on
a set of possible different values (similarly to other mathematical variables), each with an
associated probability, in contrast to other mathematical variables.
Random variables can be discrete, that is, taking any of a specified finite or countable list of values,
endowed with a probability mass function, characteristic of a probability distribution; or continuous, taking
any numerical value in an interval or collection of intervals, via a probability density function that is
characteristic of a probability distribution; or a mixture of both types. The realizations of a random variable,
that is, the results of randomly choosing values according to the variable's probability distribution function,
are called random variates.
For Example,
If we toss a coin for 10 times and we get heads 8 times then we cannot say that the 11th time if coin is
tossed then we get a head or a tail. But we are sure that we will either get a head or a tail.
Probability Distribution
Page 63 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Probability Distribution Function or PDF is the function that defines probability of outcomes based on
certain conditions.
Based on Conditions, there are majorly 5 types PDFs.
Page 64 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Normal Distribution
We come now to the most important continuous probability density function and perhaps the most
important probability distribution of any sort, the normal distribution.
On several occasions, we have observed its occurrence in graphs from, apparently, widely differing sources:
the sums when three or more dice are thrown; the binomial distribution for large values of n; and in the
hyper geometric distribution.
There are many other examples as well and several reasons, which will appear here, to call this distribution
“normal.”
If
We say that X has a normal probability distribution. A graph of a normal distribution, where we have
chosen a = 0 and b = 1, appears in figure below:
Page 65 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
The normal distribution f(x), with any mean μ and any positive deviation σ, has the following
properties:
It is symmetric around the point x = μ, which is at the same time the mode, the median and the mean
of the distribution.
It is unimodal: its first derivative is positive for x < μ, negative for x > μ, and zero only at x = μ.
Its density has two inflection points (where the second derivative of is zero and changes sign),
located one standard deviation away from the mean, namely at x = μ − σ and x = μ + σ.
Its density is log-concave.
Its density is infinitely differentiable, indeed super smooth of order 2.
Its second derivative f′′(x) is equal to its derivative with respect to its variance σ2.
Page 66 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
is normally distributed.
• In Bayesian statistics, one does not "test normality" per se, but rather computes the likelihood that
the data come from a normal distribution with given parameters μ,σ (for all μ,σ), and compares that with the
likelihood that the data come from other distributions under consideration, most simply using a Bayes factor
(giving the relative likelihood of seeing the data given different models), or more finely taking a prior
distribution on possible models and parameters and computing a posterior distribution given the computed
likelihoods.
1. Graphical methods:
An informal approach to testing normality is to compare a histogram of the sample data to a normal
probability curve. The empirical distribution of the data (the histogram) should be bell-shaped and resemble
the normal distribution. This might be difficult to see if the sample is small. In this case one might proceed
by regressing the data against the quartiles of a normal distribution with the same mean and variance as the
sample. Lack of fit to the regression line suggests a departure from normality.(see Anderson Darling
coefficient and Minitab)
A graphical tool for assessing normality is the normal probability plot, a quantile-quantile plot (QQ plot) of
the standardized data against the standard normal distribution. Here the correlation between the sample data
and normal quartiles (a measure of the goodness of fit) measures how well the data are modeled by a
normal distribution. For normal data the points plotted in the QQ plot should fall approximately on a
straight line, indicating high positive correlation. These plots are easy to interpret and also have the benefit
that outliers are easily identified.
Back-of-the-envelope test
Simple back-of-the-envelope test takes the sample maximum and minimum and computes their z-score, or
more properly t-statistic (number of sample standard deviations that a sample is above or below the sample
mean), and compares it to the 68–95–99.7 rule: if one has a 3σ event (properly, a 3s event) and substantially
fewer than 300 samples, or a 4s event and substantially fewer than 15,000 samples, then a normal
distribution will understate the maximum magnitude of deviations in the sample data.
This test is useful in cases where one faces kurtosis risk – where large deviations matter – and has the
benefits that it is very easy to compute and to communicate: non-statisticians can easily grasp that "6σ
events are very rare in normal distributions".
2. Frequentist tests:
Tests of univariate normality include D'Agostino's K-squared test, the Jarque–Bera test, the Anderson–
Darling test, the Cramér–von Mises criterion, the Lilliefors test for normality (itself an adaptation of the
Kolmogorov–Smirnov test), the Shapiro–Wilk test, the Pearson's chi-squared test, and the Shapiro–Francia
test. A 2011 paper from The Journal of Statistical Modeling and Analytics concludes that Shapiro-Wilk has
the best power for a given significance, followed closely by Anderson-Darling when comparing the
Shapiro-Wilk, Kolmogorov-Smirnov, Lilliefors, and Anderson-Darling tests.
Some published works recommend the Jarque–Bera test. But it is not without weakness. It has low power
for distributions with short tails, especially for bimodal distributions. Other authors have declined to include
Page 67 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Historically, the third and fourth standardized moments (skewness and kurtosis) were some of the earliest
tests for normality. The Jarque–Bera test is itself derived fromskewness and kurtosis estimates. Mardia's
multivariate skewness and kurtosis tests generalize the moment tests to the multivariate case. Other early
test statistics include the ratio of the mean absolute deviation to the standard deviation and of the range to
the standard deviation.
More recent tests of normality include the energy test (Székely and Rizzo) and the tests based on the
empirical characteristic function (ecf) (e.g. Epps and Pulley, Henze–Zirkler, BHEP test). The energy and
the ecf tests are powerful tests that apply for testing univariate or multivariate normality and are statistically
consistent against general alternatives.
The normal distribution has the highest entropy of any distribution for a given standard deviation. There are
a number of normality tests based on this property, the first attributable to Vasicek.
3. Bayesian tests:
Kullback–Leibler divergences between the whole posterior distributions of the slope and variance do not
indicate non-normality. However, the ratio of expectations of these posteriors and the expectation of the
ratios give similar results to the Shapiro–Wilk statistic except for very small samples, when non-
informative priors are used.
Spiegelhalter suggests using a Bayes factor to compare normality with a different class of distributional
alternatives. This approach has been extended by Farrell and Rogers-Stewart.
Page 68 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
The central limit theorem states that under certain (fairly common) conditions, the sum of many random
variables will have an approximately normal distribution.
More specifically, where X1, …, Xn are independent and identically distributed random variables with the
same arbitrary distribution, zero mean, and variance σ2; and Z is their mean scaled by
Then, as n increases, the probability distribution of Z will tend to the normal distribution with zero mean
and variance (σ2).
The central limit theorem also implies that certain distributions can be approximated by the normal
distribution, for example:
• The binomial distribution B(n, p) is approximately normal with mean np and variance np(1−p) for
large n and for p not too close to zero or one.
• The Poisson distribution with parameter λ is approximately normal with mean λ and variance λ, for
large values of λ.
• The chi-squared distribution χ2(k) is approximately normal with mean k and variance 2k, for large
k.
• The Student's t-distribution t(ν) is approximately normal with mean 0 and variance 1 when ν is
large.
Random walk
A random walk is a mathematical formalization of a path that consists of a succession of random steps.
For example, the path traced by a molecule as it travels in a liquid or a gas, the search path of a foraging
animal, the price of a fluctuating stock and the financial status of a gambler can all be modeled as random
walks, although they may not be truly random in reality. The term random walk was first introduced by
Karl Pearson in 1905.
Random walks have been used in many fields: ecology, economics, psychology, computer science, physics,
chemistry, and biology.
Page 69 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 70 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Summary
•summary(data_frame)
•summary(iris)
•Output: Mean, Median, Minimum, Maximum, 1st and 3 rd
Normality tests are used to determine if a data set is well-modeled by a normal distribution and to
compute how likely it is for a random variable underlying the data set to be normally distributed.
A random walk is a mathematical formalization of a path that consists of a succession of random steps.
Probability Distribution Function or PDF is the function that defines probability of outcomes based on
certain conditions.
A random variable, aleatory variable or stochastic variable is a variable whose value is subject to
variations due to chance
Sachin buys a chocolate bar every day during a promotion that says one out of six chocolate bars
has a gift coupon within. Answer the following questions:
•What is the distribution of the number of chocolates with gift coupons in seven days?
•What is the probability that Amir gets no chocolates with gift coupons in seven days?
•Amir gets no gift coupons for the first six days of the week. What is the chance that he will get a
one on the seventh day?
Page 71 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
•Amir buys a bar every day for six weeks. What is the probability that he gets at least three gift
coupons?
•How many days of purchase are required so that Amir’s chance of getting at least one gift
coupon is 0.95 or greater?
Solution:
Steps:
Where
n is the no. of trials
r is the number of successful outcomes
p is the probability of success, and
q is the probability of failure.
5. Number of purchase days required so that probability of success is greater than 0.95:
Page 72 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Page 73 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Facilitator Preparation
Responsibilities
Review examples provided: reflect on your own experiences and determine when to share
them.
Review all material – Facilitator Guide, Presentation, Guides and Handouts (if any)
Conduct a run through of the content. Conduct a dress rehearsal of the session as you
move through the content. Make sure you are comfortable with the tools and interactions
recommended in the facilitator guide.
Note that all examples are in italics to emphasize key learning points; however, you may
use your own professional experience to enhance the learning.
Page 74 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Principles of Facilitating
Personal Experiences
As a facilitator, you lead participants through prepared scenarios and discussions. During this
process, relate your own professional experience to add realism. Often, personal experienceson
how you helped a colleague through the career ownership process and guided them to achieving
work satisfaction are more memorable than step-by-step instructions on following the career
ownership process. Sharing experiences helps participants understand how professionals work and
think, and gives them the opportunity to apply those lessons to their own work processes. Also,
participants are more likely to remember answers if they have to think and explore on their own.
Your goal is to foster independent thinking and action rather than having participants depend on
your experience.
Experiential Learning
This workshop includes exercises designed to help participants discover the principles of guiding the
participants through the career ownership process and career satisfaction. Encourage a free-
wheeling discussion and call out important trends and insights. Make liberal use of the whiteboard
to capture and display critical participant insights.
Socratic Questions
Your goal throughout the session is to guide participants towards thinking through the scenarios
and discussion questions independently, rather than providing answer. For example:
The Reality Check worksheet provides valuable What information can you gather from
information about how time is currently spent and the Reality Check worksheet and how
what it would look like in the best case scenario. can the information be used to move
towards career satisfaction?
Page 75 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Introductions
I am <Facilitator’s Name> and I am your facilitator today.”
Briefly review the roles of the Lead Facilitator and Support Facilitator, if any.
Give a brief of your own experience and background.
“Regardless of why you’re here today, we’re all going to walk away with some key benefits – let’s discuss those briefly.”
Page 76 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
What is a Team?
A team comprises a group of people linked in a common purpose.
Teams are especially appropriate for conducting tasks that are high in complexity and have many
interdependent subtasks.
Team Work
Coming together is a beginning, keeping together is progress and working together is success. A team is a number of people
associated together in work or activity. In a good team members create an environment that allows everyone to go beyond
their limitation.
Why do we need teamwork – The overriding need of all people working together for the same organization is to make the
organization profitable.
Team Building
Page 77 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Trust
Communic
Planning
ation
Decision Problem
Making Solving
Team Development
Page 78 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Team building is any activity that builds and strengthens the team as a team. The teams that are
integrated in Spirit, Enthusiasm, Cohesiveness and Camaraderie are vitally important.
Team building fundamentals
Clear Expectations – Vision/Mission
Context – Background – Why participation in Teams?
Commitment – dedication – Service as valuable to Organization & Own
Competence – Capability – Knowledge
Charter – agreement – Assigned area of responsibility
Control – Freedom & Limitations
Collaboration – Team work
Communication
Creative Innovation
Consequences – Accountable for rewards
Coordination
Cultural Change
Roles of team member
Communicate
Don't Blame Others
Support Group Member's Ideas
No Bragging – No Full of yourself
Listen Actively
Get Involved
Coach, Don't Demonstrate
Provide Constructive Criticism
Try To Be Positive
Value Your Group's Ideas
Page 79 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
1. True or False. The organizations that display a higher level fo team work are
generally more successful.
Suggested Answer:
True
Summary
Page 80 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Key Points
Importance of Professionalism
Provide a brief overview of the session. Discuss the importance of professional behavior in the
organization.
Activity Description:
Ask the candidates “What does professionalism mean to you?”
Page 81 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Nimble. Being flexible and open to change allows these individuals to be quick on their feet
and nimble to the opportunities that they encounter on a daily basis.
Awareness. Having a high level of awareness of themselves, the marketplace, the community
and even the world helps these individuals continually stay on top of things.
Leadership. Last, but not least, professionals demonstrate exceptional leadership skills and
even more importantly self-leadership skill. For if you cannot lead yourself, you cannot lead
others.
What is professionalism?
Professionalism is the competence or set of skills that are expected from a professional.
Professionalism determines how a person is perceived by his employer, co-workers, and
casual contacts.
How long does it take for someone to form an opinion about you?
Studies have proved that it just takes six seconds for a person to form an opinion about
another person.
Empathy
Positive Attitude
Teamwork
Professional Language
Knowledge
Punctual
Confident
Emotionally stable
Grooming
What are the colours that one can opt for work wear?
A good rule of thumb is to have your pants, skirts and blazers in neutral colours. Neutrals are not only restricted to grey
brown and off white - you can also take advantage of the beautiful navies, forest greens, burgundies, tans and caramel
Page 82 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
tones around. Pair these neutrals with blouses, scarves or other accessories in accent colours - ruby red, purple, teal blue,
soft metallic and pinks are some examples.
Things to remember
Wear neat clothes at work which are well ironed and do not stink.
Ensure that the shoes are polished and the socks are clean
Cut your nails on a regular basis and ensure that your hair is in place.
Women should avoid wearing revealing clothes at work.
Remember that the way one presents oneself plays a major role in the professional world.
Suggested Responses:
False
Suggested Responses:
False
Suggested Responses:
False
Page 83 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Activity Description:
Ask the participants to pick up the items listed below and place them in
the Acceptable / Unacceptable category. Discuss the rationale of their
thoughts and categorization.
1. Polo T Shirt –
2. Golf Shoes –
3. Collared Shirt -
4. Suede Shoes –
5. Leather laced Shoes –
6. Matching Socks –
7. Backpacks –
8. Lynards in Pockets –
9. Jeans on weekdays –
10. Rolled Up Sleeves –
11. Matching Belt and Shoes –
12. Pressed Suit –
13. Knee Length Skirt –
14. Short Skirts –
15. Obvious Tatoos –
Suggested Answers:
Page 84 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Summary
Page 85 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Key Points
Effective Communication
Provide a brief overview of the session.
Prompt candidates to discuss the consequences of ineffective or unclear
communication.
We would probably all agree that effective communication is essential to workplace effectiveness.
And yet, we probably don’t spend much time thinking about how we communicate, and how we
might improve our communication skills. The purpose of building communication skills is to
achieve greater understanding and meaning between people and to build a climate of trust,
openness, and support. To a large degree, getting our work done involves working with other
people. And a big part of working well with other people is communicating effectively. Sometimes
we just don’t realize how critical effective communication is to getting the job done. So, let’s have
an experience that reminds us of the importance of effective communication. Actually, this
experience is a challenge to achieve a group result without any communication at all! Let’s give it a
shot.
Activity Description:
Ask the participants to share an experience that reminds them of the significance of
effective communication OR consequences of ineffective communication.
Page 86 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Real communication or understanding happens only when the receiver’s impressionmatches what
the sender intended through his or her expression.So the goal of effective communication is a
mutual understanding of the message.
But how do we know if the other person “gets” our message? We don’t know until wecomplete the
communication.Until a message is complete, the best we can say about its meaning is this:
The meaning of a message is not what is intended by the sender, but what is
understood by the receiver.
So what does it take to complete communication? It takes completing the loop. It takesadding one
more step. It takes feedback.
In simple terms, complete or effective communication means . . .
You say it.
I get it.
You get that I got it.
So far, then, we’ve defined effective communication and what makes it complete. Let’snow
explore the process, or circle, of communication to see how, where, and why itbreaks down.
Forms of Communication:
Page 87 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
The most common way in which we communicate is by talking to the other person. What are the other possible ways in
which you communicate?
Verbal Communication
Verbal communication refers to the use of sounds and language to relay a message. It serves as a vehicle for expressing
desires, ideas and concepts and is vital to the processes of learning and teaching. In combination with nonverbal forms
of communication, verbal communication acts as the primary tool for expression between two or more people.
Page 88 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
- Also, trying to read something into every movement others make can get in the way of
effective interactions.
1. Ambulation is the way one walks. Whether the person switches, stomps, or swaggers can
indicate how that person experiences the environment.
2. Touching is possibly the most powerful nonverbal communication form. People
communicate trust, compassion, tenderness, warmth, and other feelings through touch.
Also, people differ in their willingness to touch and to be touched. Some people are
“touchers” and others emit signals not to touch them.
3. Eye contact is used to size up the trustworthiness of another. Counselors use this
communication method as a very powerful way to gain understanding and acceptance.
Speakers use eye contact to keep the audience interested.
4. Posturing can constitute a set of potential signals that communicate how a person is
experiencing the environment. It is often said that a person who sits with his/her arms
folded and legs crossed is defensive or resistant. On the other hand, the person may just be
cold.
5. Tics are involuntary nervous spasms that can be a key to indicate one is being threatened.
For example, some people stammer or jerk when they are threatened. But these mannerisms
can easily be misinterpreted.
6. Sub-vocals are the non-words one says, such as “ugh” or “um.” They are used when one is
trying to find the right word. People use a lot of non-words trying to convey a message to
another person. Another example is the use of "you know." It is used in place of the "ugh"
and other grunts and groans commonly used.
7. Distancing is a person’s psychological space. If this space is invaded, one can become
somewhat tense, alert, or “jammed up.” People may try to move back to reestablish their
personal space. The kind of relationship and the motives toward one another determines this
personal space.
8. Gesturing carries a great deal of meaning between people, but different gestures can mean
different things to the sender and the receiver. This is especially true between cultures. Still,
Page 89 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
gestures are used to emphasize our words and to attempt to clarify our meaning.
9. Vocalism is the way a message is packaged and determines the signal that is given to
another person. For example, the message, “I trust you,” can have many meanings. “I trust
you” could imply that someone else does not. “I trust you” could imply strong sincerity. “I
trust you” could imply that the sender does not trust others.
Written Communication
Written communication involves any type of message that makes use of
the written word. Written communication is the most important and the
most effective of any mode of business communication.
Page 90 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
- Also, if the receivers of the written message are separated by distance and if they need to
clear their doubts, the response is not spontaneous.
- Written communication is time-consuming as the feedback is not immediate. The encoding
and sending of message takes time.
- Effective written communication requires great skills and competencies in language and
vocabulary use. Poor writing skills and quality have a negative impact on organization’s
reputation.
- Too much paper work and e-mails burden is involved
2 – Ensuring Connectivity
The content that comprises a piece of writing should reflect fluency and should be connected through a logical flow of
thought, in order to prevent misinterpretation and catch the attention of the reader.
Moreover, care should be taken to ensure that the flow is not brought about through a forced/deliberate use of
connectives, as this make the piece extremely uninteresting and artificial.
Page 91 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
6 – Importance of Creativity
In order to hold the readers' attention one needs to be creative to break the tedium of writing and prevent monotony
from creeping in.
This is especially true in the case of all detailed writing that seeks to hold the readers' attention.
Communication Barriers
Page 92 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Suggested Responses:
False
Page 93 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Suggested Responses:
True
Suggested Responses:
True
4. True or False? The best way to get feedback is to ask, “Do you have any
questions?”
a. True
b. False
Suggested Responses:
False
Suggested Responses:
True
Page 94 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Suggested Responses:
True
Suggested Responses:
True
Suggested Responses:
False
Page 95 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Suggested Responses:
True
Suggested Responses:
False
Suggested Responses:
True
Page 96 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
14. True or False? The use of effective visual aids by a speaker usually provides
a
significant increase in the audience’s understanding of the message
a. True
b. False
Suggested Responses:
True
15. True or False? The use of a large vocabulary helps greatly in a person’s
communication effectiveness.
a. True
b. False
Suggested Responses:
True
16. True or False? Most people can listen approximately four times faster than
they speak.
a. True
b. False
Suggested Responses:
True
Suggested Responses:
True
Page 97 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
18. True or False? In getting people to listen, subject content is more important
than the
manner in which the subject is communicated.
a. True
b. False
Suggested Responses:
True
19. True or False? People will accept a logical explanation even if it ignores
their personal
Feelings.
a. True
b. False
Suggested Responses:
True
Summary
Page 98 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Module 1: Unit – 3
SQL using R
Session Time: 180 minutes
Topic Activities
Session Plan:
Material Time
Activity Location
Needed (Minutes)
NO SQL Classroom Student Workbook
Student Workbook
Excel and R integration with R connector Classroom
Student Workbook
Check your understanding Classroom
Student Workbook
Summary Classroom
Page 99 of 126
Facilitator Guide – SSC/ Q2101 – Associate Analytics
Step-by-Step
NO SQL
SQL using R:
data(titanic3, package=”PASWR”)
colnames(titanic3)
head(titanic3)
NO SQL:
A NoSQL (originally referring to "non SQL" or "non-relational") database provides a
mechanism for storage and retrieval of data that is modeled in means other than the tabular
relations used in relational databases.
NoSQL databases are increasingly used in big data and real-time web applications. NoSQL
systems are also sometimes called "Not only SQL" to emphasize that they may support SQL-like
query languages.
There have been various approaches to classify NoSQL databases, each with different categories
and subcategories, some of which overlap.
One type (SQL database) with minor variations Many different types including key-value
Types stores, document databases, wide-column stores,
and graph databases
Developed in 1970s to deal with first wave of Developed in 2000s to deal with limitations of
Development
data storage applications SQL databases, particularly concerning scale,
History
replication and unstructured data storage
Individual records (e.g., "employees") are Varies based on database type. For example, key-
stored as rows in tables, with each column value stores function similarly to SQL databases,
storing a specific piece of data about that record but have only two columns ("key" and "value"),
(e.g., "manager," "date hired," etc.), much like a with more complex information sometimes stored
Data Storage spreadsheet. Separate data types are stored in within the "value" columns. Document databases
Model separate tables, and then joined together when do away with the table-and-row model altogether,
more complex queries are executed. For storing all relevant data together in single
example, "offices" might be stored in one table, "document" in JSON, XML, or another format,
and "employees" in another. When a user wants which can nest values hierarchically.
to find the work address of an employee, the
database engine joins the "employee" and
Structure and data types are fixed in advance. Typically dynamic. Records can add new
To store information about a new data item, the information on the fly, and unlike SQL table rows,
entire database must be altered, during which dissimilar data can be stored together as necessary.
Schemas
time the database must be taken offline. For some databases (e.g., wide-column stores), it is
somewhat more challenging to add new fields
dynamically.
Vertically, meaning a single server must be Horizontally, meaning that to add capacity, a
made increasingly powerful in order to deal database administrator can simply add more
with increased demand. It is possible to spread commodity servers or cloud instances. The
Scaling
SQL databases over many servers, but database automatically spreads data across servers
significant additional engineering is generally as necessary.
required.
Supports Yes, updates can be configured to complete In certain circumstances and at certain levels (e.g.,
Transactions entirely or not at all document level vs. database level)
Can be configured for strong consistency Depends on product. Some provide strong
x <- cbind(rnorm(20),runif(20))
colnames(x) <- c("A","B")
write.table(x,"your_path",sep=",",row.names=FALSE)
3 – Execute R code in VBA
RExcel is from my perspective the best suited tool but there is at least one alternative. You can run a batch
file within the VBA code. If R.exe is in your PATH, the general syntax for the batch file (.bat) is:
Summary
For integration of SQL and R we use SQLDF package.
A NoSQL (originally referring to "non SQL" or "non-relational") database provides a
mechanism for storage and retrieval of data that is modeled in means other than the tabular
relations used in relational databases.
NoSQL support SQL like query language.
NoSQL is used primarily in compliment to Big Data tools.
Excel can be integrated with R using R-Connector.
How to execute VBA code in R tool?
Module 1 UNIT – 4
Correlation and Regression
Topic Session Goals
Correlation and Regression By the end of this session, you will be able
to:
1. Make Regression models
2. Find Correlation
3. Understand Multi Collinearity
4. Work on Multiple Regression
5. Work with Dummy variables
Session Plan:
Activity Location
Basic Regression Analysis Classroom
Correlation Classroom
Heteroscedasticity Classroom
We get the intercept “C” and the slope “m” of the equation – Y=mX+C
The fit information displays four charts: Residuals vs. Fitted, Normal Q-Q, Scale-Location, and
Residuals vs. Leverage.
Below are the various graphs representing values of regression
OLS Regression
OLS:- Ordinary least squares (OLS) or linear least squares is a method for estimating the unknown
parameters in a linear regression model, with the goal of minimizing the differences between the
observed responses in some arbitrary dataset and the responses predicted by the linear approximation of
the data.
This is applied in both simple linear and multiple regression where the common assumptions are
(1) The model is linear in the coefficients of the predictor with an additive random error term
(2) The random error terms are
normally distributed with 0 mean and
a variance that doesn't change as the values of the predictor covariates (i.e. IVs) change.
Regression Modeling
Regression modeling or analysis is a statistical process for estimating the relationships among
variables. It includes many techniques for modeling and analyzing several variables, when the
focus is on the relationship between a dependent variable and one or more independent
variables (or 'predictors'). More specifically, regression analysis helps one understand how the
typical value of the dependent variable (or 'criterion variable') changes when any one of the
independent variables is varied, while the other independent variables are held fixed. Most
commonly, regression analysis estimates the conditional expectation of the dependent variable
given the independent variables – that is, the average value of the dependent variable when the
independent variables are fixed. Less commonly, the focus is on a quantile, or other location
parameter of the conditional distribution of the dependent variable given the independent
variables. In all cases, the estimation target is a function of the independent variables called the
regression function. In regression analysis, it is also of interest to characterize the variation of
the dependent variable around the regression function which can be described by a probability
distribution.
Regression analysis is widely used for prediction and forecasting, where its use has substantial
overlap with the field of machine learning. Regression analysis is also used to understand which
among the independent variables are related to the dependent variable, and to explore the forms
of these relationships. In restricted circumstances, regression analysis can be used to infer
causal relationships between the independent and dependent variables. However this can lead to
illusions or false relationships, so caution is advisable; for example, correlation does not imply
causation.
Many techniques for carrying out regression analysis have been developed. Familiar methods
such as linear regression and ordinary least squares regression are parametric, in that the
regression function is defined in terms of a finite number of unknown parameters that are
estimated from the data. Nonparametric regression refers to techniques that allow the regression
function to lie in a specified set of functions, which may be infinite-dimensional.
The performance of regression analysis methods in practice depends on the form of the data
generating process, and how it relates to the regression approach being used. Since the true
form of the data-generating process is generally not known, regression analysis often depends to
some extent on making assumptions about this process. These assumptions are sometimes
testable if a sufficient quantity of data is available. Regression models for prediction are often
useful even when the assumptions are moderately violated, although they may not perform
optimally. However, in many applications, especially with small effects or questions of
causality based on observational data, regression methods can give misleading results.
In a narrower sense, regression may refer specifically to the estimation of continuous response
variables, as opposed to the discrete response variables used in classification. The case of a
continuous output variable may be more specifically referred to as metric regression to
distinguish it from related problems.
Regression residuals
The residual of an observed value is the difference between the observed value and the estimated value
of the quantity of interest.
Because a linear regression model is not always appropriate for the data, you should assess the
appropriateness of the model by defining residuals and examining residual plots.
Residuals
The difference between the observed value of the dependent variable (y) and the predicted value (ŷ) is
called the residual (e). Each data point has one residual.
Both the sum and the mean of the residuals are equal to zero. That is, Σ e = 0 and e = 0.
Residual Plots
A residual plot is a graph that shows the residuals on the vertical axis and the independent variable on
the horizontal axis. If the points in a residual plot are randomly dispersed around the horizontal axis, a
linear regression model is appropriate for the data; otherwise, a non-linear model is more appropriate.
Below the table on the left shows inputs and outputs from a simple linear regression analysis, and the
chart on the right displays the residual (e) and independent variable (X) as a residual plot.
x 60 70 80 85 95
y 70 65 70 95 85
The residual plot shows a fairly random pattern - the first residual is positive, the next two are negative,
the fourth ispositive, and the last residual is negative. This random pattern indicates that a linear model
provides a decent fit to the data.
Below, the residual plots show three typical patterns. The first plot shows a random pattern, indicating
a good fit for a linear model. The other plot patterns are non-random (U-shaped and inverted U),
suggesting a better fit for a non-linear model.
Correlation
Correlation Coefficients:-
•r : correlation coefficient
•+1 : Perfectly positive
•-1 : Perfectly negative
•0 – 0.2 : No or very weak association
•0.2 – 0.4 : Weak association
•0.4 – 0.6 : Moderate association
•0.6 – 0.8 : Strong association
•0.8 – 1 : Very strong to perfect association
Heteroscedasticity:
A collection of random variables is heteroscedastic (or 'heteroskedastic' from Ancient Greek hetero
“different” and skedasis “dispersion”) if there are sub-populations that have different variabilities from
others. Here "variability" could be quantified by the variance or any other measure of statistical
dispersion. Thus heteroscedasticity is the absence of homoscedasticity.
The existence of heteroscedasticity is a major concern in the application of regression analysis,
including the analysis of variance, as it can invalidate statistical tests of significance that assume that
the modeling errors are uncorrelated and uniform—hence that their variances do not vary with the
effects being modeled. For instance, while the ordinary least squares estimator is stillunbiased in the
presence of heteroscedasticity, it is inefficient because the true variance and covariance are
underestimated. Similarly, in testing for differences between sub-populations using a location test,
some standard tests assume that variances within groups are equal.
Test of Heteroscedasticity:-
Tests in regression
• Levene's test
• Goldfeld–Quandt test
• Park test
• Glejser test
• Brown–Forsythe test
• Harrison–McCabe test
• Breusch–Pagan test
• White test
• Cook–Weisberg test
Fixes:-
There are four common corrections for heteroscedasticity. They are:
View logarithmized data. Non-logarithmized series that are growing exponentially often appear
to have increasing variability as the series rises over time. The variability in percentage terms
may, however, be rather stable.
Use a different specification for the model (different X variables, or perhaps non-linear
transformations of the X variables).
Apply a weighted least squares estimation method, in which OLS is applied to transformed or
weighted values of X and Y. The weights vary over observations, usually depending on the
changing error variances. In one variation the weights are directly related to the magnitude of
the dependent variable, and this corresponds to least squares percentage regression.
Heteroscedasticity-consistent standard errors (HCSE), while still biased, improve upon OLS
estimates. HCSE is a consistent estimator of standard errors in regression models with
heteroscedasticity. This method corrects for heteroscedasticity without altering the values of the
coefficients. This method may be superior to regular OLS because if heteroscedasticity is
present it corrects for it, however, if the data is homoscedastic, the standard errors are
equivalent to conventional standard errors estimated by OLS. Several modifications of the
White method of computing heteroscedasticity-consistent standard errors have been proposed as
corrections with superior finite sample properties.
Autocorrelation
A more flexible test, covering autocorrelation of higher orders and applicable whether or not the
regressors include lags of the dependent variable, is the Breusch–Godfrey test. This involves an
auxiliary regression, wherein the residuals obtained from estimating the model of interest are
regressed on (a) the original regressors and (b) k lags of the residuals, where k is the order of
the test. The simplest version of the test statistic from this auxiliary regression is TR2, where T
is the sample size and R2 is the coefficient of determination. Under the null hypothesis of no
autocorrelation, this statistic is asymptotically distributed as x2 with k degrees of freedom.
Multicollinearity
In statistics, multicollinearity (also collinearity) is a phenomenon in which two or more predictor
variables in a multiple regression model are highly correlated, meaning that one can be linearly
predicted from the others with a substantial degree of accuracy. In this situation the coefficient
estimates of the multiple regressions may change erratically in response to small changes in the model
or the data. Multicollinearity does not reduce the predictive power or reliability of the model as a
whole, at least within the sample data set; it only affects calculations regarding individual predictors.
That is, a multiple regression model with correlated predictors can indicate how well the entire bundle
of predictors predicts the outcome variable, but it may not give valid results about any individual
predictor, or about which predictors are redundant with respect to others.
In case of perfect multicollinearity the predictor matrix is singular and therefore cannot be inverted.
Under these circumstances, for a general linear model y=Xβ+ε, the ordinary estimator
5) Condition number test: The standard measure of ill-conditioning in a matrix is the condition
index. It will indicate that the inversion of the matrix is numerically unstable with finite-precision
numbers (standard computer floats and doubles). This indicates the potential sensitivity of the
computed inverse to small changes in the original matrix. The Condition Number is computed by
finding the square root of (the maximum eigenvalue divided by the minimum eigenvalue). If the
Condition Number is above 30, the regression may have significant multicollinearity; multicollinearity
exists if, in addition, two or more of the variables related to the high condition number have high
proportions of variance explained. One advantage of this method is that it also shows which variables
are causing the problem.
6) Farrar–Glauber test: If the variables are found to be orthogonal, there is no multicollinearity; if
the variables are not orthogonal, then multicollinearity is present. C. Robert Wichers has argued that
Farrar–Glauber partial correlation test is ineffective in that a given partial correlation may be
compatible with different multicollinearity patterns. The Farrar–Glauber test has also been criticized by
other researchers.
7) Perturbing the data. Multicollinearity can be detected by adding random noise to the data and
re-running the regression many times and seeing how much the coefficients change.
8) Construction of a correlation matrix among the explanatory variables will yield indications as to
the likelihood that any given couplet of right-hand-side variables is creating multicollinearity problems.
Correlation values (off-diagonal elements) of at least .4 are sometimes interpreted as indicating a
multicollinearity problem. This procedure is, however, highly problematic and cannot be
recommended. Intuitively, correlation describes a bivariate relationship, whereas collinearity is a
multivariate phenomenon.
The general purpose of multiple regressions (the term was first used by Pearson, 1908) is to learn more
about the relationship between several independent or predictor variables and a dependent or criterion
variable.
For example,
A real estate agent might record for each listing the size of the house (in square feet), the number of
bedrooms, the average income in the respective neighborhood according to census data, and a
subjective rating of appeal of the house. Once this information has been compiled for various houses it
would be interesting to see whether and how these measures relate to the price for which a house is
sold. For example, you might learn that the number of bedrooms is a better predictor of the price for
which a house sells in a particular neighborhood than how "pretty" the house is (subjective rating).
Dummy Variables
In regression analysis, a dummy variable (also known as an indicator variable, design variable, Boolean
indicator, categorical variable, binary variable, or qualitative variable) is one that takes the value 0 or 1
to indicate the absence or presence of some categorical effect that may be expected to shift the
outcome. Dummy variables are used as devices to sort data into mutually exclusive categories (such as
smoker/non-smoker, etc.).
In other words, Dummy variables are "proxy" variables or numeric stand-ins for qualitative facts in a
regression model. In regression analysis, the dependent variables may be influenced not only by
quantitative variables (income, output, prices, etc.), but also by qualitative variables (gender, religion,
geographic region, etc.). A dummy independent variable (also called a dummy explanatory variable)
which for some observation has a value of 0 will cause that variable's coefficient to have no role in
influencing the dependent variable, while when the dummy takes on a value 1 its coefficient acts to
alter the intercept.
For example,
Suppose Gender is one of the qualitative variables relevant to a regression. Then, female and male
would be the categories included under the Gender variable. If female is arbitrarily assigned the value
of 1, then male would get the value 0. Then the intercept (the value of the dependent variable if all
other explanatory variables hypothetically took on the value zero) would be the constant term for males
but would be the constant term plus the coefficient of the gender dummy in the case of females.
Check your understanding In the context of regression analysis, which of the following statements are
true?
I. When the sum of the residuals is greater than zero, the data set is nonlinear.
II. A random pattern of residuals supports a linear model.
III. A random pattern of residuals supports a non-linear model.
(A) I only
(B) II only
(C) III only
(D) I and II
(E) I and III
Summary Solution
Regression is method of establishing relation between two or more variables.
Correlation shows the extent of relation
The correct answer is (B). A random pattern of residuals supports a linear model; a
Correlation coefficient lies between -1 and supports
non-random pattern 1. a non-linear model. The sum of the residuals is
Ordinary least squares (OLS)
always or linear
zero, least the
whether squares is aismethod
data set linear orfor estimating the
nonlinear.
unknown parameters in a linear regression model.
Multiple regressions find relationship between several independent or predictor variables
and a dependent or criterion variable.
Multicollinearity (also collinearity) is a phenomenon in which two or more predictor
variables in a multiple regression model are highly correlated.
A dummy variableis one that takes the value 0 or 1 to indicate the absence or presence of
some categorical effect.
Autocorrelation, also known as serial correlation or cross-autocorrelation
Module 1 UNIT – 5
Understand the verticals and requirements gathering
Session Time: 180 minutes
Topic Activities
Understand the Verticals and Requirement Gathering By the end of this session, you will be
able to:
1. Solve Engg. & Manufac. Issues
2. Create Business Models
Session Plan:
Activity Location
Understand systems viz. Engineering
Design, Manufacturing, smart utilities,
Classroom
production lines, Automotive industries,
Tech system
Summary Classroom
Step-by-Step
Understand systems viz. Engineering Design, Manufacturing, smart utilities, production lines,
Automotive industries, Tech system
Engineering Design:
The engineering design process is a methodical series of steps that engineers use in creating
functional products and processes. The process is highly iterative - parts of the process often need
to be repeated many times before production phase can be entered - though the part(s) that get
iterated and the number of such cycles in any given project can be highly variable.
One framing of the engineering design process delineates the following stages: research,
conceptualization, feasibility assessment, establishing design requirements, preliminary
design, detailed design, production planning and tool design, and production.
Manufacturing:
Manufacturing is the production of merchandise for use or sale using labour and machines, tools,
chemical and biological processing, or formulation. The term may refer to a range of human
activity, from handicraft to high tech, but is most commonly applied to industrial production, in
which raw materials are transformed into finished goods on a large scale. Such finished goods may
be used for manufacturing other, more complex products, such as aircraft, household appliances or
automobiles, or sold to wholesalers, who in turn sell them to retailers, who then sell them to end
users – the "consumers".
Manufacturing takes turns under all types of economic systems. In a free market economy,
manufacturing is usually directed toward the mass production of products for sale to consumers at
a profit. In a collectivist economy, manufacturing is more frequently directed by the state to supply
a centrally planned economy. In mixed market economies, manufacturing occurs under some
degree of government regulation.
Modern manufacturing includes all intermediate processes required for the production and
integration of a product's components. Some industries, such as semiconductor and steel
manufacturers use the term fabrication instead.
The manufacturing sector is closely connected with engineering and industrial design. Examples
of major manufacturers in North America include General Motors Corporation, General Electric,
Procter & Gamble, General Dynamics, Boeing, Pfizer, and Precision Cast parts. Examples in
Europe include Volkswagen Group, Siemens, and Michelin. Examples in Asia include Sony,
Huawei, Lenovo, Toyota, Samsung, and Bridgestone.
SMART Utilities:
When S.M.A.R.T. data indicates a possible imminent drive failure, software running on the host
system may notify the user so stored data can be copied to another storage device, preventing data
loss, and the failing drive can be replaced.
Understand the business problem related to engineering, Identify the critical issues. Set business
objectives.
The BA process can solve problems and identify opportunities to improve business performance. In
the process, organizations may also determine strategies to guide operations and help achieve
competitive advantages. Typically, solving problems and identifying strategic opportunities to follow
are organization decision-making tasks. The latter, identifying opportunities can be viewed as a
problem of strategy choice requiring a solution.
There are many different approaches that can be used to gather information about a business. They
include the following:
The business analyst should use one-on-one interviews early in the business analysis project to gage
the strengths and weaknesses of potential project participants and to obtain basic information about the
business. Large meetings are not a good use of time for data gathering.
Facilitated work sessions are a good mechanism for validating and refining “draft”
requirements. They are also useful to prioritize final business requirements. Group dynamics can
often generate even better ideas.
Primary or local data is collected by the business owner and can be collected by survey, focus group
or observation. Third party static data is purchased in bulk without a specific intent in mind. While
easy to get (if you have the cash) this data is not specific to your business and can be tough to sort
through as you often get quite a bit more data than you need to meet your objective. Dynamic data is
collected through a third party process in near real-time from an event for a specific purpose (read into
that VERY expensive).
Three key questions you need to ask before making a decision about the best method for your firm.
Is the data collection for a stand-alone event or for part of a broader data collection effort?
Business intelligence (BI) is the set of techniques and tools for the transformation of raw data into
meaningful and useful information for business analysis purposes. BI technologies are capable of
handling large amounts of unstructured data to help identify, develop and otherwise create new
strategic business opportunities. The goal of BI is to allow for the easy interpretation of these large
volumes of data. Identifying new opportunities and implementing an effective strategy based on
insights can provide businesses with a competitive market advantage and long-term stability.
BI technologies provide historical, current and predictive views of business operations. Common
functions of business intelligence technologies are reporting, online analytical processing, analytics,
data mining, process mining, complex event processing, business performance management,
benchmarking, text mining, predictive analytics and prescriptive analytics.
BI can be used to support a wide range of business decisions ranging from operational to strategic.
Basic operating decisions include product positioning or pricing. Strategic business decisions include
priorities, goals and directions at the broadest level. In all cases, BI is most effective when it combines
data derived from the market in which a company operates (external data) with data from company
sources internal to the business such as financial and operations data (internal data). When combined,
external and internal data can provide a more complete picture which, in effect, creates an
"intelligence" that cannot be derived by any singular set of data.
Business intelligence can be applied to the following business purposes, in order to drive business value.
Measurement – program that creates a hierarchy of performance metrics (see also Metrics
Reference Model) and benchmarking that informs business leaders about progress towards
business goals (business process management).
Analytics – program that builds quantitative processes for a business to arrive at optimal
decisions and to perform business knowledge discovery. Frequently involves: data mining,
process mining, statistical analysis, predictive analytics, predictive modeling, business process
modeling, data lineage, complex event processing and prescriptive analytics.
Reporting/enterprise reporting – program that builds infrastructure for strategic reporting to
serve the strategic management of a business, not operational reporting. Frequently involves
data visualization, executive information system and OLAP.
Collaboration/collaboration platform – program that gets different areas (both inside and
outside the business) to work together through data sharing and electronic data interchange.
Knowledge management – program to make the company data-driven through strategies and
practices to identify, create, represent, distribute, and enable adoption of insights and
experiences that are true business knowledge. Knowledge management leads to learning
management and regulatory compliance.
In addition to the above, business intelligence can provide a pro-active approach, such as alert
functionality that immediately notifies the end-user if certain conditions are met. For example, if some
business metric exceeds a pre-defined threshold, the metric will be highlighted in standard reports, and
the business analyst may be alerted via e-mail or another monitoring service. This end-to-end process
requires data governance, which should be handled by the expert.
1. Keep it VERY simple. I recommend one page with 3-4 questions maximum. Customers are
Summary
Engineering design process is a methodical series of steps that engineers use in creating
functional products and processes.
Manufacturing is the production of merchandise for use or sale using labor and machines,
tools, chemical and biological processing, or formulation.
Assembly line or Production line concept was first used by Henry Ford in automobile
industry. It reduces production time drastically.
Most of the critical Business problems are solved with help of Data Analytics.