Business Research Method Notes Mba
Business Research Method Notes Mba
HASSANPUR
NOTES
What is Research?
Objectives of Research:
The main aim of research is to find out the truth which is hidden and which
has not been discovered as yet.
Though each research study has its own specific purpose, we may think
of research objectives as falling into a number of following broad
groupings:
Let say there’s an automobile company that is planning to launch a car that runs
on CNG. In order to promote cleaner fuel, the company will be involved in
developing different plans and strategies to identify the demand for the car they
are intending to launch. Other than this, the company will also look for
competitors, the target audience, keeping in mind the distribution of CNG in
India. Hence the researches are conducted on various ideas to formulate a
sustainable and more efficient design.
Types of Research :
1. DescriptiveResearch
2. Exploratory Research
3. Applied Research
4. Fundamental Research
5. Quantitative Research
6. Qualitative Research
7. Conceptual Research
8. Empirical Research
9. Experimental Research
10. Historical Research
Graduate students learn about research methods used in their particular field
of study. Whatever methods are used, there are many resources to support
research, and any number of variations to the basic methods. Choose a
method or variation that is manageable in your first years as a new faculty
member to be certain to get your research agenda underway. Often the
biggest challenge is often to get started, so establish research priorities early
on, and create a plan to impliment them.
Quantitative Research
Qualitative Research
Collaborative Research
There are many ways to collaborate and thereby enrich your work as a faculty
member. Multiple perspectives offer a more complete view of an issue under
study. Whether presenting or publishing, having several iterations of a study
in different settings offers a means of validating findings. Collaboration may
take many forms, including crossing disciplines, types of institutions, or
engaging the community outside the campus environment. Explore the ways
that collaboration may enrich your research. Some of the issues and
challenges associated with collaborative research are explored in this section.
Practitioner Research
3. Exploratory research
a. Involves exploring a general aspect.
b. Includes studying of a problem, about which nothing or a very little is known.
c. Follows a very formal approach of research.
d. Helps in exploring new ideas.
e. Helps in gathering information to study a specific problem very minutely.
f. Helps in knowing the feasibility in attempting a study.
4. Descriptive research
a. Simplest form of research.
b. More specific in nature and working than exploratory research.
c. It involves a mutual effort.
d. Helps in identifying various features of a problem.
e. Restricted to the problems that are describable and not arguable and the
problems in which valid standards can be developed for standards.
f. Existing theories can be easily put under test by empirical observations.
g. Underlines factors that may lead to experimental research.
h. It consumes a lot of time.
i. It is not directed by hypothesis.
5. Diagnostic study
a. Quite similar to the descriptive research.
b. Identifies the causes of the problems and then solutions for these problems.
c. Related to causal relations.
d. It is directed by hypothesis.
e. Can be done only where knowledge is advanced.
6. Evaluation study
a. Form of applied research.
b. Studies the development project.
c. Gives access to social or economical programmes.
d. Studies the quality and also the quantity of an activity.
7. Action research
a. Type of evaluation study.
b. Is a concurrent evaluation study.
In this article, we discuss what a theoretical framework is, how it differs from a
conceptual framework, why you’d use a theoretical framework and how to create one
in four steps, plus we provide an example for you to use when creating your own
framework.
What is a theoretical framework?
Before defining what a theoretical framework is, it’s important to first understand
what a theory is.
Researchers use theoretical frameworks to explain the theories they're using within
their research and provide their own research with context by identifying the
assumptions that inform their work. A theoretical framework is often integrated into a
literature review section near the beginning of a paper or experiment but can also be
included as its own chapter or section.
The terms “theoretical framework” and “conceptual framework” are often used
interchangeably but there are key differences between the two:
The same research topic can be approached differently within different theoretical
frameworks. For example, in psychology, different kinds of therapists might address
trauma using various theories. Behavioral therapists may prefer to treat symptoms of
trauma using exposure therapy while another therapist might recommend an eye
movement desensitization and reprocessing (EMDR) session.
Here's how to create a theoretical framework for your study or research paper in four
steps:
The first step in creating a theoretical framework is to define your research objective
and then gather supporting documents. Consider what you hope to achieve with your
study and what new information you aim to bring to your chosen field.
You can also think about the subject you're planning to study and how your research
might affect the knowledge the scientific community already has about the subject
matter. To prepare for a successful study, define each of your objectives clearly to
help your audience understand them.
Once you define your objectives, you can write a problem statement to define the
purpose of the study and what problems it's addressing. For example, if you're
researching the moons of Mars you might create a problem statement like this:
Rather than providing in-depth information, try using concise and straightforward
language in your problem statement to explain your study’s purpose in an easy-to-
read manner.
A research question is an inquiry that you aim to address with your study results.
These questions often provide a basis for your theories or hypotheses. They can also
help your audience understand the study’s purpose and the information you're
seeking. This can clarify to your audience whether your study was successful.
A researcher studying Mars’ moons might ask the following research questions:
• How can I prove the astral body I'm observing is another moon?
• What tools can I use to verify my findings?
• What qualities am I looking for to verify the object shares an orbit with Mars?
• What factors affect the object's behavior in space?
• Why has no one else observed the object until now?
If you plan to answer more than one question with your study, consider adding a
section of subquestions to support your main research question.
4. Create a literature review to highlight relevant theories
A literature review is a summary or analysis of all the research that you read before
doing your own research. It collects key sources on a topic and outlines how those
sources and other researchers' theories align with your study and how you're using
their research to develop your theories. Since some research requires several theories
from a variety of studies, it's helpful to define why each theory applies to your
research in your theoretical framework.
For example, if you're studying Mars' moons, you might evaluate research from
astronomers, astrophysicists or other scientists to determine whether their work
supports or denies the existence of another Mars moon.
• 2. As noted, a research hypothesis is more than just a topic. It has two elements
(variables) that are in relation to each other. Remember that, within the word
"hypothesis" is the word "thesis." Your hypothesis is what you propose to “prove” by
your research. As a result of your research, you will arrive at a conclusion, a theory,
or understanding that will be useful or applicable beyond the research itself.
• 3. Avoid judgmental words in your hypothesis. Value judgments are subjective and
are not appropriate for a hypothesis. You should strive to be objective. Therefore the
use of personal opinion is to be avoided.
• 4. Your hypothesis must involve an issue or question that cannot be answered
exclusively by the discipline of law. You should try to limit your inquiry to the
literatures of 2 or 3 disciplines. It is best to choose a hypothesis where you already
have some level of familiarity with the disciplines that are most relevant to the topic.
• 5. Be sure that each term in your hypothesis is clearly understood and defined; do
not deal in generalities or assume that the reader knows the meaning of a technical
term.
• 6. Specify, if appropriate, whether you will be dealing with state or federal law or
both on a comparative basis if appropriate.
• 7. Know that your hypothesis may change over time as your research progresses.
You must obtain the professor's approval of your hypothesis, as well as any
modifications to your hypothesis, before proceeding with any work on the topic. Your
will be expressing your hypothesis in 3 ways:
• As a one-sentence hypothesis
• As a research question
• As a title for your paper Your hypothesis will become part of your research
proposal
What is literature survey in research?
A literature review is a comprehensive summary of previous research on a topic.
The literature review surveys scholarly articles, books, and other sources relevant to
a particular area of research. The review should enumerate, describe, summarize,
objectively evaluate and clarify this previous research
Different types of literature reviews
• Narrative or Traditional literature reviews. Narrative or
Traditional literature reviews critique and summarise a body of literature about the
thesis topic. ...
• Scoping Reviews. ...
• Systematic Quantitative Literature Review. ...
• Cochrane Reviews. ...
• Campbell Collaboration.
A good research always starts with a good problem. You can observe people or
things, visit places, read print materials, or consult experts to find the research
problem that is right for you. The research problem guides you in formulating the
hypothesis and interpretation of your findings so that you can formulate the right
conclusion. A good research problem is important because it is the basis of all
subsequent research activities you are going to undertake. Factors like area of
interest, availability of fund, socio-economic significance of the study, and the safety
measures to be undertaken should be considered in finding a good research problem.
2. Formulation of Hypothesis
After finding your research problem, the next step is to formulate your own
hypothesis. A hypothesis is a theoretical statement in solving a logical relationship
between variables. Do not be afraid if your hypothesis proves to be incorrect after the
experimentation because it is only considered as an educated guess. Always
remember that when you formulate a hypothesis, it should be based on the research
problem being solved.
A research problem is vague at first. To give you a vivid picture of the whole
research, you shall read various publications or surf the internet to become aware of
the previous works already done. In doing so, it could spur an idea that can be the
subject of your investigation. The review of related literature can be taken from
science books, magazines, journals, newspapers, or even in the internet.
A research design is the blueprint of the research you are going to undertake. It serves
as the work plan of the whole study not only because it entails the resources needed
in conducting the research but also the ways these resources are utilized.
5. Actual experimentation
Actual experimentation is an implementation of the research design. In actual
experimentation, you have to conduct an experiment to prove the validity of the
hypothesis you have formulated. Actual experimentation includes the methodology
that you have followed in doing your research. The methodology should be carefully
planned prior to the actual experimentation to ensure the validity and accuracy of the
result.
This is the heart of the research process because this is part where the findings of the
research can be found. You can use table (not the table in your kitchen) and graph to
interpret the results of your research.
Conclusion is a statement where you will present the solution to the proposed
problem based on the findings of the investigation. They are tied up to the questions
investigated. Your conclusion will show whether or not your experiment worked. It
should answer your hypothesis and research problem. In your concluding statement
you can also infer on the possible benefits to society that your results might present.
You can state any plans you might have to continue working on other aspects related
to your area of study. We must remember that recommendations are based on
conclusions and conclusions are based on findings
Research Proposal
A research proposal describes what you will investigate, why it's important, and
how you will conduct your research. The format of a research proposal varies
between fields, but most proposals will contain at least these elements: Title page.
Introduction. Literature review.
The purpose of the research proposal (it’s job, so to speak) is to convince your
research supervisor, committee or university that your research is suitable (for the
requirements of the degree program) and manageable (given the time and resource
constraints you will face).
• Cover Page: Contains your project title, your name, your supervisor's name,
program/department, institution or affiliation, and date.
• Table of Contents: Outlines the contents of your entire proposal with
respective page numbers.
• Introduction: Contains background and context, a problem statement, research
questions, and the rationale behind the study.
• Literature review: Contains key concepts and theories that serve as the
framework for your study as well as any gaps in research.
• Research design and methods: Contains research objectives, method, and
potential limitations
• Implications: Explains how the study can be applied to the existing field of
knowledge on the topic.
• Reference list: A list of references used to write the proposal.
• Research schedule: A timeline of research phases and how they will achieve
the objective and meet deadlines.
In compliance with APA style, you can use these sections as headings for your
document as well. Using section headings makes information more organized for the
reader and allows them to follow the author's thoughts more clearly.
WRITING STYLE
Besides the contents of your proposal, you also need to pay attention to your writing
style. It is going to be different from other papers or documents you may have had to
write in the past. According to Academic Writer (n.d.), the following are some of the
main elements of writing style. These are important to making your proposal sound
respectful and professional.
Language
Instead of using common language, which is the type of language we use in normal
conversations, you want to use the "language of research" or the "language of
science." This means that if a term has two meanings, you should only use the term
for the meaning that is the most relevant to your research. For example, if a chemist
uses the word "element" in a proposal, they use it only in the context of its scientific
definition. This prevents the reader from getting confused throughout the
document. Avoid creating new terms in your proposal and be sure to clearly define
unfamiliar words at the beginning of the proposal (Locke et al., 2007). Lastly, you
also want to avoid using first person in your proposal ("I will...") as it does not
demonstrate professionalism in writing.
Tone
The tone of your writing should be professional and serious. In other words, use
"academic voice" in your proposal writing. Academic voice is meant to convey your
thoughts and distinguish them from other authors (Robbins, 2016). It is comprised of
three elements ("What are the three elements," n.d.):
These elements make your academic writing unique from other writers and present
your thoughts in a professional manner.
Clarity
You want to ensure that your writing is precise so that readers have a clear
understanding of your project. Proposals should exclude excessive jargon (technical
terms), slang, and abbreviations. They should also make logical comparisons
between ideas to prevent readers from getting confused or lost ("Academic Writer,"
n.d.). Here are some general tips for ensuring clarity in your writing:
• Using a term consistently throughout your paper (it refers to the same meaning
throughout the document).
• Do not use excessive jargon or technical terms, and make sure you define any
new terms.
• Draw comparisons between concepts to avoid ambiguity. This requires using
proper word choice and sentence structure.
Conciseness
Continuity
• Be mindful of how you use punctuation marks. This includes commas, dashes,
and hyphens.
• Use transitional words (and, or, therefore, etc.) to maintain flow.
Tone
Verb Tense
• Use appropriate verb tenses to reflect series of events and timelines in your
proposal.
Unit-2
Research Design
• Data collection
• Measurement
• Data Analysis
The research problem an organization faces will determine the design, not vice-versa.
The design phase of a study determines which tools to use and how they are used.
Impactful research usually creates a minimum bias in data and increases trust in the
accuracy of collected data. A design that produces the slightest margin of error in
experimental research is generally considered the desired outcome. The essential
elements are:
A proper design sets your study up for success. Successful research studies provide
insights that are accurate and unbiased. You’ll need to create a survey that meets all
of the main characteristics of a design. There are four key characteristics:
• Neutrality: When you set up your study, you may have to make assumptions
about the data you expect to collect. The results projected in the research should
be free from bias and neutral. Understand opinions about the final evaluated
scores and conclusions from multiple individuals and consider those who agree
with the results.
• Reliability: With regularly conducted research, the researcher expects similar
results every time. You’ll only be able to reach the desired results if your design
is reliable. Your plan should indicate how to form research questions to ensure
the standard of results.
• Validity: There are multiple measuring tools available. However, the only
correct measuring tools are those which help a researcher in gauging results
according to the objective of the research. The questionnaire developed from
this design will then be valid.
• Generalization: The outcome of your design should apply to a population and
not just a restricted sample. A generalized method implies that your survey can
be conducted on any part of a population with similar accuracy.
The above factors affect how respondents answer the research questions, so they
should balance all the above characteristics in a good design.
Exploratory Research
Consider a scenario where a juice bar owner feels that increasing the variety of juices
will enable an increase in customers. However, he is not sure and needs more
information. The owner intends to conduct exploratory research to find out; hence, he
decides to do exploratory research to find out if expanding their juices selection will
enable him to get more customers or if there is a better idea.
Content Index
While it may sound difficult to research something that has very little information
about it, there are several methods which can help a researcher figure out the best
research design, data collection methods and choice of subjects. There are two ways
in which research can be conducted namely primary and secondary.. Under these two
types, there are multiple methods which can used by a researcher. The data gathered
from these research can be qualitative or quantitative. Some of the most widely
used research designs include the following:
Primary research is information gathered directly from the subject. It can be through
a group of people or even an individual. Such a research can be carried out directly
by the researcher himself or can employ a third party to conduct it on their behalf.
Primary research is specifically carried out to explore a certain problem which
requires an in-depth study.
For example: A survey is sent to a given set of audience to understand their opinions
about the size of mobile phones when they purchase one. Based on such information
organization can dig deeper into the topic and make business related decision.
• Interviews: While you may get a lot of information from public sources, but
sometimes an in person interview can give in-depth information on the subject
being studied. Such a research is a qualitative research method. An interview
with a subject matter expert can give you meaningful insights that a generalized
public source won’t be able to provide. Interviews are carried out in person or
on telephone which have open-ended questions to get meaningful information
about the topic.
For example: An interview with an employee can give you more insights to find out
the degree of job satisfaction, or an interview with a subject matter expert of quantum
theory can give you in-depth information on that topic.
• Focus groups: Focus group is yet another widely used method in exploratory
research. In such a method a group of people is chosen and are allowed to
express their insights on the topic that is being studied. Although, it is important
to make sure that while choosing the individuals in a focus group they should
have a common background and have comparable experiences.
For example: A focus group helps a research identify the opinions of consumers if
they were to buy a phone. Such a research can help the researcher understand what
the consumer value while buying a phone. It may be screen size, brand value or even
the dimensions. Based on which the organization can understand what are consumer
buying attitudes, consumer opinions, etc.
• Online research: In today’s world, this is one of the fastest way to gather
information on any topic. A lot of data is readily available on the internet and
the researcher can download it whenever he needs it. An important aspect to be
noted for such a research is the genuineness and authenticity of the source
websites that the researcher is gathering the information from.
For example: A researcher needs to find out what is the percentage of people that
prefer a specific brand phone. The researcher just enters the information he needs in a
search engine and gets multiple links with related information and statistics.
However, a few things have to be kept in mind while researching from these sources.
Government agencies have authentic information but sometimes may come with a
nominal cost. Also, research from educational institutions is generally overlooked,
but in fact educational institutions carry out more number of research than any other
entities.
For example: A company has low sales. It can be easily explored from available
statistics and market literature if the problem is market related or organization related
or if the topic being studied is regarding financial situation of the country, then
research data can be accessed through government documents or commercial sources.
• Case study research: Case study research can help a researcher with finding
more information through carefully analyzing existing cases which have gone
through a similar problem. Such analysis are very important and critical
especially in today’s business world. The researcher just needs to make sure he
analyses the case carefully in regards to all the variables present in the previous
case against his own case. It is very commonly used by business organizations
or social sciences sector or even in the health sector.
For example: A particular orthopedic surgeon has the highest success rate for
performing knee surgeries. A lot of other hospitals or doctors have taken up this case
to understand and benchmark the method in which this surgeon does the procedure to
increase their success rate.
• Identify the problem: A researcher identifies the subject of research and the
problem is addressed by carrying out multiple methods to answer the questions.
• Create the hypothesis: When the researcher has found out that there are no
prior studies and the problem is not precisely resolved, the researcher will create
a hypothesis based on the questions obtained while identifying the problem.
• Further research: Once the data has been obtained, the researcher will
continue his study through descriptive investigation. Qualitative methods are
used to further study the subject in detail and find out if the information is true
or not.
• The researcher has a lot of flexibility and can adapt to changes as the research
progresses.
• It is usually low cost.
• It helps lay the foundation of a research, which can lead to further research.
• It enables the researcher understand at an early stage, if the topic is worth
investing the time and resources and if it is worth pursuing.
• It can assist other researchers to find out possible causes for the problem, which
can be further studied in detail to find out, which of them is the most likely
cause for the problem.
• Even though it can point you in the right direction towards what is the answer, it
is usually inconclusive.
• The main disadvantage of exploratory research is that they provide qualitative
data. Interpretation of such information can be judgmental and biased.
• Most of the times, exploratory research involves a smaller sample, hence the
results cannot be accurately interpreted for a generalized population.
• Many a times, if the data is being collected through secondary research, then
there is a chance of that data being old and is not updated.
Descriptive Research
Descriptive research aims to accurately and systematically describe a population,
situation or phenomenon. It can answer what, where, when and how questions, but
not why questions.
It is useful when not much is known yet about the topic or problem. Before you can
research why something happens, you need to understand how, when and where it
happens.
Surveys
Survey research allows you to gather large volumes of data that can be analyzed for
frequencies, averages and patterns. Common uses of surveys include:
Observations
Observations allow you to gather data on behaviours and phenomena without having
to rely on the honesty and accuracy of respondents. This method is often used by
psychological, social and market researchers to understand how people act in real-life
situations.
Observation of physical entities and phenomena is also an important part of research
in the natural sciences. Before you can develop testable hypotheses, models or
theories, it’s necessary to observe and systematically describe the subject under
investigation.
Case studies
A case study can be used to describe the characteristics of a specific subject (such as
a person, group, event or organization). Instead of gathering a large volume of data to
identify patterns across time or location, case studies gather detailed data to identify
the characteristics of a narrowly defined subject.
Rather than aiming to describe generalizable facts, case studies often focus on
unusual or interesting cases that challenge assumptions, add complexity, or reveal
something new about a research problem.
Diagnostic Design:
Diagnostic refers to scientific differentiation among various conditions
• clinical practice.
Its objective includes screening and classification personality description,
• problem solving strategy of the helper faced with problems and crises on the job.
It consists of:
• requires an expertness and sensibility that might well be called an art. It requires a
skilful interviewer to obtain a maximum amount of useful data.
• The interviewers must try to establish a feeling of mental trust with the
• persons being interviewed being careful not to ask questions that might alarm them.
His attitude should convey a spirit of cooperation, acceptance and empathy
• corroborate findings of the other two areas of assessment. Diagnostic tests provide
a microscopic view of the component elements of
• some area of performance. Such tests enable the diagnostician to analyze the
individual's functioning within specific subskill areas and supply direction for
remediation.
Formal standardised tests are useful in testing particular aspects of behaviour
• however, it is important to know the limitations of the test and to use the
information in proper perspectives. Collaborative Diagnosis Many of the problems
which have been attached through field experiment
• have infact been problems related to highly developed professional skills in such
areas as human relations training, therapy, community organizations etc. For such
problems there are already available highly skilled professional
• people who may be called upon to conduct the manipulation. Once the basic role
relationships are worked out, there should be a
• collaborative diagnosis of the situation by the researcher and at least some part of
the client organization. 8 The purpose of this diagnosis is to assess various factors
that will be involved
• in executing the research design the resistance that may be encountered, the
dynamics of the situation, in regard to the problem of bring about a change. The
ways of going about such a collaborative diagnosis will varies
• tremendously depending upon the problem to be studied, the setting, etc., all
relevant experiences and information should be used, but whatever the source, a
useful diagnosis will have to be formulated in theoretical terms. The next step should
be joint planning based on an adequate diagnosis of the
• action that must be taken in order to manipulate the independent variable and to
control other possibly confounding factors. Among the descriptive and diagnostic
designs, descriptive research is
• with the relationships between variables, the testing of hypothesis, and the
development of generalization, principles, or theories that have universal validity. It
concerns with functional relationships.
Experimental research
Experimental research is research conducted with a scientific approach using two sets
of variables. The first set acts as a constant, which you use to measure the differences
of the second set. Quantitative research methods, for example, are experimental.
If you don’t have enough data to support your decisions, you must first determine the
facts. Experimental research gathers the data necessary to help you make better
decisions.
The classic experimental design definition is, “The methods used to collect data in
experimental studies.”
2. Systematic sampling
Individuals are selected at regular intervals from the sampling frame. The intervals
are chosen to ensure an adequate sample size. If you need a sample size n from a
population of size x, you should select every x/nth individual for the sample. For
example, if you wanted a sample size of 100 from a population of 1000, select every
1000/100 = 10th member of the sampling frame.
Systematic sampling is often more convenient than simple random sampling, and it is
easy to administer. However, it may also lead to bias, for example if there are
underlying patterns in the order of the individuals in the sampling frame, such that the
sampling technique coincides with the periodicity of the underlying pattern. As a
hypothetical example, if a group of students were being sampled to gain their
opinions on college facilities, but the Student Record Department’s central list of all
students was arranged such that the sex of students alternated between male and
female, choosing an even interval (e.g. every 20th student) would result in a sample of
all males or all females. Whilst in this example the bias is obvious and should be
easily corrected, this may not always be the case.
3. Stratified sampling
In this method, the population is first divided into subgroups (or strata) who all share
a similar characteristic. It is used when we might reasonably expect the measurement
of interest to vary between the different subgroups, and we want to ensure
representation from all the subgroups. For example, in a study of stroke outcomes, we
may stratify the population by sex, to ensure equal representation of men and women.
The study sample is then obtained by taking equal sample sizes from each stratum. In
stratified sampling, it may also be appropriate to choose non-equal sample sizes from
each stratum. For example, in a study of the health outcomes of nursing staff in a
county, if there are three hospitals each with different numbers of nursing staff
(hospital A has 500 nurses, hospital B has 1000 and hospital C has 2000), then it
would be appropriate to choose the sample numbers from each
hospital proportionally (e.g. 10 from hospital A, 20 from hospital B and 40 from
hospital C). This ensures a more realistic and accurate estimation of the health
outcomes of nurses across the county, whereas simple random sampling would over-
represent nurses from hospitals A and B. The fact that the sample was stratified
should be taken into account at the analysis stage.
Stratified sampling improves the accuracy and representativeness of the results by
reducing sampling bias. However, it requires knowledge of the appropriate
characteristics of the sampling frame (the details of which are not always available),
and it can be difficult to decide which characteristic(s) to stratify by.
4. Clustered sampling
In a clustered sample, subgroups of the population are used as the sampling unit,
rather than individuals. The population is divided into subgroups, known as clusters,
which are randomly selected to be included in the study. Clusters are usually already
defined, for example individual GP practices or towns could be identified as clusters.
In single-stage cluster sampling, all members of the chosen clusters are then included
in the study. In two-stage cluster sampling, a selection of individuals from each
cluster is then randomly selected for inclusion. Clustering should be taken into
account in the analysis. The General Household survey, which is undertaken annually
in England, is a good example of a (one-stage) cluster sample. All members of the
selected households (clusters) are included in the survey.1
Cluster sampling can be more efficient that simple random sampling, especially
where a study takes place over a wide geographical region. For instance, it is easier to
contact lots of individuals in a few GP practices than a few individuals in many
different GP practices. Disadvantages include an increased risk of bias, if the chosen
clusters are not representative of the population, resulting in an increased sampling
error.
2. Quota sampling
This method of sampling is often used by market researchers. Interviewers are given
a quota of subjects of a specified type to attempt to recruit. For example, an
interviewer might be told to go out and select 20 adult men, 20 adult women, 10
teenage girls and 10 teenage boys so that they could interview them about their
television viewing. Ideally the quotas chosen would proportionally represent the
characteristics of the underlying population.
Whilst this has the advantage of being relatively straightforward and potentially
representative, the chosen sample may not be representative of other characteristics
that weren’t considered (a consequence of the non-random nature of sampling). 2
4. Snowball sampling
This method is commonly used in social sciences when investigating hard-to-reach
groups. Existing subjects are asked to nominate further subjects known to them, so
the sample increases in size like a rolling snowball. For example, when carrying out a
survey of risk behaviours amongst intravenous drug users, participants may be asked
to nominate other users to be interviewed.
Snowball sampling can be effective when a sampling frame is difficult to identify.
However, by selecting friends and acquaintances of subjects already investigated,
there is a significant risk of selection bias (choosing a large number of people with
similar characteristics or views to the initial individual identified).
Bias in sampling
There are five important potential sources of bias that should be considered when
selecting a sample, irrespective of the method used. Sampling bias may be introduced
when:1
1. Any pre-agreed sampling rules are deviated from
2. People in hard-to-reach groups are omitted
3. Selected individuals are replaced with others, for example if they are difficult to
contact
4. There are low response rates
5. An out-of-date list is used as the sample frame (for example, if it excludes people
who have recently moved to an area)
Further potential problems with sampling strategies are covered in chapter 8 of this
section (“Sources of variation, its measurement and control”).
What is Scale ?
The process of developing a continuum for placing the measured objects is called
scaling. These measured objects may involve interests, opinions. attitude and other
responses of the individuals. Thus, the next operation involved in measurement is the
scaling process. In this, different qualitative aspects are associated with quantitative
components.
Scale of Measurement in Research Methodology :
Measurement is followed by scaling procedure in a research activity. Researchers
always face the problem of how to measure or quantify abstract concepts and how to
relate one individual's response to that with another response. Hence, the problem is
two-fold. First is, how to quantify a response in absolute terms, and the second is,
how to relate it with other responses. This problem is resolved by scaling process, or
assigning the numbers or values to responses and positioning them on a common
scale. Scaling can then be defined as, "creating a continuum with two extreme limits
and few immediate values between them". Hence, scaling is the process or procedure
of quantifying and measuring abstract concepts like attitudes, opinions, emotions, etc.
Scaling also quantitatively determines the position of an individual in a range of
values.
Definition of Scaling
In a scale, the values are progressively placed according their value or magnitude in
such a manner that moving forward in a scale will depict an item to be always higher
than the immediate lower one.
Process of Scaling
Among all the commonly used direct techniques to measure attitudes, like word-
association tests, sentence-completion tests, story-telling. performance of 'objective'
tasks. physiological reactions, etc., the 'self-report method' is the most popular. In this
method, the respondent is asked to give his opinion about a particular object freely.
Described below is a step-by-step guide on how to construct a self-report :
1) Arbitrary Scales :
In this technique, large number of items or categories, which are able to represent the
concept under study are collected and (then) measured in terms of same
characteristic. Subjective selection of the researcher plays an important role here.
Firstly, few related statements or items are selected by the researcher, and then these
statements are filtered for being included in the measuring instrument. A list of
filtered statements is presented to the respondents and they are required to tick the
most suitable one.
2) Cumulative Scales :
A series of statements is used in developing scale for the respondents in cumulative
scales. These are also called 'Louis Guttman's scales. These scales require
respondents to show their agreement or disagreement over the statements. These
scales are called cumulative scales because a cumulative series is formed by different
statements here. For example, the confirmation of an extreme position item should
also lead to the confirmation of less extreme position items.
3) Consensus Scaling :
This method of scaling was given by L.L. Thurstone. In this, a panel of judges uses
different attributes like attitude. relevance and ambiguity factor so as to select items
or categories. Generally, differential scales are created by this method, which focus
on measuring attitudes of respondents concerning specific issues like religion, war,
etc. These are the most suitable scales for recording only a single attitude of the given
concept. A high level of cost and effort is needed to create such scales. Judges' own
attitudes may affect their functioning of assigning values to different statements.
4) Item Analysis :
In this type of scale construction technique, a group of respondents is given a test
created by selecting different individual items. Following steps are involved in this
technique :
• Firstly, large number of items expressing their positive or negative aspects is
selected.
• Then, a group of respondents assigns scores to them.
• Then for each respondent, total scores for all items are calculated. Then using
these scores, they are divided into four constructs. Two middle constructs are
then eliminated from the calculation.
• Then average score is calculated separately for each item.
5) Factor Scales :
Several different techniques are involved in constructing such scales. The purpose of
using different scales is to study multi-dimensional items, to find the relations
between different dimensions and to identify any other dimensions of the items. In
the end, limited set of factors are available each having interrelated dimensions.
Scales developed through this technique are as follows :
i) Q-Sort Technique :
In this, the respondent is asked to sort the different given statements into prefixed
categories.
v) Standardized Instruments :
In this, an available measuring instrument is selected for data collection. No new
measuring instruments are developed. The opinion of expert is taken for selecting
such standardized instruments.
Any scale must satisfy the following criteria to prove its efficacy :
1) Validity :
Validity is the most critical aspect. It establishes or indicates the extent to which the
scale does what it is supposed to do. Does it measure what it was intended to
measure? In other words, whether or not the test has been useful or worthwhile.
2) Reliability :
This indicates whether or not the findings of a measuring instrument can be relied
upon. To know this, the researcher needs to answer some questions, such as, Are the
outcomes accurate? Can they be replicated? Are the findings consistent?, etc. While
reliability contributes to validity, it is not true the other way round. A reliable
instrument may or may not be necessarily valid.
3) Practicality :
Practicality refers to ease and economy, of constructing the measuring instrument as
well as administration and interpretation of the test outcome. It measures the
achievability, and practicality of an instrument. The benefits or results should justify
the costs.
4) Sensitivity :
This refers to how well the test is standardized. It tests how much an instrument is
able to measure accurately. For example, a test which requires respondents to merely
say 'yes' or 'no', may not be very sensitive. On the other hand, if the respondent is
asked to rule on a '5' or "7"-point scale, the scale may be considered to be highly
sensitive. It may however be noted that all scales need not be highly sensitive. It
largely depends upon the requirement of the test.
5) Generalisability :
This refers to whether or not one can generalize the findings, i.e., whether or not the
findings can be applied to both similar and different situations. Whether or not the
sample or respondents selected can be said to represent larger population. For
example, exit polls can be said to have a high degree of generalisability.
6) Economy :
This factor emphasizes on the aspect that whether the instrument, is economical to be
constructed and conducted. Tests are expensive to develop and administer. Therefore,
the results or benefits of the tests must justify or warrant the costs.
7) Convenience :
A measuring instrument is convenient, if it is easy to conduct. The instruments that
provide guidelines to use are much more convenient than those instruments that do
not include this feature. Generally, it is considered that the requirement for
convenience increases with the level of complexity of a measuring instrument.
Importance of Scaling
In business or management research, scaling is very crucial for the research process.
It helps in measuring. and analyzing attitudes of different individuals. The exact
behavior of an individual is reflected by such attitude analysis. Number of attitude
measuring scales has been developed by researchers. For example, in order to
measure the attitude of an individual about a particular tourist place, product or
election candidate, i.e., if he visits, buys or votes, respectively, a suitable scale is
developed. The different facts describing the importance of scaling are as follows :
1) Attitude Scoring :
Scaling is particularly used for altitude scoring of an individual. In scaling, with the
help of an individual's responses a particular number or point is selected on given
scale, which represents the attitude of that individual.
2) Broad Application :
Different management research processes as well as scientific inventions use scaling
as their crucial element. Data collection methods like interviews, observations,
surveys, etc,, also use scaling for attitude measurement of respondents.
3) Hypothesis Testing :
Scaling is also functional in hypothesis testing. Without effective measure of attitudes
of different respondents, it is not easy to test hypothesis about the population.
4) Dimension Checking :
Scaling is useful in determining the dimensional aspects of different quantitative
concepts or items. It helps in checking whether a particular item is single dimensional
or multi-dimensional.
5) Others :
• It is an essential element of the exploratory research.
• It is used to check whether a set of questions is measuring single aspect or
multiple aspects.
Levels of Measurements
There are four different scales of measurement. The data can be defined as being one
of the four scales. The four types of scales are:
• Nominal Scale
• Ordinal Scale
• Interval Scale
• Ratio Scale
Nominal Scale
A nominal scale is the 1st level of measurement scale in which the numbers serve as
“tags” or “labels” to classify or identify the objects. A nominal scale usually deals
with the non-numeric variables or the numbers that do not have any value.
Characteristics of Nominal Scale
• The numbers don’t define the object characteristics. The only permissible aspect
of numbers in the nominal scale is “counting.”
Example:
An example of a nominal scale measurement is given below:
What is your gender?
M- Male
F- Female
Here, the variables are used as tags, and the answer to this question should be either
M or F.
Ordinal Scale
The ordinal scale is the 2nd level of measurement that reports the ordering and ranking
of data without establishing the degree of variation between them. Ordinal represents
the “order.” Ordinal data is known as qualitative data or categorical data. It can be
grouped, named and also ranked.
Characteristics of the Ordinal Scale
• Along with the information provided by the nominal scale, ordinal scales give
the rankings of those variables
• The interval properties are not known
• The surveyors can quickly analyse the degree of agreement concerning the
identified order of variables
Example:
• Often
• Not often
• Not at all
• Agree
• Neutral
• Disagree
• Totally disagree
Interval Scale
The interval scale is the 3rd level of measurement scale. It is defined as a quantitative
measurement scale in which the difference between the two variables is meaningful.
In other words, the variables are measured in an exact manner, not as in a relative
way in which the presence of zero is arbitrary.
Characteristics of Interval Scale:
• The interval scale is quantitative as it can quantify the difference between the
values
• It allows calculating the mean and median of the variables
• To understand the difference between the variables, you can subtract the values
between the variables
• The interval scale is the preferred scale in Statistics as it helps to assign any
numerical values to arbitrary assessment such as feelings, calendar types, etc.
Example:
• Likert Scale
• Net Promoter Score (NPS)
• Bipolar Matrix Table
Ratio Scale
The ratio scale is the 4th level of measurement scale, which is quantitative. It is a type
of variable measurement scale. It allows researchers to compare the differences or
intervals. The ratio scale has a unique feature. It possesses the character of the origin
or zero points.
Characteristics of Ratio Scale:
• Ratio scale has a feature of absolute zero
• It doesn’t have negative numbers, because of its zero-point feature
• 76 – 85 kgs
• 86 – 95 kgs
Unit-3
What is Data?
Data is a collection of facts, figures, objects, symbols, and events gathered from
different sources. Organizations collect data with various data collection methods
to make better decisions. Without data, it would be difficult for organizations to
make appropriate decisions, so data is collected from different audiences at various
points in time.
Although data is a valuable asset for every organization, it does not serve any purpose
until analyzed or processed to get the desired results.
Data collection methods are techniques and procedures used to gather information for
research purposes. These methods can range from simple self-reported surveys to
more complex experiments and can involve either quantitative or qualitative
approaches to data gathering.
Data collection methods play a crucial role in the research process as they determine
the quality and accuracy of the data collected. Here are some mejor importance of
data collection methods.
The importance of data collection methods cannot be overstated, as it plays a key role
in the overall success and validity of the research study.
The choice of data collection method depends on the research question being
addressed, the type of data needed, and the resources and time available. You can
categorize data collection methods into primary methods of data collection and
secondary methods of data collection.
Primary data is collected from first-hand experience and is not used in the past. The
data gathered by primary data collection methods are specific to the research’s motive
and highly accurate.
Primary data collection methods can be divided into two categories: quantitative
methods and qualitative methods.
Quantitative Methods:
Quantitative techniques for market research and demand forecasting usually use
statistical tools. In these techniques, demand is forecasted based on historical data.
These methods of primary data collection are generally used to make long-term
forecasts. Statistical methods are highly reliable as subjectivity is minimal in these
methods.
Time Series Analysis: The term time series refers to a sequential order of values of a
variable, known as a trend, at equal time intervals. Using patterns, an organization
can predict the demand for its products and services for the projected time.
Smoothing Techniques: In cases where the time series lacks significant trends,
smoothing techniques can be used. They eliminate a random variation from the
historical demand. It helps in identifying patterns and demand levels to estimate
future demand. The most common methods used in smoothing demand forecasting
techniques are the simple moving average method and the weighted moving average
method.
Barometric Method: Also known as the leading indicators approach, researchers use
this method to speculate future trends based on current developments. When the past
events are considered to predict future events, they act as leading indicators.
Qualitative Methods:
Qualitative methods are especially useful in situations when historical data is not
available. Or there is no need of numbers or mathematical calculations.
Qualitative research is closely associated with words, sounds, feeling, emotions,
colors, and other elements that are non-quantifiable. These techniques are based on
experience, judgment, intuition, conjecture, emotion, etc.
Quantitative methods do not provide the motive behind participants’ responses, often
don’t reach underrepresented populations, and span long periods to collect the data.
Hence, it is best to combine quantitative methods with qualitative methods.
Surveys: Surveys are used to collect data from the target audience and gather insights
into their preferences, opinions, choices, and feedback related to their products and
services. Most survey software often has a wide range of question types to select.
You can also use a ready-made survey template to save time and effort. Online
surveys can be customized as per the business’s brand by changing the theme, logo,
etc. They can be distributed through several distribution channels such as email,
website, offline app, QR code, social media, etc. Depending on the type and source of
your audience, you can select the channel.
Once the data is collected, survey software can generate various reports and run
analytics algorithms to discover hidden insights. A survey dashboard can give you
statistics related to response rate, completion rate, filters based on demographics,
export and sharing options, etc. Integrating survey builder with third-party apps can
maximize the effort spent on online real-time data collection.
Polls: Polls comprise one single or multiple-choice question. You can go for polls
when it is required to have a quick pulse of the audience’s sentiments. Because they
are short in length, it is easier to get responses from people.
Like surveys, online polls can also be embedded into various platforms. Once the
respondents answer the question, they can also be shown how they stand compared to
others’ responses.
Focus Groups: In a focus group, a small group of people, around 8-10 members,
discuss the common areas of the problem. Each individual provides his insights on
the issue concerned. A moderator regulates the discussion among the group members.
At the end of the discussion, the group reaches a consensus.
Secondary data is the data that has been used in the past. The researcher can
obtain data from the data sources, both internal and external, to the organization.
• Government reports
• Press releases
• Business journals
• Libraries
• Internet
The secondary data collection methods can also involve quantitative and qualitative
techniques. Secondary data is easily available and hence, less time-consuming and
expensive than primary data. However, with the secondary data collection methods,
the authenticity of the data gathered cannot be verified.
The secondary data collection methods can also involve quantitative and qualitative
techniques. Secondary data is easily available, less time-consuming, and more
expensive than primary data. However, with the secondary data collection methods,
the authenticity of the data gathered cannot be verified.
Regardless of the data collection method of your choice, there must be direct
communication with decision-makers so that they understand and commit to acting
according to the results.
For this reason, we must pay special attention to the analysis and presentation of the
information obtained. Remember that these data must be useful and functional to us,
so the data collection method used has much to do with it.
Conclusion
The conclusion you obtain from your investigation will set the course of the
company’s decision-making, so present your report clearly, and list the steps you
followed to obtain those results.
Make sure that whoever will take the corresponding actions understands the
importance of the information collected and that it gives them the solutions they
expect.
Remember that at QuestionPro, we can help you collect data easily and efficiently.
Request a demo and learn about all the tools we have for you.
The Data Analysis Process is nothing but gathering information by using a proper
application or tool which allows you to explore the data and find a pattern in it. Based
on that information and data, you can make decisions, or you can get ultimate
conclusions.
Data Collection •
Data Cleaning •
Data Analysis •
Data Interpretation•
Data Visualization•
Data Interpretation
Data interpretation is the process of reviewing data through some predefined processes
which will help assign some meaning to the data and arrive at a relevant conclusion. It
involves taking the result of data analysis. Data analysis is the process of ordering,
categorizing, manipulating, and summarizing data to obtain answers to research
questions. It is usually the first step taken towards data interpretation.
It is evident that the interpretation of data is very important, and as such needs to be
done properly. Therefore, researchers have identified some data interpretation methods
to aid this process.
Data interpretation methods are how analysts help people make sense of numerical data
that has been collected, analyzed and presented. Data, when collected in raw form, may
be difficult for the layman to understand, which is why analysts need to break down
the information gathered so that others can make sense of it. For example, when
founders are pitching to potential investors, they must interpret data (e.g. market size,
growth rate, etc.) for better understanding. There are 2 main methods in which this can
be done, namely; quantitative methods and qualitative methods.
Qualitative Data Interpretation Method
The qualitative data interpretation method is used to analyze qualitative data, which is
also known as categorical data. This method uses texts, rather than numbers or patterns
to describe data. Qualitative data is usually gathered using a wide variety of person-
toperson techniques, which may be difficult to analyze compared to the quantitative
research method. Unlike the quantitative data which can be analyzed directly after it
has been collected and sorted, qualitative data needs to first be coded into numbers
before it can be analyzed. This is because texts are usually cumbersome, and will take
more time and result in a lot of errors if analyzed in its original state. Coding done by
the analyst should also be documented so that it can be reused by others and also
analyzed. There are 2 main types of qualitative data, namely; nominal and ordinal data.
These 2 data types are both interpreted using the same method, but ordinal data
interpretation is quite easier than that of nominal data. In most cases, ordinal data is
usually labelled with numbers during the process of data collection, and coding may
not be required. This is different from nominal data that still needs to be coded for
proper interpretation.
The quantitative data interpretation method is used to analyze quantitative data, which
is also known as numerical data. This data type contains numbers and is therefore
analyzed with the use of numbers and not texts. Quantitative data are of 2 main types,
namely; discrete and continuous data. Continuous data is further divided into interval
data and ratio data, with all the data types being numeric. Due to its natural existence
as a number, analysts do not need to employ the coding technique on quantitative data
before it is analyzed. The process of analyzing quantitative data involves statistical
modelling techniques such as standard deviation, mean and median.Some of the
statistical methods used in analyzing quantitative data are highlighted below:
Mean•
The mean is a numerical average for a set of data and is calculated by dividing the sum
of the values by the number of values in a dataset. It is used to get an estimate of a large
population from the dataset obtained from a sample of the population.
Standard deviation•
This technique is used to measure how well the responses align with or deviates from
the mean. It describes the degree of consistency within the responses; together with the
mean, it provides insight into data sets.
Frequency distribution•
This technique is used to assess the demography of the respondents or the number of
times a particular response appears in research. It is extremely keen on determining the
degree of intersection between data points.
It helps to make informed decisions and not just through guessing or predictions. •
It is cost-efficient •
The insights obtained can be used to set and identify trends in data.•
Conclusion
Data interpretation and analysis is an important aspect of working with data sets in any
field or research and statistics. They both go hand in hand, as the process of data
interpretation involves the analysis of data. Data interpretation is very important, as it
helps to acquire useful information from a pool of irrelevant ones while making
informed decisions. It is found useful for individuals, businesses, and researchers.
Hypothesis Testing
Hypothesis testing is one of the most important concepts in Statistics which is heavily
In hypothesis testing, Statistical tests are used to check whether the null hypothesis is
So, In this article, we will be discussing the statistical test for hypothesis testing
The basic principle behind the parametric tests is that we have a fixed set of parameters
that are used to determine a probabilistic model that may be used in Machine Learning
as well.
Parametric tests are those tests for which we have prior knowledge of the population
distribution which is possible with the help of the Central Limit Theorem.
• Mean
• Standard Deviation
population assumptions. There are many parametric tests available from which some
• To find the confidence interval for the population means with the help of known
standard deviation.
• To determine the confidence interval for population means along with the unknown
standard deviation.
• To find the confidence interval for the difference of two means, with an unknown value
of standard deviation.
Non-parametric Tests
In Non-Parametric tests, we don’t make any assumption about the parameters for the
given population or the population we are studying. In fact, these tests don’t depend on
the population.
Hence, there is no fixed set of parameters is available, and also there is no distribution
This is also the reason that nonparametric tests are also referred to as distribution-free
tests.
In modern days, Non-parametric tests are gaining popularity and an impact of influence
• The main reason is that there is no need to be mannered while using parametric tests.
• The second reason is that we do not require to make assumptions about the population
• Most of the nonparametric tests available are very easy to apply and to understand also
T-Test
the sample size is small (i.e, less than 30) and when the population standard deviation
is not available.
A T-test can be a:
One Sample T-test: To compare a sample mean with that of the population mean.
where,
where,
Conclusion:
• If the value of the test statistic is greater than the table value -> Rejects the null
hypothesis.
• If the value of the test statistic is less than the table value -> Do not reject the null
hypothesis.
Z-Test
is known and the sample size is large (i.e, greater than 30).
One Sample Z-test: To compare a sample mean with that of the population mean.
F-Test
variance.
5. It is calculated as:
F = s12/s22
6. By changing the variance in the ratio, F-test has become a very flexible test. It can
3. It is used to test the significance of the differences in the mean values among more
4. It uses F-test to statistically test the equality of means and the relative variance
between them.
Chi-Square Test
theoretically.
frequencies.
6. If there is no difference between the expected and observed frequencies, then the
7. It is also known as the “Goodness of fit test” which determines whether a particular
8. It is calculated as:
• No one of the groups should contain very few items, say less than 10.
• The reasonably large overall number of items. Normally, it should be at least 50,
11. Chi-square as a parametric test is used as a test for population variance based on
sample variance.
12. If we take each one of a collection of sample variances, divide them by the known
population variance and multiply these quotients by (n-1), where n means the number
2. This test is used to investigate whether two independent samples were selected from
3. It is a true non-parametric counterpart of the T-test and gives the most accurate
estimates of significance especially when sample sizes are small and the population is
4. It is based on the comparison of every observation in the first sample with every
U1 = R1 – n1(n1+1)/2
where n1 is the sample size for sample 1, and R1 is the sum of ranks in Sample 1.
U2 = R2 – n2(n2+1)/2
When consulting the significance tables, the smaller values of U 1 and U2 are used. The
U1 + U2 = { R1 – n1(n1+1)/2 } + { R2 – n2(n2+1)/2 }
Knowing that R1+R2 = N(N+1)/2 and N=n1+n2, and doing some algebra, we find that
U1 + U2 = n1*n2
Kruskal-Wallis H-test
2. This test is used for comparing two or more independent samples of equal or
4. One-Way ANOVA is the parametric equivalent of this test. And that’s why it is also
Another popular nonparametric test for matched or paired data is called the Wilcoxon
Signed Rank Test. Like the Sign Test, it is based on difference scores, but in addition
to analyzing the signs of the differences, it also takes into account the magnitude of
the observed differences.
Let's use the Wilcoxon Signed Rank Test to re-analyze the data in Example 4 on page
5 of this module. Recall that this study assessed the effectiveness of a new drug
designed to reduce repetitive behaviors in children affected with autism. A total of 8
children with autism enroll in the study and the amount of time that each child is
engaged in repetitive behavior during three hour observation periods are measured
both before treatment and then again after taking the new medication for a period of 1
week. The data are shown below.
The next step is to rank the difference scores. We first order the absolute values of the
difference scores and assign rank from 1 through n to the smallest through largest
absolute values of the difference scores, and assign the mean rank when there are ties
in the absolute values of the difference scores.
The final step is to attach the signs ("+" or "-") of the observed differences to each
rank as shown below.
Similar to the Sign Test, hypotheses for the Wilcoxon Signed Rank Test concern the
population median of the difference scores. The research hypothesis can be one- or
two-sided. Here we consider a one-sided test.
The test statistic for the Wilcoxon Signed Rank Test is W, defined as the smaller of
W+ (sum of the positive ranks) and W- (sum of the negative ranks). If the null
hypothesis is true, we expect to see similar numbers of lower and higher ranks that
are both positive and negative (i.e., W+ and W- would be similar). If the research
hypothesis is true we expect to see more higher and positive ranks (in this example,
more children with substantial improvement in repetitive behavior after treatment as
compared to before, i.e., W+ much larger than W-).
In this example, W+ = 32 and W- = 4. Recall that the sum of the ranks (ignoring the
signs) will always equal n(n+1)/2. As a check on our assignment of ranks, we have
n(n+1)/2 = 8(9)/2 = 36 which is equal to 32+4. The test statistic is W = 4.
Next we must determine whether the observed test statistic W supports the null or
research hypothesis. This is done following the same approach used in parametric
testing. Specifically, we determine a critical value of W such that if the observed
value of W is less than or equal to the critical value, we reject H0 in favor of H1, and
if the observed value of W exceeds the critical value, we do not reject H 0.
Note that when we analyzed the data previously using the Sign Test, we failed to find
statistical significance. However, when we use the Wilcoxon Signed Rank Test, we
conclude that the treatment result in a statistically significant improvement at α=0.05.
The discrepant results are due to the fact that the Sign Test uses very little
information in the data and is a less powerful test.
Example:
The test statistic for the Wilcoxon Signed Rank Test is W, defined as the smaller of
W+ and W- which are the sums of the positive and negative ranks, respectively.
The critical value of W can be found in the table of critical values. To determine the
appropriate critical value from Table 7 we need sample size (n=15) and our two-sided
level of significance (α=0.05). The critical value for this two-sided test with n=15 and
α=0.05 is 25 and the decision rule is as follows: Reject H0 if W < 25.
Because the before and after systolic blood pressures measures are paired, we
compute difference scores for each patient.
Systolic Blood Pressure Systolic Blood Pressure
Patient Before Exercise Program After Exercise Program (B
1 125 118
2 132 134
3 138 130
4 120 124
5 125 105
6 127 130
7 136 130
8 139 132
9 131 123
10 132 128
11 135 126
12 136 140
13 128 135
14 127 126
15 130 132
The next step is to rank the ordered absolute values of the difference scores using the
approach outlined in Section 10.1. Specifically, we assign ranks from 1 through n to
the smallest through largest absolute values of the difference scores, respectively, and
assign the mean rank when there are ties in the absolute values of the difference
scores.
Ordered Absolute
Observed Differences
Values of Differences
7 1
-2 -2
8 -2
-4 -3
20 -4
-3 -4
6 4
7 6
8 -7
4 7
9 7
-4 8
-7 8
1 9
-2 20
The final step is to attach the signs ("+" or "-") of the observed differences to each
rank as shown below.
In this example, W+ = 89 and W- = 31. Recall that the sum of the ranks (ignoring the
signs) will always equal n(n+1)/2. As a check on our assignment of ranks, we have
n(n+1)/2 = 15(16)/2 = 120 which is equal to 89 + 31. The test statistic is W = 31.
• Step 5. Conclusion.
The modules on hypothesis testing presented techniques for testing the equality of
means in two independent samples. An underlying assumption for appropriate use of
the tests described was that the continuous outcome was approximately normally
distributed or that the samples were sufficiently large (usually n 1> 30 and n2> 30) to
justify their use based on the Central Limit Theorem. When comparing two
independent samples when the outcome is not normally distributed and the samples
are small, a nonparametric test is appropriate.
In contrast, the null and two-sided research hypotheses for the nonparametric test are
stated as follows:
H0: The two populations are equal versus
This test is often performed as a two-sided test and, thus, the research hypothesis
indicates that the populations are not equal as opposed to specifying directionality. A
one-sided research hypothesis is used if interest lies in detecting a positive or
negative shift in one population as compared to the other. The procedure for the test
involves pooling the observations from the two samples into one combined sample,
keeping track of which sample each observation comes from, and then ranking lowest
to highest from 1 to n1+n2, respectively.
Example:
Placebo 7 5 6 4 12
New Drug 3 6 4 2 1
In this example, the outcome is a count and in this sample the data do not follow a
normal distribution.
Note that if the null hypothesis is true (i.e., the two populations are equal), we expect
to see similar numbers of episodes of shortness of breath in each of the two treatment
groups, and we would expect to see some participants reporting few episodes and
some reporting more episodes in each group. This does not appear to be the case with
the observed data. A test of hypothesis is needed to determine whether the observed
data is evidence of a statistically significant difference in populations.
The first step is to assign ranks and to do so we order the data from smallest to
largest. This is done on the combined or total sample (i.e., pooling the data from the
two treatment groups (n=10)), and assigning ranks from 1 to 10, as follows. We also
need to keep track of the group assignments in the total sample.
Total Sample R
(Ordered Smallest to Largest)
Placebo New Drug Placebo New Drug Placebo
7 3 1
5 6 2
6 4 3
4 2 4 4 4.5
12 1 5 6
6 6 7.5
7 9
12 10
Note that the lower ranks (e.g., 1, 2 and 3) are assigned to responses in the new drug
group while the higher ranks (e.g., 9, 10) are assigned to responses in the placebo
group. Again, the goal of the test is to determine whether the observed data support a
difference in the populations of responses. Recall that in parametric tests (discussed
in the modules on hypothesis testing), when comparing means between two groups,
we analyzed the difference in the sample means relative to their variability and
summarized the sample information in a test statistic. A similar approach is employed
here. Specifically, we produce a test statistic based on the ranks.
First, we sum the ranks in each group. In the placebo group, the sum of the ranks is
37; in the new drug group, the sum of the ranks is 18. Recall that the sum of the ranks
will always equal n(n+1)/2. As a check on our assignment of ranks, we have n(n+1)/2
= 10(11)/2=55 which is equal to 37+18 = 55.
For the test, we call the placebo group 1 and the new drug group 2 (assignment of
groups 1 and 2 is arbitrary). We let R1 denote the sum of the ranks in group 1 (i.e.,
R1=37), and R2 denote the sum of the ranks in group 2 (i.e., R2=18). If the null
hypothesis is true (i.e., if the two populations are equal), we expect R1 and R2 to be
similar. In this example, the lower values (lower ranks) are clustered in the new drug
group (group 2), while the higher values (higher ranks) are clustered in the placebo
group (group 1). This is suggestive, but is the observed difference in the sums of the
ranks simply due to chance? To answer this we will compute a test statistic to
summarize the sample information and look up the corresponding value in a
probability distribution.
Test Statistic for the Mann Whitney U Test
The test statistic for the Mann Whitney U Test is denoted U and is the smaller of
U1 and U2, defined below.
where R1 = sum of the ranks for group 1 and R2 = sum of the ranks for group 2.
In our example, U=3. Is this evidence in support of the null or research hypothesis?
Before we address this question, we consider the range of the test statistic U in two
different situations.
Situation #1
Consider the situation where there is complete separation of the groups, supporting
the research hypothesis that the two populations are not equal. If all of the higher
numbers of episodes of shortness of breath (and thus all of the higher ranks) are in the
placebo group, and all of the lower numbers of episodes (and ranks) are in the new
drug group and that there are no ties, then:
and
Situation #2
Consider a second situation where low and high scores are approximately evenly
distributed in the two groups, supporting the null hypothesis that the groups are
equal. If ranks of 2, 4, 6, 8 and 10 are assigned to the numbers of episodes of
shortness of breath reported in the placebo group and ranks of 1, 3, 5, 7 and 9 are
assigned to the numbers of episodes of shortness of breath reported in the new drug
group, then:
and
Thus, smaller values of U support the research hypothesis, and larger values of U
support the null hypothesis.
Key Concept:
For any Mann-Whitney U test, the theoretical range of U is from 0
(complete separation between groups, H0 most likely false and H1 most
likely true) to n1*n2 (little evidence in support of H1).
In every test, we must determine whether the observed U supports the null or research
hypothesis. This is done following the same approach used in parametric testing.
Specifically, we determine a critical value of U such that if the observed value of U is
less than or equal to the critical value, we reject H0 in favor of H1 and if the observed
value of U exceeds the critical value we do not reject H0.
The critical value of U can be found in the table below. To determine the appropriate
critical value we need sample sizes (for Example: n 1=n2=5) and our two-sided level
of significance (α=0.05). For Example 1 the critical value is 2, and the decision rule is
to reject H0 if U < 2. We do not reject H0 because 3 > 2. We do not have statistically
significant evidence at α =0.05, to show that the two populations of numbers of
episodes of shortness of breath are not equal. However, in this example, the failure to
reach statistical significance may be due to low power. The sample data suggest a
difference, but the sample sizes are too small to conclude that there is a statistically
significant difference.
Example:
A new approach to prenatal care is proposed for pregnant women living in a rural
community. The new program involves in-home visits during the course of pregnancy
in addition to the usual or regularly scheduled visits. A pilot randomized trial with 15
pregnant women is designed to evaluate whether women who participate in the
program deliver healthier babies than women receiving usual care. The outcome is
the APGAR score measured 5 minutes after birth. Recall that APGAR scores range
from 0 to 10 with scores of 7 or higher considered normal (healthy), 4-6 low and 0-3
critically low. The data are shown below.
Usual Care 8 7 6 2 5 8
New Program 9 9 7 8 10 9
Because APGAR scores are not normally distributed and the samples are small (n 1=8
and n2=7), we use the Mann Whitney U test. The test statistic is U, the smaller of
where R1 and R2 are the sums of the ranks in groups 1 and 2, respectively.
The first step is to assign ranks of 1 through 15 to the smallest through largest values
in the total sample, as follows:
Total Sample R
(Ordered Smallest to Largest)
Usual Care New Program Usual Care New Program Usual Care
8 9 2 1
7 8 3 2
6 7 5 3
2 8 6 6 4.5
5 10 7 7 7
8 9 7 7
7 6 8 8 10.5
3 8 8 10.5
9
9
10
R1=45.5
Next, we sum the ranks in each group. In the usual care group, the sum of the ranks is
R1=45.5 and in the new program group, the sum of the ranks is R2=74.5. Recall that
the sum of the ranks will always equal n(n+1)/2. As a check on our assignment of
ranks, we have n(n+1)/2 = 15(16)/2=120 which is equal to 45.5+74.5 = 120.
• Step 5. Conclusion:
We reject H0 because 9.5 < 10. We have statistically significant evidence at α =0.05
to show that the populations of APGAR scores are not equal in women receiving
usual prenatal care as compared to the new program of prenatal care.
Example:
A clinical trial is run to assess the effectiveness of a new anti-retroviral therapy for
patients with HIV. Patients are randomized to receive a standard anti-retroviral
therapy (usual care) or the new anti-retroviral therapy and are monitored for 3
months. The primary outcome is viral load which represents the number of HIV
copies per milliliter of blood. A total of 30 participants are randomized and the data
are shown below.
Standard
7500 8000 2000 550 1250 1000 2250 6800 3400 63
Therapy
New Therapy 400 250 800 1400 8000 7400 1020 6000 920 14
Because viral load measures are not normally distributed (with outliers as well as
limits of detection (e.g., "undetectable")), we use the Mann-Whitney U test. The test
statistic is U, the smaller of
where R1 and R2 are the sums of the ranks in groups 1 and 2, respectively.
The critical value can be found in the table of critical values based on sample sizes
(n1=n2=15) and a two-sided level of significance (α=0.05). The critical value 64 and
the decision rule is as follows: Reject H 0 if U < 64.
The first step is to assign ranks of 1 through 30 to the smallest through largest values
in the total sample. Note in the table below, that the "undetectable" measurement is
listed first in the ordered values (smallest) and assigned a rank of 1.
Next, we sum the ranks in each group. In the standard anti-retroviral therapy group,
the sum of the ranks is R1=245; in the new anti-retroviral therapy group, the sum of
the ranks is R2=220. Recall that the sum of the ranks will always equal n(n+1)/2. As a
check on our assignment of ranks, we have n(n+1)/2 = 30(31)/2=465 which is equal
to 245+220 = 465. We now compute U1 and U2, as follows,
We do not reject H0 because 100 > 64. We do not have sufficient evidence to
conclude that the treatment groups differ in viral load.
Unit-4
Multiple Regression
Multiple regression generally explains the relationship between multiple
independent or predictor variables and one dependent or criterion variable. A
dependent variable is modeled as a function of several independent variables with
corresponding coefficients, along with the constant term. Multiple regression
requires two or more predictor variables, and this is why it is called multiple
regression.
Multiple Regression
Simple linear regression is a model that assesses the relationship between a dependent
variable and an independent variable. The simple linear model is expressed using the
following equation:
Y = a + bX + ϵ
Where:
• Y – Dependent variable
• X – Independent (explanatory) variable
• a – Intercept
• b – Slope
• ϵ – Residual (error)
Regression Analysis – Multiple Linear Regression
Multiple linear regression analysis is essentially similar to the simple linear model,
with the exception that multiple independent variables are used in the model. The
mathematical representation of multiple linear regression is:
Where:
• Y – Dependent variable
• X1, X2, X3 – Independent (explanatory) variables
• a – Intercept
• b, c, d – Slopes
• ϵ – Residual (error)
Multiple linear regression follows the same conditions as the simple linear model.
However, since there are several independent variables in multiple linear analysis,
there is another mandatory condition for the model:
Factor Analysis
Factor analysis is a technique that is used to reduce a large number of variables into
fewer numbers of factors. This technique extracts maximum common variance from
all variables and puts them into a common score. As an index of all variables, we can
use this score for further analysis. Factor analysis is part of general linear model
(GLM) and this method also assumes several assumptions: there is linear
relationship, there is no multicollinearity, it includes relevant variables into analysis,
and there is true correlation between variables and factors. Several methods are
available, but principal component analysis is used most commonly.
Types of factoring:
There are different types of methods used to extract the factor from the data set:
4. Maximum likelihood method: This method also works on correlation metric but
it uses maximum likelihood method to factor.
Factor loading:
Factor loading is basically the correlation coefficient for the variable and
factor. Factor loading shows the variance explained by the variable on that particular
factor. In the SEM approach, as a rule of thumb, 0.7 or higher factor loading
represents that the factor extracts sufficient variance from that variable.
Eigenvalues: Eigenvalues is also called characteristic roots. Eigenvalues shows
variance explained by that particular factor out of the total variance. From the
commonality column, we can know how much variance is explained by the first
factor out of the total variance. For example, if our first factor explains 68% variance
out of the total, this means that 32% variance will be explained by the other factor.
Factor score: The factor score is also called the component score. This score is of
all row and columns, which can be used as an index of all variables and can be used
for further analysis. We can standardize this score by multiplying a common
term. With this factor score, whatever analysis we will do, we will assume that all
variables will behave as factor scores and will move.
Criteria for determining the number of factors: According to the Kaiser Criterion,
Eigenvalues is a good criteria for determining a factor. If Eigenvalues is greater than
one, we should consider that a factor and if Eigenvalues is less than one, then we
should not consider that a factor. According to the variance extraction rule, it should
be more than 0.7. If variance is less than 0.7, then we should not consider that a
factor.
Assumptions:
Cluster Analysis
Cluster analysis foundations rely on one of the most fundamental, simple and very
often unnoticed ways (or methods) of understanding and learning, which is grouping
“objects” into “similar” groups. This process includes a number of different
algorithms and methods to make clusters of a similar kind. It is also a part of data
management in statistical analysis.
When we try to group a set of objects that have similar kind of characteristics,
attributes these groups are called clusters. The process is called clustering. It is a
very difficult task to get to know the properties of every individual object instead, it
would be easy to group those similar objects and have a common structure of
properties that the group follows.
What is Cluster Analysis?
Cluster analysis is a multivariate data mining technique whose goal is to groups
objects (eg., products, respondents, or other entities) based on a set of user selected
characteristics or attributes. It is the basic and most important step of data mining and
a common technique for statistical data analysis, and it is used in many fields such as
data compression, machine learning, pattern recognition, information retrieval etc.
Clusters should exhibit high internal homogeneity and high external heterogeneity.
What does this mean?
When plotted geometrically, objects within clusters should be very close together and
clusters will be far apart.
Related Articles:
Centroid-based Clustering
In this type of clustering, clusters are represented by a central entity, which may or
may not be a part of the given data set. K-Means method of clustering is used in this
method, where k are the cluster centers and objects are assigned to the nearest cluster
centres.
Distribution-based Clustering
It is a type of clustering model closely related to statistics based on the modals of
distribution. Objects that belong to the same distribution are put into a single
cluster.This type of clustering can capture some complex properties of objects like
correlation and dependence between attributes.
Density-based Clustering
In this type of clustering, clusters are defined by the areas of density that are higher
than the remaining of the data set. Objects in sparse areas are usually required to
separate clusters.The objects in these sparse points are usually noise and border
points in the graph.The most popular method in this type of clustering is DBSCAN.
To learn more on the cluster and other statistics-related topics, visit BYJU’S.
Applications and Examples
It is the principal job of exploratory data mining, and a common method for statistical
data analysis. It is used in many fields, such as machine learning, image analysis,
pattern recognition, information retrieval, data compression, bioinformatics and
computer graphics.
It can be used to examine patterns of antibiotic resistance, to incorporate
antimicrobial compounds according to their mechanism of activity, to analyse
antibiotics according to their antibacterial action.
Cluster analysis can be a compelling data-mining means for any organization that
wants to recognise discrete groups of customers, sales transactions, or other kinds of
behaviours and things. For example, insurance providing companies use cluster
analysis to identify fraudulent claims and banks apply it for credit scoring.
What Is a Research Report?
A research report is a well-crafted document that outlines the processes, data, and
findings of a systematic investigation. It is an important document that serves as a
first-hand account of the research process, and it is typically considered an objective
and accurate source of information.
So how do you recognize a research report when you see one? Here are some of the
basic features that define a research report.
The research report is classified based on two things; nature of research and target
audience.
Nature of Research
This is the type of report written for qualitative research. It outlines the methods,
processes, and findings of a qualitative method of systematic investigation. In
educational research, a qualitative research report provides an opportunity for one to
apply his or her knowledge and develop skills in planning and executing qualitative
research projects.
In this type of research report, the researcher presents quantitative data to support the
research process and findings. Unlike a qualitative research report that is mainly
descriptive, a quantitative research report works with numbers; that is, it is numerical
in nature.
Target Audience
Also, a research report can be said to be technical or popular based on the target
audience. If you’re dealing with a general audience, you would need to present a
popular research report, and if you’re dealing with a specialized audience, you would
submit a technical report.
A technical research report is a detailed document that you present after carrying out
industry-based research. This report is highly specialized because it provides
information for a technical audience; that is, individuals with above-average
knowledge in the field of study.
Examples of technical research reports include legal and medical research reports.
A popular research report is one for a general audience; that is, for individuals who
do not necessarily have any knowledge in the field of study. A popular research
report aims to make information accessible to everyone.
It is written in very simple language, which makes it easy to understand the findings
and recommendations. Examples of popular research reports are the information
contained in newspapers and magazines.
• Knowledge Transfer: As already stated above, one of the reasons for carrying
out research is to contribute to the existing body of knowledge, and this is made
possible with a research report. A research report serves as a means to
effectively communicate the findings of a systematic investigation to all and
sundry.
• Identification of Knowledge Gaps: With a research report, you’d be able to
identify knowledge gaps for further inquiry. A research report shows what has
been done while hinting at other areas needing systematic investigation.
• In market research, a research report would help you understand the market
needs and peculiarities at a glance.
• A research report allows you to present information in a precise and concise
manner.
• It is time-efficient and practical because, in a research report, you do not have to
spend time detailing the findings of your research work in person. You can
easily send out the report via email and have stakeholders look at it.
A lot of detail goes into writing a research report, and getting familiar with the
different requirements would help you create the ideal research report. A research
report is usually broken down into multiple sections, which allows for a concise
presentation of information.
• Title
This is the title of your systematic investigation. Your title should be concise and
point to the aims, objectives, and findings of a research report.
• Table of Contents
This is like a compass that makes it easier for readers to navigate the research report.
• Abstract
An abstract is always brief; typically 100-150 words and goes straight to the point.
The focus of your research abstract should be the 5Ws and 1H format – What, Where,
Why, When, Who and How.
• Introduction
Here, the researcher highlights the aims and objectives of the systematic investigation
as well as the problem which the systematic investigation sets out to solve. When
writing the report introduction, it is also essential to indicate whether the purposes of
the research were achieved or would require more work.
In the introduction section, the researcher specifies the research problem and also
outlines the significance of the systematic investigation. Also, the researcher is
expected to outline any jargons and terminologies that are contained in the research.
• Literature Review
• An Account of Investigation
This is a detailed account of the research process, including the methodology, sample,
and research subjects. Here, you are expected to provide in-depth information on the
research process including the data collection and analysis procedures.
• Findings
In this section, you are expected to present the results of the systematic investigation.
• Discussion
This section further explains the findings of the research, earlier outlined. Here, you
are expected to present a justification for each outcome and show whether the results
are in line with your hypotheses or if other research studies have come up with
similar results.
• Conclusions
This is a summary of all the information in the report. It also outlines the significance
of the entire study.
This section contains a list of all the primary and secondary research sources.
As is obtainable when writing an essay, defining the context for your research report
would help you create a detailed yet concise document. This is why you need to
create an outline before writing so that you do not miss out on anything.
The idea of a research report is to present some sort of abridged version of your
systematic investigation. In your report, you should exclude irrelevant information
while highlighting only important data and findings.
• Include Illustrations
Your research report should include illustrations and other visual representations of
your data. Graphs, pie charts, and relevant images lend additional credibility to your
systematic investigation.
A good research report title is brief, precise, and contains keywords from your
research. It should provide a clear idea of your systematic investigation so that
readers can grasp the entire focus of your research from the title.
Every research aims at solving a specific problem or set of problems, and this should
be at the back of your mind when writing your research report. Understanding the
problem would help you to filter the information you have and include only important
data in your report.
This is somewhat similar to the point above because, in some way, the aim of your
research report is intertwined with the objectives of your systematic investigation.
Identifying the primary purpose of writing a research report would help you to
identify and present the required information accordingly.
Knowing your target audience plays a crucial role in data collection for a research
report. If your research report is specifically for an organization, you would want to
present industry-specific information or show how the research findings are relevant
to the work that the company does.
• Create Surveys/Questionnaires
A survey is a research method that is used to gather data from a specific group of
people through a set of questions. It can be either quantitative or qualitative.
You can seamlessly create an online questionnaire for your research on Formplus.
With the multiple sharing options available in the builder, you would be able to
administer your survey to respondents in little or no time.
Formplus also has a report summary tool that you can use to create custom visual
reports for your research.
Step-by-step guide on how to create an online questionnaire using Formplus
In the Formplus builder, you can easily create different online questionnaires for your
research by dragging and dropping preferred fields into your form. To access the
Formplus builder, you will need to create an account on Formplus.
Once you do this, sign in to your account and click on Create new form to begin.
1. Edit Form Title: Click on the field provided to input your form title, for
example, “Research Questionnaire.”
2. Edit Form: Click on the edit icon to edit the form.
3. Add Fields: Drag and drop preferred form fields into your form in the
Formplus builder inputs column. There are several field input options for
questionnaires in the Formplus builder.
4. Edit fields
5. Click on “Save”
6. Form Customization: With the form customization options in the form
builder, you can easily change the outlook of your form and make it more
unique and personalized. Formplus allows you to change your form theme, add
background images, and even change the font according to your needs.
7. Multiple Sharing Options: Formplus offers various form-sharing options,
which enables you to share your questionnaire with respondents easily. You can
use the direct social media sharing buttons to share your form link to your
organization’s social media pages. You can also send out your survey form as
email invitations to your research subjects too. If you wish, you can share your
form’s QR code or embed it on your organization’s website for easy access.
Conclusion
Always remember that a research report is just as important as the actual systematic
investigation because it plays a vital role in communicating research findings to
everyone else. This is why you must take care to create a concise document
summarizing the process of conducting any research.
In this article, we’ve outlined essential tips to help you create a research report. When
writing your report, you should always have the audience at the back of your mind, as
this would set the tone for the document.
Difference Between Reference and Bibliography
BASIS FOR
REFERENCE BIBLIOGRAPHY
COMPARISON
Includes Only in-text citations, that Both in-text citations and other
have been used in the sources, that are used to generate
assignment or project. idea.
Used for Thesis and Dissertation Journal Papers and Research wor
Includes Only in-text citations, that Both in-text citations and other
have been used in the sources, that are used to generate
assignment or project. idea.
Used for Thesis and Dissertation Journal Papers and Research wor
Definition of Reference
Reference can be understood as the act of giving credit to or mentioning the name of,
someone or something. In research methodology, it denotes the items which you have
reviewed and referred to, in the text, in your research work. It is nothing but a way to
acknowledge or indirectly showing gratitude, towards the sources from where the
information is gathered.
While using references, one thing is to be noted that you go for reliable sources only,
because it increases credence and also supports your arguments. It may include,
books, research papers, or articles from magazines, journals, newspapers, etc.,
interview transcripts, internet sources such as websites, blogs, videos watched, and so
forth.
These are used to inform the reader about the sources of direct quotations, tables,
statistics, photos etc. that are included in the research work.
Definition of Bibliography
At the end of the research report, bibliography is added, which contains a list of
books, magazines, journals, websites or other publications which are in some way
relevant to the topic under study, that has been consulted by the researcher during the
research. In finer terms, it comprises of all the references cited in the form of
footnotes and other important works that the author has studied.
Types of Bibliography
Introduction
These guidelines originated as the "Criteria for Evaluating a Bibliography,"
adopted by the Reference Services Division Board in 1971. A revision of those
guidelines under the current name, "Guidelines for the Preparation of a
Bibliography" was approved by the RASD board of directors in 1982 and can be
found in RQ22 (Fall 1982): 31‐32. The RASD Bibliography Committee based
many parts of the 1982 revision on the "Criteria for the Evaluation of Enumerative
Bibliographies" prepared by the Committee on Bibliographical Services for Canada
in 1979."
"The 1992 committee has proposed less extensive revisions than were adopted in
1982 but wants to strengthen wording regarding annotations and multiple points of
access and to make clearer what is meant by standard bibliographic form."
The original Guidelines were adopted in 1971 by the Reference Services Division
Board and revised and updated in 1982 and in 1992 by the Bibliography Committee
of the Collection Development and Evaluation Section of the Reference and User
Services Association. The 2001 revision was prepared by RUSA’s Standards and
Guidelines Committee to reflect technological developments since the wide
dissemination of print or electronic bibliographies and to make more explicit sound
principles involved in the preparation of a bibliography, regardless of its format.
They are intended for use by all in the library community.
Ensure that the bibliography fills a significant need in order to justify its
compilation.
Scope
Methodology
Organization
Principles of organization
Organize the material suitably for both the subject and the targeted users.
Annotations/Notes
Informative- Use informative notes chiefly when the nature or reason for
inclusion of a title is not clear. Limit use of this minimal level of
description to those bibliographies that approach comprehensiveness for
the area they are covering.
5.2 In each case, provide succinct and informative annotations or notes written
on a level suitable for the intended users. If another source is drawn upon for
the annotation, acknowledge the source appropriately.
Bibliographic Form
Accuracy
Cumulation
Distribution