BRM Imp Questions With Answers
BRM Imp Questions With Answers
BRM Imp Questions With Answers
Research Methodology
(2810006)
2. What is a Research Problem? State the main issue which should receive
the attention of the researcher. Give examples to illustrate your answer.
Answer:
Problem Definition is the first and most crucial step in the research process
- Main function is to decide what you want to find out about.
- The way you formulate a problem determines almost every step that follows.
Hypothesis:
a. A statement which is accepted temporarily as true and on this basis the research work
confines,
b. A tentative assumption drawn from knowledge and theory which is used as a
guide in the investigation of other facts and theories that are yet unknown,
c. States what the researchers are looking for. Hence, a hypothesis looks
forward.
d. It is a tentative supposition or provisional guess, generalization which seems
to explain the situation under observation.
Research Design:
A research design is a framework or blueprint for conducting the marketing research
project. It details the procedures necessary for obtaining the information needed to
structure or solve marketing research problems.
2. Longitudinal Design:
A fixed sample (or samples) of population elements is measured repeatedly on
the same variables
A longitudinal design differs from a cross-sectional design in that the sample
or samples remain the same over time
Research Methodology (2810006) 5
7. Describe the Qualitative vs. Quantitative Research
Answer:
Qualitative Quantitative
Research Research
Written:
Non-written:
Time-series based:
national:
-usa.gov
www.hist.umn.edu/~rmccaa/IPUMSI
10. In business research errors occurs other than sampling error. What are
different non-sampling errors?
Answer:
Non-Sampling Error: Non-sampling errors can be attributed to sources other than
sampling, and they may be random or nonrandom: including errors in problem definition,
approach, scales, questionnaire design, interviewing methods, and data preparation and
analysis. Non-sampling errors consist of non-response errors and response errors.
Non-response error arises when some of the respondents included in the sample do not
respond.
Response error arises when respondents give inaccurate answers or their answers are
misrecorded or misanalyzed.
Measurement Error
Questioning Error
Recording Error
Cheating Error
Inability Error
Unwillingness Error
ADVANTAGES:
-can be used to collect many different kinds of information
-Quick and low cost as compared to observation and experimental method.
LIMITATIONS:
-Respondents reluctance to answer questions asked by unknown interviewers about
things they consider private.
-Busy people may not want to take the time
-may try to help by giving pleasant answers
-unable to answer because they cannot remember or never gave a thought to what they
do and why
-may answer in order to look smart or well informed.
Telephone Interviewing:
- quick method
- more flexible as interviewer can explain questions not understood by the respondent
- depending on respondents answer they can skip some Qs and probe more on others
- allows greater sample control
- response rate tends to be higher than mail
Drawbacks:
-Cost per respondent higher
-Some people may not want to discuss personal Qs with interviewer
-Interviewers manner of speaking may affect the respondents answers
-Different interviewers may interpret and record response in a variety of ways
-under time pressure ,data may be entered without actually interviewing
Personal Interviewing:
It is very flexible and can be used to collect large amounts of information. Trained
interviewers are can hold the respondents attention and are available to clarify difficult
questions. They can guide interviews, explore issues, and probe as the situation requires.
Personal interview can be used in any type of questionnaire and can be conducted fairly
quickly.
Interviewers can also show actual products, advertisements, packages and observe and
record their reactions and behaviour.
This takes two forms-
Individual- Intercept interviewing
Group - Focus Group Interviewing
12. Define projective techniques. Explain with illustration four different types
of projective techniques.
Answer:
Because researchers are often looking for hidden or suppressed meanings, Projective
techniques can be used within the interview structures. Some of these techniques
include:
Observation can yield information which people are normally unwilling or unable to provide.
e.g. Observing numerous plates containing uneaten portions the same menu items indicates
that food is not satisfactory.
Types of Observation:
1. Structured for descriptive research
2. Unstructuredfor exploratory research
3. Disguised Observation
4. Undisguised Observation
5. Natural Observation
6. Contrived Observation
Observation Methods:
1. Personal Observation
2. Mechanical Observation
3. Audit
4. Content Analysis
5. Trace Analysis
Advantages:
They permit measurement of actual behavior rather than reports of intended or preferred
behavior.
Limitations:
The reasons for the observed behavior may not be determined since little is known about
the underlying motives, beliefs, attitudes, and preferences.
Selective perception (bias in the researcher's perception) can bias the data.
Observational data are often time-consuming and expensive, and it is difficult to observe
certain forms of behavior.
In some cases, the use of observational methods may be unethical, as in observing people
without their knowledge or consent.
It is best to view observation as a complement to survey methods, rather than as being in
competition with them.
14. What are experiments? What is experimental Design? Explain the three
types of most widely accepted experimental research designs.
Answer:
The process of examining the truth of a statistical hypothesis, relating to some research problem,
is known as an experiment. For example, we can conduct an experiment to examine the
usefulness of a certain newly developed drug. Experiments can be of two types viz., absolute
experiment and comparative experiment. If we want to determine the impact of a fertilizer on the
yield of a crop, it is a case of absolute experiment; but if we want to determine the impact of
one fertilizer as compared to the impact of some other fertilizer, our experiment then will be
termed as a comparative experiment. Often, we undertake comparative experiments when we talk
of designs of experiments.
the test units and how these units are to be divided into homogeneous subsamples,
what independent variables or treatments are to be manipulated,
what dependent variables are to be measured; and
how the extraneous variables are to be controlled.
15. What is primary and secondary data? Write its advantages and
disadvantages?
Answer:
Primary data are originated by a researcher for the specific purpose of addressing the
problem at hand. The collection of primary data involves all six steps of the marketing
research process.
Secondary data are data that have already been collected for purposes other than the
problem at hand. These data can be located quickly and inexpensively.
Advantages of Primary Data:
1. Reliability
2. Specific to problem on hand
Disadvantages
Extraneous variables are all variables other than the independent variables that affect
the response of the test units, e.g., store size, store location, and competitive effort.
17. What are the evaluating factors used for secondary data sources? Also
explain how each of the five factors influences on evaluation of the
secondary sources.
Answer:
1. Personal Observation:
A researcher observes actual behavior as it occurs.
The observer does not attempt to manipulate the phenomenon being observed but merely
records what takes place.
For example, a researcher might record traffic counts and observe traffic flows in a
department store.
2. Mechanical Observation:
Do not require respondents' direct participation.
The AC Nielsen audimeter
Turnstiles that record the number of people entering or leaving a building.
On-site cameras (still, motion picture, or video)
Optical scanners in supermarkets
Do require respondent involvement.
Eye-tracking monitors
Pupilometers
Psychogalvanometers
Voice pitch analyzers
Devices measuring response latency
3. Audit:
The researcher collects data by examining physical records or performing inventory
analysis.
Data are collected personally by the researcher.
The data are based upon counts, usually of physical objects.
Research Methodology (2810006) 17
Retail and wholesale audits conducted by marketing research suppliers were discussed in
the context of syndicated data.
4. Content Analysis:
The objective, systematic, and quantitative description of the manifest content of a
communication.
The unit of analysis may be words, characters (individuals or objects), themes
(propositions), space and time measures (length or duration of the message), or topics
(subject of the message).
Analytical categories for classifying the units are developed and the communication is
broken down according to prescribed rules.
5. Trace Analysis:
Data collection is based on physical traces, or evidence, of past behavior.
The number of different fingerprints on a page was used to gauge the readership of
various advertisements in a magazine.
The position of the radio dials in cars brought in for service was used to estimate share of
listening audience of various radio stations.
The age and condition of cars in a parking lot were used to assess the affluence of
customers.
20. What is measurement? What are the different measurement scales used in
research? What are the essential differences among them? Explain in
brief with example.
Answer:
1. Nominal Scale:
The numbers serve only as labels or tags for identifying and classifying objects.
When used for identification, there is a strict one-to-one correspondence between the
numbers and the objects.
The numbers do not reflect the amount of the characteristic possessed by the objects.
The only permissible operation on the numbers in a nominal scale is counting.
2. Ordinal Scale:
A ranking scale in which numbers are assigned to objects to indicate the relative
extent to which the objects possess some characteristic.
Can determine whether an object has more or less of a characteristic than some other
object, but not how much more or less.
Any series of numbers can be assigned that preserves the ordered relationships
between the objects.
In addition to the counting operation allowable for nominal scale data, ordinal scales
permit the use of statistics based on centiles, e.g., percentile, quartile, median.
3. Interval Scale:
Numerically equal distances on the scale represent equal values in the characteristic
being measured.
It permits comparison of the differences between objects.
The location of the zero point is not fixed. Both the zero point and the units of
measurement are arbitrary.
It is not meaningful to take ratios of scale values.
Statistical techniques that may be used include all of those that can be applied to
nominal and ordinal data, and in addition the arithmetic mean, standard deviation, and
other statistics commonly used in marketing research.
4. Ratio Scale:
Possesses all the properties of the nominal, ordinal, and interval scales.
It has an absolute zero point.
It is meaningful to compute ratios of scale values.
All statistical techniques can be applied to ratio data.
a. Likert Scale: The Likert scale requires the respondents to indicate a degree of
agreement or disagreement with each of a series of statements about the stimulus
objects.
b. Semantic Differential Scale: The semantic differential is a seven-point rating scale
with end points associated with bipolar labels that have semantic meaning. The
negative adjective or phrase sometimes appears at the left side of the scale and
sometimes at the right. This controls the tendency of some respondents,
particularly those with very positive or very negative attitudes, to mark the right-
or left-hand sides without reading the labels. Individual items on a semantic
differential scale may be scored on either a -3 to +3 or a 1 to 7 scale.
c. Stapel Scale: The Stapel scale is a unipolar rating scale with ten categories numbered
from -5 to +5, without a neutral point (zero). This scale is usually presented
vertically. The data obtained by using a Stapel scale can be analyzed in the same way
as semantic differential data.
1. Reliability:
Consistency of the instrument is called as Reliability.
Reliability can be defined as the extent to which measures are free from random
error, XR. If XR = 0, the measure is perfectly reliable.
If the same scale/instrument is used and we use the scale/instrument for the similar
conditions then the measured data should be consistent (same).
The systematic error does not affect the reliability but random error does. So, for
better reliability, the scale should be free from random errors.
If random error=O, the scale is most reliable.
Methods of Reliability:
In test-retest reliability, respondents are administered identical sets of scale
items at two different times and the degree of similarity between the two
measurements is determined. If the association or correlation between test and
retest result is high, we can conclude that scale is highly reliable. The problem
with this are: It is illogical to conduct research on same scale twice and sometimes
impossible also (Mall intercept), interval between test and retest influence
reliability, change in attitude between test and retest.
In alternative-forms reliability, two equivalent forms of the scale are
constructed and the same respondents are measured at two different times, with a
different form being used each time. The scales are similar to each other and the
same groups are measured. 2 instruments: A and B. the correlation or association
2. Validity:
A measure has validity if it measures what it is supposed to measure.
For example, a study on consumer likeliness of taste in a restaurant, should have
questions related to taste only and so it is called valid instrument. If it includes
question of service provided, it is invalid scale for measuring likeliness of taste in a
restaurant.
An instrument is valid if it is free from systematic and random errors.
The validity of a scale may be defined as the extent to which differences in observed
scale scores reflect true differences among objects on the characteristic being
measured, rather than systematic or random error. Perfect validity requires that there
be no measurement error (XO = XT, XR = 0, XS = 0).
3. Generalizability:
It is the degree to which a study based on a sample applies to a universe of
generalizations.
24. Explain the reasons for sampling. Elaborate the process of sampling.
Answer:
1. Non-Probability Sampling
Sampling techniques that do not use chance selection procedures. Rather, they rely on
the personal judgment of the researcher.
2. Probability Sampling
A sampling procedure in which each element of the population has a fixed probabilistic
chance of being selected for the sample.
b. Systematic Sampling:
The sample is chosen by selecting a random starting point and then picking every
ith element in succession from the sampling frame.
The sampling interval, i, is determined by dividing the population size N by the
sample size n and rounding to the nearest integer.
When the ordering of the elements is related to the characteristic of interest,
systematic sampling increases the representativeness of the sample.
If the ordering of the elements produces a cyclical pattern, systematic sampling
may decrease the representativeness of the sample.
For example, there are 100,000 elements in the population and a sample of 1,000
is desired. In this case the sampling interval, i, is 100. A random number
between 1 and 100 is selected. If, for example, this number is 23, the sample
consists of elements 23, 123, 223, 323, 423, 523, and so on.
Advantages: Can increase representativeness, easier to implement than SRS,
sample frame not necessary
Disadvantages: Can decrease representativeness
d. Cluster Sampling:
The target population is first divided into mutually exclusive and collectively
exhaustive subpopulations, or clusters.
Then a random sample of clusters is selected, based on a probability sampling
technique such as SRS.
For each selected cluster, either all the elements are included in the sample (one-
stage) or a sample of elements is drawn probabilistically (two-stage).
Elements within a cluster should be as heterogeneous as possible, but clusters
themselves should be as homogeneous as possible. Ideally, each cluster should be a
small-scale representation of the population.
In probability proportionate to size sampling, the clusters are sampled with
probability proportional to size. In the second stage, the probability of selecting a
sampling unit in a selected cluster varies inversely with the size of the cluster.
Types of Cluster Sampling: 1. One Stage Cluster Sampling 2. Two Stage Cluster
Sampling 3. Multi Stage Cluster Sampling
Advantages: Easy to implement and cost effective
Disadvantages: Imprecise, Difficult to compute and interpret results
Type I Error:
Type I error occurs when the sample results lead to the rejection of the null hypothesis
when it is in fact true. The probability of type I error is also called the alpha and level of
significance.
Type II Error
Type II error occurs when, based on the sample results, the null hypothesis is not
rejected when it is in fact false. The probability of type II error is denoted by beta. Unlike, which
is specified by the researcher, the magnitude of depends on the actual value of the population
parameter (proportion).
Parametric Tests:
Parametric methods make assumptions about the underlying distribution from which sample
populations are selected.
Parametric tests are based on models with some assumptions. If the assumptions hold good,
these tests offer a more powerful tool for analysis. It usually assumes certain properties of the
parent population from which we draw samples. Assumptions like observations come from a
normal population, sample size is large, assumptions about population parameters like mean,
variance, etc., must hold good before parametric tests can be used. But these are situations
when the researcher cannot or does not want to make such assumptions. In such situations we
use statistical methods for testing hypotheses which are called non-parametric tests because
such tests do not depend on any assumption about the parameters of the parent population.
Besides, most non-parametric tests assume only nominal or ordinal data, whereas parametric
tests require measurement equivalent to at least an interval or ratio scale. As a result, non-
parametric tests need more observations than parametric tests to achieve the same size of
type-I and Type-II errors.
All these tests are based on the assumptions of normality, the source of data is considered to
be normally distributed.
Non-Parametric Tests:
As the name implies, non-parametric tests do not require parametric assumptions because
interval data are converted to rank ordered data. Handling of rank ordered data is considered
strength of non-parametric tests.
All practical data follow normal distribution. Under such situations, one can estimate the
parameters such as mean, variance, etc., and use the standard tests, they are known as
parametric tests. The practical data may be non-normal and/or it may not be possible to
estimate the parameters of the data. The tests which are used for such situations are called
non-parametric tests. Since, these tests are based on data which are free from distribution and
parameter; these tests are known as non-parametric tests.
The test technique makes use of one or more values obtained from sample data to arrive at a
probability statement about the hypothesis. But no such assumptions are made in case of non-
parametric tests.
In a statistical test, two kinds of assertions are involved, an assertion directly related to the
purpose of the investigation and other assertions to make a probability statement. The former
is an assertion to be tested and is technically called a hypothesis, whereas the set of all other
assertions is called the model. When we apply a test without a model, it is a distribution free
test or the non-parametric test. Non-parametric tests do not make an assumption about the
parameters of the population and thus do not make use of the parameters of the distribution.
In other words, under non-parametric tests we do not assume that a particular distribution is
applicable, or that a certain value is attached to a parameter of the population.
Popular non-parametric tests are Chi-Square Test, Kolmogorov-Smirnov Test, Run Test,
Median, Mann-Whitney U Test, Wilcoxon Signed Rank Test, Kruskal Wallis Test
Coding:
Before permitting statistical analysis, a researcher has to prepare data for analysis. This
preparation is done by data coding. Coding of data is probably the most crucial step in the
analytical process. Codes or categories are tags or labels for allocating units of meaning to
the descriptive or inferential information compiled during a study. In coding, each answer is
identified and classified with a numerical score or other symbolic characteristics for
processing the data in computers. While coding, researchers generally select a convenient
way of entering the data in a spreadsheet. The data can be conveniently exported or imported
to other software such as SPSS and SAS. The character of information occupies the column
position in the spreadsheet with the specific answers to the questions entered in a row. Thus,
each row of the spreadsheet will indicate the respondents answers on the column heads.
Raw data, particularly qualitative data can be very interesting to look at, but do not
express any relationship unless anlysed systematically.
Qualitative data are textual, non-numerical and unstructured.
Editing:
Editing is actually checking of the questionnaire for suspicious, inconsistent, intelligible, and
incomplete answers visible from careful study of the questionnaire. Field editing is a good way
to deal with the answering problems in the questionnaire and is usually done by the supervisor on
the same day when the interviewer has conducted the interview. Editing on the same day of the
interview has various advantages. In the case of any inconsistent or illogical answering, the
supervisor can ask the interviewer about it, and the interviewer can also answer well because he
or she has just conducted the interview and the possibility of forgetting the incident if minimal.
The interviewer can then recall and discuss the answer and situation with the supervisor. In this
manner, a logical answer to the question may be arrived. Field editing can also identify any fault
in the process of the interview just on time and can be rectified without much delay in the
interview process. In some interview techniques such as mail technique, field editing is not
possible. In such a situation, in-house editing is done, which rigorously examines the
questionnaire for any deficiency and a team of experts edit and code these data.
Editing includes:
Detects errors and omissions,
Corrects them when possible, and
Certifies that minimum data quality standards are achieved
Field Editing
translation of ad hoc abbreviations and symbols used during data collection
validation of the field results.
Central Editing
Tabulation of Data:
General Purpose Tables: The general purpose table is informative. Its main aim is to
present a comprehensive set of data in such an arrangement that any particular item can be
easily located.
Special Purpose Table: Special purpose tables are relatively small in size and are designed
to present as effectively as possible the findings of a particular study for which it is designed.
After conducting any research, the researcher needs to draft the work. Drafting is a scientific
procedure and needs a systematic and careful articulation. When the organization for which
the researcher is conducting the research has provided guidelines to prepare the research
report, one has to follow them. However, if there is no guideline available from the
sponsoring organization, then the researcher can follow a reasonable pattern of presenting a
written research report. In fact, there is no single universally accepted guideline available for
organization of a written report. In fact, the organization of the written report depends on the
type of the target group that it addresses. However, following is the format for the written
report. Although this format presents a broad guideline for report organization, in real sense,
the researcher enjoys flexibility to either include or exclude the items from the given list.
1. Title Page
2. Letter of Transmittal
3. Letter of Authorization
4. Table of Contents (including list of figures, tables and charts)
5. Executive Summary
Objective
Concise statement of the methodology
Results
Conclusions
Recommendations
6. Body
Introduction
Research Objective
Research Methodology (Sample, Sample Size, Sample Technique, Scaling
Technique, Data source, Research Design, Instrument for data collection,
Contact Methods, Test statistics, Field Work)
Results and findings
Conclusions and Recommendations
Limitations of the research
7. Appendix
Copies of data collection forms
Statistical output details
General tables that are not included in the body
Bibliography
Other required support material
Research Methodology (2810006) 35
32. Explain Univariate, Bivariate and Multivariate analysis with examples.
Answer:
Empirical testing typically involves inferential statistics. This means that an inference will be
drawn about some population based on observations of a sample representing that
population. Statistical analysis can be divided into several groups:
Univariate analysis is the simplest form of quantitative (statistical) analysis. The analysis is
carried out with the description of a single variable and its attributes of the applicable unit of
analysis. For example, if the variable age was the subject of the analysis, the researcher
would look at how many subjects fall into a given age attribute categories.
Univariate analysis contrasts with bivariate analysis the analysis of two variables
simultaneously or multivariate analysis the analysis of multiple variables simultaneously.
Univariate analysis is also used primarily for descriptive purposes, while bivariate and
multivariate analysis are geared more towards explanatory purposes. Univariate analysis is
commonly used in the first stages of research, in analyzing the data at hand, before being
supplemented by more advance, inferential bivariate or multivariate analysis.
There are several tools used in univariate analysis; their applicability depends on whether we
are dealing with a continuous variable (such as age) or a discrete variable (such as gender).
Another set of measures used in the univariate analysis, complementing the study of the
central tendency, involves studying the statistical dispersion. Those measurements look at
how the values are distributed around values of central tendency. The dispersion measures
most often involve studying the range, interquartile range, and the standard deviation.
Bivariate analysis is one of the simplest forms of the quantitative (statistical) analysis. It
involves the analysis of two variables (often denoted as X, Y), for the purpose of determining
the empirical relationship between them. In order to see if the variables are related to one
another, it is common to measure how those two variables simultaneously change together
(see also covariance).
Bivariate analysis can be helpful in testing simple hypotheses of association and causality
checking to what extent it becomes easier to know and predict a value for the dependent
variable if we know a case's value on the independent variable (see also correlation).
Bivariate analysis can be contrasted with univariate analysis in which only one variable is
analysed. Furthermore, the purpose of a univariate analysis is descriptive. Subgroup
comparison the descriptive analysis of two variables can be sometimes seen as a very
simple form of bivariate analysis (or as univariate analysis extended to two variables). The
major differentiating point between univariate and bivariate analysis, in addition to looking at
more than one variable, is that the purpose of a bivariate analysis goes beyond simply
descriptive: it is the analysis of the relationship between the two variables.
Types of analysis
Common forms of bivariate analysis involve creating a percentage table, a scatterplot graph,
or the computation of a simple correlation coefficient. For example, a bivariate analysis
intended to investigate whether there is any significant difference in earnings of men and
women might involve creating a table of percentages of the population within various
categories, using categories based on gender and earnings:
over 50,000$ 8% 1%
The types of analysis that are suited to particular pairs of variables vary in accordance with
the level of measurement of the variables of interest (e.g., nominal/categorical, ordinal,
interval/ratio).
Bivariate analysis is a simple (two variable) variant of multivariate analysis (where multiple
relations between multiple variables are examined simultaneously).
Step 1: Define the nature of the relationship whether the values of the independent variables
relate to the values of the dependent variable or not.
Step 4: Identify the strength of the relationship, i.e. the degree to which the values of the
independent variable explain the variation in the dependent variable.
Multivariate statistical analysis tests hypotheses and models involving multiple (three or
more) variables or sets of variables.
Multivariate statistics concerns understanding the different aims and background of each of
the different forms of multivariate analysis, and how they relate to each other. The practical
implementation of multivariate statistics to a particular problem may involve several types of
univariate and multivariate analysis in order to understand the relationships between
variables and their relevance to the actual problem being studied.
There are many different models, each with its own type of analysis:
33. Explain in brief the following and their specific application in data
presentation
a. Cross Tabulation
Cross tabulation approach is especially useful when the data are in nominal form. Under it
we classify each variable into two or more categories and then cross classify the variables in
these subcategories. Then we look for interactions between them which may be symmetrical,
reciprocal or asymmetrical. A symmetrical relationship is one in which the two variables vary
together, but we assume that neither variable is due to the other. A reciprocal relationship
exists when the two variables mutually influence or reinforce each other. Asymmetrical
relationship is said to exist if one variable (the independent variable) is responsible for
another variable (the dependent variable). The cross classification procedure begins with a
two-way table which indicates whether there is or there is not an interrelationship between
the variables. This sort of analysis can be further elaborated in which case a third factor is
introduced into the association through cross-classifying the three variables. By doing so we
find conditional relationship in which factor X appears to affect factor Y only when factor Z is
held constant. The correlation, if any, found through this approach is not considered a very
Powerful form of statistical correlation and accordingly we use some other methods when
data happen to be either ordinal or interval or ratio data.
Pareto analysis is a statistical technique in decision making that is used for selection
of a limited number of tasks that produce significant overall effect. It uses the Pareto
principle the idea that by doing 20% of work, 80% of the advantage of doing the
entire job can be generated. Or in terms of quality improvement, a large majority of
problems (80%) are produced by a few key causes (20%).
Pareto analysis is a formal technique useful where many possible courses of action
are competing for attention. In essence, the problem-solver estimates the benefit
delivered by each action, then selects a number of the most effective actions that
deliver a total benefit reasonably close to the maximal possible one.
This technique helps to identify the top portion of causes that need to be addressed to
resolve the majority of problems. Once the predominant causes are identified, then
tools like the Ishikawa diagram or Fish-bone Analysis can be used to identify the root
causes of the problems. While it is common to refer to pareto as "20/80", under the
assumption that, in all situations, 20% of causes determine 80% of problems, this
ratio is merely a convenient rule of thumb and is not nor should it be considered
immutable law of nature.
c. Box Plots
A maths test is marked out of 50. The marks for the class are shown below:
This data can be more easily interpreted if we represent it in a stem and leaf
diagram.
The tens form the 'stem' and the units form the 'leaves'.
It is usual for the numbers to be ordered. So, for example, the row
f. Histogram
g. Pie charts
A pie chart (or a circle graph) is a circular chart divided into sectors, illustrating
proportion. In a pie chart, the arc length of each sector (and consequently its
central angle and area), is proportional to the quantity it represents. When angles
are measured with 1 turn as unit then a number of percent is identified with the
same number of centiturns. Together, the sectors create a full disk. It is named for
its resemblance to a pie which has been sliced. The size of the sectors are
calculated by converting between percentage and degrees or by the use of a
percentage protractor. The earliest known pie chart is generally credited to
William Playfair's Statistical Breviary of 1801.
The pie chart is perhaps the most widely used statistical chart in the business
world and the mass media. However, it has been criticized,[5] and some
recommend avoiding it, pointing out in particular that it is difficult to compare
different sections of a given pie chart, or to compare data across different pie
charts. Pie charts can be an effective way of displaying information in some cases,
in particular if the intent is to compare the size of a slice with the whole pie, rather
than comparing the slices among them. Pie charts work particularly well when the
slices represent 25 to 50% of the data, but in general, other plots such as the bar
1. The purpose of the research should be clearly defined and common concepts be
used.
2. The research procedure used should be described in sufficient detail to permit
another researcher to repeat the research for further advancement, keeping the
continuity of what has already been attained.
3. The procedural design of the research should be carefully planned to yield results
that are as objective as possible.
4. The researcher should report with complete frankness, flaws in procedural design
and estimate their effects upon the findings.
5. The analysis of data should be sufficiently adequate to reveal its significance and
the methods of analysis used should be appropriate. The validity and reliability of the
data should be checked carefully.
6. Conclusions should be confined to those justified by the data of the research and
limited to those for which the data provide an adequate basis.
7. Greater confidence in research is warranted if the researcher is experienced, has a
good reputation in research and is a person of integrity.
36. Explain one tailed and two tailed tests of significance with example.
Answer:
One Tailed Test: A statistical test in which the critical area of a distribution is one-sided so
that it is either greater than or less than a certain value, but not both. If the sample that is
being tested falls into the one-sided critical area, the alternative hypothesis will be accepted
instead of the null hypothesis. The one-tailed test gets its name from testing the area under
one of the tails (sides) of a normal distribution, although the test can be used in other non-
normal distributions as well.
Research Methodology (2810006) 47
An example of when one would want to use a one-tailed test is in the error rate of a
factory. Lets say a label manufacturer wants to make sure that errors on labels are
below 1%. It would be too costly to have someone check every label, so the factory
selects random samples of the labels and test whether errors exceed 1% with
whatever level of significance they choose. This represents the implementation of
a one-tailed test.
Example
Either of these two alternative hypotheses would lead to a one-sided test. Presumably,
we would want to test the null hypothesis against the first alternative hypothesis since
it would be useful to know if there is likely to be less than 50 matches, on average, in
a box (no one would complain if they get the correct number of matches in a box or
more). Yet another alternative hypothesis could be tested against the same null,
leading this time to a two-sided test:
Here, nothing specific can be said about the average number of matches in a box;
only that, if we could reject the null hypothesis in our test, we would know that the
average number of matches in a box is likely to be less than or greater than 50.
Two Tailed Test: A statistical test in which the critical area of a distribution is two
sided and tests whether a sample is either greater than or less than a certain range of
values. If the sample that is being tested falls into either of the critical areas, the
alternative hypothesis will be accepted instead of the null hypothesis. The two-tailed
test gets its name from testing the area under both of the tails (sides) of a normal
distribution, although the test can be used in other non-normal distributions.
Either of these two alternative hypotheses would lead to a one-sided test. Presumably,
we would want to test the null hypothesis against the first alternative hypothesis since
it would be useful to know if there is likely to be less than 50 matches, on average, in
a box (no one would complain if they get the correct number of matches in a box or
more). Yet another alternative hypothesis could be tested against the same null,
leading this time to a two-sided test:
Here, nothing specific can be said about the average number of matches in a box;
only that, if we could reject the null hypothesis in our test, we would know that the
average number of matches in a box is likely to be less than or greater than 50.
Null Hypothesis: A type of hypothesis used in statistics that proposes that no statistical
significance exists in a set of given observations. The null hypothesis attempts to show that
no variation exists between variables, or that a single variable is no different than zero. It is
presumed to be true until statistical evidence nullifies it for an alternative hypothesis.
The null hypothesis assumes that any kind of difference or significance you see in a
set of data is due to chance.
For example, Chuck sees that his investment strategy produces higher average returns
than simply buying and holding a stock. The null hypothesis claims that there is no
difference between the two average returns, and Chuck has to believe this until he
proves otherwise. Refuting the null hypothesis would require showing statistical
significance, which can be found using a variety of tests. If Chuck conducts one of
these tests and proves that the difference between his returns and the buy-and-hold
returns is significant, he can then refute the null hypothesis.
Alternate Hypothesis:
The alternative hypothesis, H1, is a statement of what a statistical hypothesis test is set
up to establish. For example, in a clinical trial of a new drug, the alternative
The final conclusion once the test has been carried out is always given in terms of the
null hypothesis. We either "Reject H0 in favour of H1" or "Do not reject H0". We
never conclude "Reject H1", or even "Accept H1".
If we conclude "Do not reject H0", this does not necessarily mean that the null
hypothesis is true, it only suggests that there is not sufficient evidence against H 0 in
favour of H1. Rejecting the null hypothesis then, suggests that the alternative
hypothesis may be true.
Summary:
5. Both the null and alternative hypotheses are necessary in statistical hypothesis testing in
scientific, medical, and other research studies.