SlideShare ist ein Scribd-Unternehmen logo
1 von 48
HOW TO DEVELOP SELF
REPORTED SCALE
Presenter:
Mahesh Chand
Nursing Tutor
Introduction
Researchers sometimes are unable to identify an
appropriate instrument to operationalize a
construct. This may occur when the construct is
new, but often it is due to limitations of existing
instruments. Because this situation occurs fairly
often, this chapter provides an overview of the
steps involved in developing high-quality self-
report scales.
Steps
1. Conceptualization and item generation
2. Preliminary evaluation of items
3. Field testing the instruments
4. Analysis of scale development data
5. Scale refinement and validation
6. Interpretation of scale score
7. Critiquing
 Conceptualization the construct
 Deciding the type of scale
 Generating an item pool
 Making decision about item features
 Wording the items
CONCEPTUALIZATION AND ITEM GENERATION
You will not be able to quantify an attribute adequately
unless you thoroughly understand the underlying
construct you wish to capture.
• Underlying construct: You cannot develop items to
produce the right score, and you cannot expect good
content and construct validity, if you are unclear about
the construct. Thus, the first step in scale development
is to become an expert on the construct.
• Complex construct: Complex constructs have a number
of different facets or dimensions, and it is important to
identify and understand each one.
Conceptualizing the Construct
Deciding on the Type of Scale
• Before items can be generated, you need to
decide on the type of scale you wish to create
because item characteristics vary by scale type.
E.g. Traditional summated rating (Likert type) scale
Generating an Item Pool
• An early step in scale construction is to develop a pool
of possible items for the scale.
• Items—which collectively constitute the operational
definition of the construct—need to be carefully
crafted to reflect the latent variable they are designed
to measure.
• Here are some possible sources for generating an item
pool-
i. Existing instruments
ii. The literature
iii. Concept analysis
iv. In-depth qualitative research
v. Clinical observations
Making Decisions about Item Features
You need to make decisions about such issues as the
number of items to develop, the number and form of
the response options, whether to include positively
and negatively worded items, and how to deal with
time.
• Number of Items: There is no magic formula for
how many items should be developed, but our
advice is to generate a very large pool of items. As
you proceed, many items will be discarded. Longer
scales tend to be more reliable, so starting with a
large number of items.
E.g.: In your final scale (e.g., 30 to 40 items for a 10-
item scale), but the minimum should be 50% more
(e.g., 15 items for a 10-item scale).
• Response Options:
Scale items involve both a stem
(declarative statement) and response options. Likert scales
involve response options on a continuum of agreement,
such as frequency (never/always), importance (very
important /unimportant), quality (excellent/very poor),
and likelihood (highly likely/impossible).
The midpoint can be labeled with such phrases as “neither
agree nor disagree,” “undecided,” “agree and disagree
equally,” or simply “?”.
Strongly disagree, disagree, undecided, agree, strongly agree
Never, almost never, sometimes, often, almost always
Very important, important, somewhat important, of little importance, unimportant
• Positive and Negative Stems:
Psychometricians advised scale developers to
deliberately include both positively and negatively
worded statements and to reverse score negative
items. E.g. Two items for a scale of depression: “I
frequently feel sad,” and “I don’t often feel sad.”
Research suggests that acceptance can be
minimized by putting the most positive response
options (e.g. strongly agree) at the end of the list
rather than at the beginning.
• Item Intensity: The intensity of the statements
should be similar and fairly strongly worded. If
items are worded such that almost anyone
would agree with them.
• For E.g. “Good health is important” would
generate almost universal agreement
• Item Time Frames: A time frame should not emerge
as a consequence of item development. You should
decide in advance, based on your conceptual
understanding of the construct and the needs for
which the scale is being constructed, how to deal
with time.
• 2-week period was chosen because it parallels the
duration of symptoms required for a diagnosis of
major depressive episode.
Postpartum Depression Screening Scale asks
respondents to indicate their emotional state in the
previous 2 weeks— for example, over the last 2 weeks I:
“ . . . had trouble sleeping even when my baby was
asleep” and “ . . . felt like a failure as a mother”
Wording the Items
• Items should be worded in such a manner that
every respondent is answering the same
question.
a) Clarity
b) Jargon (technical or special words)
c) Length
d) Double negatives
e) Double-barelled items.(Insects and Snakes)
PRELIMINARY EVALUATION OF ITEMS
Preliminary evaluation of the items can be made under following:
• Internal review
• Input for target population
• External review by expert
• Internal Review:
Once a large item pool has been
generated, it is time for critical appraisal. It is vital
importance to assess the scale’s readability, unless the
scale is intended for a highly educated population. There
are different approaches for assessing the reading level of
written documents.
Set A Set B
• I am frequently exhausted. I am often tired.
• I invariably get insufficient sleep. I don’t get enough sleep.
Set A have a Flesch-Kincaid grade level of 12.0 and a Flesch reading
ease score of 4.8. Set B, by contrast, has a grade level of 1.8 and a
reading ease score of 89.4.
• It is clear from the foregoing analysis that the second set of items
would be superior for a population that includes people with
limited education.
• Input from the Target Population:
In the next step, the initial
pool of items is pretested. a small sample of people
(20 to 40 or so) representing the target population
is invited to complete the items.
• In cognitive interviews, people are asked to
reflect upon their interpretation of the items and
their answers so that the underlying process of
response selection is better understood.
There are two basic approaches to cognitive
interviewing-
• The think-aloud method, wherein respondents
are asked to explain step by step how they
processed the question and arrived at an
answer.
• A second approach is to conduct an interview
in which the interviewer uses a series of
targeted probes that encourage reflection
about underlying cognitive processes.
• As an alternative or supplement to pretests,
focus groups can also be used at this stage in
scale development.
External Review by Experts:
External review of the
revised items by a panel of experts should be
undertaken to assess the scale’s content validity. It
is advisable to undertake two rounds of review.
 First to refine faulty items or to add new items
to cover the domain adequately.
 Second to formally assess the content validity of
the items and scale.
It can be done under following steps-
• Selecting and Recruiting the Experts:
In the initial phase of a
two-part review, we advise having an expert panel
of 8 to 12 members, with a good mix in terms of
roles and disciplines.
• The second panel for formally assessing the
content validity of a more refined set of items
should consist of three to five experts in the
content area.
E.g.: For a scale designed to measure fear of dying
in the elderly, the experts might include nurses,
gerontologists, and psychiatrists.
• Preliminary Expert Review: Content Validation
of Items:
The experts’ job is to evaluate individual items and the
overall scale using guidelines established by the scale
developer.
The first panel of experts is usually invited to rate each
item in terms of several characteristics, such as clarity of
wording; relevance of the item to the construct; and
appropriateness for the target population.
The relevance is most often rated as follows: 1 = not
relevant, 2 = somewhat relevant, 3 = quite relevant, 4 =
highly relevant.
• The scale items shown below have been developed to measure one dimension of the
construct of safe sexual behaviors among adolescents
Item
s
Not
relevant
Somewhat
relevant
Quite
relevant
Highly
relevant
1 I asked my partner about his/her
sexual history before having
intercourse
1 2 3 4
2 I don’t have sex without asking the
person if he/she has been tested for
HIV/AIDS
1 2 3 4
3 When I am having sex with someone
for the first time insist that we use a
condom
1 2 3 4
An expert panel should be asked to consider whether the items taken as a whole
adequately cover the construct domain.
Standard method for computing an item-level content
validity index (I-CVI) =
Number giving a rating of 3 or 4 on the 4-point scale
---------------------------------------------------------
Divided by the number of experts.
For example, if five experts rated an item as 3 and one
rated the item as 2, the I-CVI would be .83.
This means that there must be 100% agreement among
raters when there are four or fewer experts. When there
are five to eight experts, one rating of “not relevant” can
be tolerated.
Items with lower than desired I-CVIs need careful scrutiny.
• Content Validation of the Scale:
In the second round of
content validation, a smaller group of experts (three
to five) can be used to evaluate the relevance of the
revised set of items and to compute the scale content
validity.
• First, experts who rated every item as “highly
relevant” (or “not relevant”) may not be sufficiently
discriminating.
• Second, an expert who gave high ratings to items
that were judged by most others to not be relevant
might be problematic.
• Third, the proportion of items judged relevant
should be computed for all judges.
• For example, if an expert rated 8 out of 10 items
as relevant, the proportion for that judge would be
.80. The pattern across experts can be examined
for “outliers.” If the average proportion across
raters is, for example, .80, you might consider not
inviting back for a second round of experts whose
average proportion was either very low (e.g., .50)
or very high (e.g., 1.0).
• On a 10- item scale, for example, if the I-CVIs for 5
items were .80 and the I-CVIs for the remaining 5
items were 1.00, then the S-CVI/Ave would be .90.
An S-CVI/Ave of .90 or higher is desirable.
FIELD TESTING THE INSTRUMENT
At this point, you will have whittled down and
refined your items based on your own and others’
careful scrutiny. It can be done under-
• Developing sampling plan
• Developing a data collection plan
• Preparing for data collection
• Developing a Sampling Plan:
The sample for
testing the scale should be representative of the
population for whom the scale has been devised
and should be large enough to support complex
analyses.
• Recommendations range from 3-4 people per
item to 40-50 per item, with 10 per item often
being recommended. That means that if you
have 20 items, your sample should be at least
200.
• Developing a Data Collection Plan:
Decisions have to be
made concerning how to administer the
instrument (e.g., by mailed or distributed
questionnaires, over the Internet) and what to
include in the instrument.
• The instrument should include the scale items
and basic demographic information.
• Preparing for Data Collection:
As in all data collection
efforts, care should be taken to make the instrument
attractive, professional looking, and easy to
understand. Colleagues, mentors, or family members
should be asked to evaluate the appearance of the
instrument before it is reproduced.
• Instructions for completing the instrument should
be clear, and a readability assessment of the
instructions is useful. There should be no ambiguity
about what is expected of respondents.
ANALYSIS OF SCALE DEVELOPMENT DATA
Analysis of developed data can be done under
following steps:
• Basic item analysis
• Exploratory factor analysis
• Internal consistency analysis
• Test – retest reliability analysis
• Basic Item Analysis:
Inter-item correlations
between .30 and .70 are often recommended, with
correlations lower than .30 suggesting little
compatibility with the underlying construct and
ones higher than .70 suggesting over no helpful.
However, the evaluation depends on the number of
items in the scale. An average inter-item correlation
of .57 is needed to achieve a coefficient alpha of .80
on a 3-item scale, but an average of only .29 is
needed for a 10-item scale.
• Exploratory Factor Analysis:
Factor analysis complex
interrelationships among items and identifies items that
“go together” as unified concepts.
E.g.: 50 items measuring women’s attitudes toward
menopause. We could form a scale by adding together
scores from several individual items, but which items
should be combined? Would it be reasonable to combine
all 50 items? Probably not, because the 50 items are not all
tapping the same thing—there are various dimensions to
women’s attitude toward menopause.
One dimension may relate to aging, reproductive ability,
sexuality, avoidance of monthly menstruation. Women’s
attitude on one dimension may be independent of their
attitude on another.
• Internal Consistency Analysis:
After a final set of
items is selected based on the item analysis and
factor analysis, an analysis should be undertaken to
calculate coefficient alpha. Alpha, it may be
recalled, provides an estimate of a key
measurement property of a multi-item scale,
internal consistency.
They also calculate the value of coefficient alpha for
the scale and for a hypothetical scale with each
individual item removed. Sometimes, removal of a
faulty item actually increases alpha.
• Test–Retest Reliability Analysis:
Test–retest reliability
a particularly important indicator of a scale’s
quality. When the time interval is too brief,
carryover effects can lead to artificially high
estimates of reliability.
• Some experts advise that the time interval
between measurements should be around 1
to 2 weeks.
SCALE REFINEMENT AND VALIDATION
Scale refinement and validation can be made under
following steps:
• Revising the scale
• Scoring the scale
• Conducting a validation study
• Revising the Scale:
The analyses undertaken in
the development study often suggest the need to
revise or add items. For example, if subscale alpha
coefficients are lower than .80 or so, consideration
should be given to adding items for subsequent
testing.
• Before deciding that your scale is finalized, it is a
good idea to examine the content of the items in
the scale.
• Scoring the Scale:
Scoring the scale is often
easy with Likert type items: items scores are
typically just added together with reverse scoring
of items. A system of weighting the items might
be considered.
E.g. A scale to assess a person’s risk of a disease
or condition( cardiovascular disease) might
benefit from weighting some items( high BP)
more heavily than others.
• Conducting a validation study:
Designing a validation
study entails much of the same issues (and advice)
as designing a development study, in terms of
sample composition, sample size, and data
collection strategies. Thus, we focus here on
analyses undertaken in a validation study.
• Confirmatory factory analysis: CFA involves
testing a measurement model, which specifies
the hypothesized relationships among underlying
latent and manifest variables. The measurement
model specifies the hypothesized factor
structure.
The latent variables provide a method for
evaluating relationships between observed
variables (the items) and unobserved
variables(factors)
1 2 3 4 5
Physical
Fatigue
Mental
Fatigue
6 7 8 9 10
e1 e2 e3 e4 e5 e6 e7 e8 e9 e10
The analysis would also indicate whether the overall model fit is
good, based on a goodness of fit statistic. If the hypothesized
model is not a good fit to the data, the measurement model
could be re-specified and retested.
INTERPRETATION OF SCALE SCORES
After the scale refinement and validation next
steps comes to interpretation of score.
It is done in form of:
• Percentile
• Standard score
• Norms
• Cutoff marks
• Percentiles:
A percentile indicates the
percentage of people who score below a
particular score value. Percentiles can range
from the 0th to the 99th percentile, and the
50th percentile corresponds to the median.
Percentile values are most useful when they are
determined on the basis of a large,
representative sample.
• Standard scores:
– Standard scores are expressed in terms of their
relative distance from the mean, in standard
deviation units.
– A standard score of 0.0 corresponds to a raw score
exactly at the scale’s mean.
– A standard score of 1.0 corresponds to a score 1
SD above the mean, and −1.0 corresponds to
below the mean.
– Standard scores can be readily calculated from
raw scores once the mean and SD have been
computed.
• Norms:
– Norms are often established for key demographic
characteristics, such as age and gender.
– Norms are often expressed in terms of percentiles.
For example, an adult male with a score of 72 on the
scale might be at the 80th percentile, but a female
with the same score might be at the 85th percentile.
• Guidelines for norming instruments have been
developed by Nunnally and Bernstein(1994).
• Cutoff Points:
Interpretation of scores can often
be facilitated when the instrument developer
establishes cut points for classification purposes.
Cut points are defined in terms of percentiles
E.g.: Children’s weights, those below the 5th
percentile are often interpreted as underweight
whereas those above the 95th percentile are
considered overweight.
CRITIQUING SCALE DEVELOPMENT STUDIES
• Articles on scale development appear regularly in
many nursing journals. If you are planning to use
a scale in a substantive study, you should
carefully review the methods used to construct
the scale and test its measurement properties.
You should scrutinize whether the evidence
regarding the scale’s psychometric adequacy is
sufficiently sound to merit its use.
Guidelines for Critiquing Scale
Development and Assessment Reports
• Does the report offer a clear definition of the
construct? Did it provide sufficient context for the
study through a summary of the literature and
discussion of relevant theory? Is the population for
whom the scale intended adequately described?
• Does the report indicate how items were generated?
Do the procedures seem sound? Was information
provided about the reading level of scale items?
• Does the report describe content validation efforts,
and was the description thorough? Is there evidence of
good content validity?
• Were appropriate efforts made to refine the scale
(e.g., through pretests, item analysis)?
• Was the development or validation sample of
participants appropriate in terms of
representativeness, size, and heterogeneity?
• Was factor analysis used to examine or validate
the scale’s dimensionality? If yes, did the report
offer evidence to support the factor structure and
the naming of factors?
• Were appropriate methods used to assess the
scale’s internal consistency and reliability? Were
estimates of reliability and internal consistency
sufficiently high?
References
• Denise F Polit and Beck, Nursing research 10th
edition, Wolters Kluwer publication, page No.
595-637, 499-505, 505-507
• KP Neeraja Textbook of Nursing Education, 1st
edition, Jaypee publication page No. 432-440.
• Polit, D. F. (2014). Getting serious about test-
retest reliability: A critique of retest research and
some recommendations. Quality of Life Research,
1713–1720.
• Suresh K Sharma Nursing research and statistics,
2nd edition .for reference

Weitere ähnliche Inhalte

Was ist angesagt?

Qualitative analysis\slideshow
Qualitative analysis\slideshowQualitative analysis\slideshow
Qualitative analysis\slideshowPaulo Azenha
 
Session 2 2018
Session 2 2018Session 2 2018
Session 2 2018Sue Hines
 
Critical Thinking: Its Basic Elements
Critical Thinking: Its Basic ElementsCritical Thinking: Its Basic Elements
Critical Thinking: Its Basic ElementsPatrisius Djiwandono
 
Survey Instrument Validity OERdifferentiation.com
Survey Instrument Validity OERdifferentiation.comSurvey Instrument Validity OERdifferentiation.com
Survey Instrument Validity OERdifferentiation.comOpen Education Consortium
 
David Hand:Trustworthiness of statistical analysis
David Hand:Trustworthiness of statistical analysisDavid Hand:Trustworthiness of statistical analysis
David Hand:Trustworthiness of statistical analysisjemille6
 
The Art and Science of Survey Research
The Art and Science of Survey ResearchThe Art and Science of Survey Research
The Art and Science of Survey ResearchSiobhan O'Dwyer
 
ATTITUDE MEASUREMENT SCALES-PSYCHOLOGY
ATTITUDE MEASUREMENT SCALES-PSYCHOLOGYATTITUDE MEASUREMENT SCALES-PSYCHOLOGY
ATTITUDE MEASUREMENT SCALES-PSYCHOLOGYQuratulaintahir1
 
Psych Investigations Revision PowerPoint
Psych Investigations Revision PowerPointPsych Investigations Revision PowerPoint
Psych Investigations Revision PowerPointJamie Davies
 
PSYC 1301 Chapter 1
PSYC 1301 Chapter 1PSYC 1301 Chapter 1
PSYC 1301 Chapter 1tterrell01
 
Tools in Qualitative Research: Validity and Reliability
Tools in Qualitative Research: Validity and ReliabilityTools in Qualitative Research: Validity and Reliability
Tools in Qualitative Research: Validity and ReliabilityDr. Sarita Anand
 
Critical thinking ...
Critical thinking ...Critical thinking ...
Critical thinking ...Vasudevan BK
 
Critical thinking measurement tools and assessment rubrics report
Critical thinking measurement tools and assessment rubrics   reportCritical thinking measurement tools and assessment rubrics   report
Critical thinking measurement tools and assessment rubrics reportLira Lei Ann Bondoc
 
Bmgt 311 chapter_8
Bmgt 311 chapter_8Bmgt 311 chapter_8
Bmgt 311 chapter_8Chris Lovett
 

Was ist angesagt? (17)

Data collection reliability
Data collection reliabilityData collection reliability
Data collection reliability
 
Qualitative analysis\slideshow
Qualitative analysis\slideshowQualitative analysis\slideshow
Qualitative analysis\slideshow
 
Session 2 2018
Session 2 2018Session 2 2018
Session 2 2018
 
Critical Thinking: Its Basic Elements
Critical Thinking: Its Basic ElementsCritical Thinking: Its Basic Elements
Critical Thinking: Its Basic Elements
 
Survey Instrument Validity OERdifferentiation.com
Survey Instrument Validity OERdifferentiation.comSurvey Instrument Validity OERdifferentiation.com
Survey Instrument Validity OERdifferentiation.com
 
David Hand:Trustworthiness of statistical analysis
David Hand:Trustworthiness of statistical analysisDavid Hand:Trustworthiness of statistical analysis
David Hand:Trustworthiness of statistical analysis
 
Researchmethodology
ResearchmethodologyResearchmethodology
Researchmethodology
 
Likert Scale
Likert ScaleLikert Scale
Likert Scale
 
The Art and Science of Survey Research
The Art and Science of Survey ResearchThe Art and Science of Survey Research
The Art and Science of Survey Research
 
ATTITUDE MEASUREMENT SCALES-PSYCHOLOGY
ATTITUDE MEASUREMENT SCALES-PSYCHOLOGYATTITUDE MEASUREMENT SCALES-PSYCHOLOGY
ATTITUDE MEASUREMENT SCALES-PSYCHOLOGY
 
Instrumentation
InstrumentationInstrumentation
Instrumentation
 
Psych Investigations Revision PowerPoint
Psych Investigations Revision PowerPointPsych Investigations Revision PowerPoint
Psych Investigations Revision PowerPoint
 
PSYC 1301 Chapter 1
PSYC 1301 Chapter 1PSYC 1301 Chapter 1
PSYC 1301 Chapter 1
 
Tools in Qualitative Research: Validity and Reliability
Tools in Qualitative Research: Validity and ReliabilityTools in Qualitative Research: Validity and Reliability
Tools in Qualitative Research: Validity and Reliability
 
Critical thinking ...
Critical thinking ...Critical thinking ...
Critical thinking ...
 
Critical thinking measurement tools and assessment rubrics report
Critical thinking measurement tools and assessment rubrics   reportCritical thinking measurement tools and assessment rubrics   report
Critical thinking measurement tools and assessment rubrics report
 
Bmgt 311 chapter_8
Bmgt 311 chapter_8Bmgt 311 chapter_8
Bmgt 311 chapter_8
 

Ähnlich wie How to develop self reported scale ppt

Reliability & Validity
Reliability & ValidityReliability & Validity
Reliability & ValidityIkbal Ahmed
 
Scale development -- Steps
Scale development -- StepsScale development -- Steps
Scale development -- StepsKhalid Mahmood
 
Laos Session 2: Introduction to Modern Assessment Theory (continued) (EN)
Laos Session 2:  Introduction to Modern Assessment Theory (continued) (EN)Laos Session 2:  Introduction to Modern Assessment Theory (continued) (EN)
Laos Session 2: Introduction to Modern Assessment Theory (continued) (EN)NEQMAP
 
Validity.pptx
Validity.pptxValidity.pptx
Validity.pptxrupasi13
 
Validity
ValidityValidity
Validitymanish740
 
validity of scale.pdf
validity of scale.pdfvalidity of scale.pdf
validity of scale.pdfSrijoniChaki
 
Research methodology
Research methodologyResearch methodology
Research methodologysanjay agnihotri
 
Non Standarized Tests (Using Nursing Approch).pptx
Non Standarized Tests (Using Nursing Approch).pptxNon Standarized Tests (Using Nursing Approch).pptx
Non Standarized Tests (Using Nursing Approch).pptxvirengeeta
 
eeMba ii rm unit-3.1 measurement & scaling a
eeMba ii rm unit-3.1 measurement & scaling aeeMba ii rm unit-3.1 measurement & scaling a
eeMba ii rm unit-3.1 measurement & scaling aRai University
 
Week 3 additional information to help with interactive
Week 3 additional information to help with interactiveWeek 3 additional information to help with interactive
Week 3 additional information to help with interactiveLinda Foster
 
Indexes scales and typologies
Indexes scales and typologiesIndexes scales and typologies
Indexes scales and typologiesmaryjune Jardeleza
 
Attitude scales
Attitude scalesAttitude scales
Attitude scalesKeshav Sharma
 
Measurement scaling
Measurement   scalingMeasurement   scaling
Measurement scalingrajalakshmi7409
 
22.10.17 instrumentation and data collection
22.10.17 instrumentation and data collection22.10.17 instrumentation and data collection
22.10.17 instrumentation and data collectionAdibah Latif
 
Research Methodology Chapter 3
Research Methodology Chapter 3Research Methodology Chapter 3
Research Methodology Chapter 3Pulchowk Campus
 
research.pptx
research.pptxresearch.pptx
research.pptxssuser9843b7
 
Measurement, scaling and sampling
Measurement, scaling and samplingMeasurement, scaling and sampling
Measurement, scaling and samplingRajThakuri
 
Ssr test construction admin and scoring
Ssr test construction admin and scoringSsr test construction admin and scoring
Ssr test construction admin and scoringVictor Jr. Bactol
 

Ähnlich wie How to develop self reported scale ppt (20)

Reliability & Validity
Reliability & ValidityReliability & Validity
Reliability & Validity
 
Aiou Code 837
Aiou Code 837Aiou Code 837
Aiou Code 837
 
Scale development -- Steps
Scale development -- StepsScale development -- Steps
Scale development -- Steps
 
Laos Session 2: Introduction to Modern Assessment Theory (continued) (EN)
Laos Session 2:  Introduction to Modern Assessment Theory (continued) (EN)Laos Session 2:  Introduction to Modern Assessment Theory (continued) (EN)
Laos Session 2: Introduction to Modern Assessment Theory (continued) (EN)
 
Validity.pptx
Validity.pptxValidity.pptx
Validity.pptx
 
Data collection
Data collectionData collection
Data collection
 
Validity
ValidityValidity
Validity
 
validity of scale.pdf
validity of scale.pdfvalidity of scale.pdf
validity of scale.pdf
 
Research methodology
Research methodologyResearch methodology
Research methodology
 
Non Standarized Tests (Using Nursing Approch).pptx
Non Standarized Tests (Using Nursing Approch).pptxNon Standarized Tests (Using Nursing Approch).pptx
Non Standarized Tests (Using Nursing Approch).pptx
 
eeMba ii rm unit-3.1 measurement & scaling a
eeMba ii rm unit-3.1 measurement & scaling aeeMba ii rm unit-3.1 measurement & scaling a
eeMba ii rm unit-3.1 measurement & scaling a
 
Week 3 additional information to help with interactive
Week 3 additional information to help with interactiveWeek 3 additional information to help with interactive
Week 3 additional information to help with interactive
 
Indexes scales and typologies
Indexes scales and typologiesIndexes scales and typologies
Indexes scales and typologies
 
Attitude scales
Attitude scalesAttitude scales
Attitude scales
 
Measurement scaling
Measurement   scalingMeasurement   scaling
Measurement scaling
 
22.10.17 instrumentation and data collection
22.10.17 instrumentation and data collection22.10.17 instrumentation and data collection
22.10.17 instrumentation and data collection
 
Research Methodology Chapter 3
Research Methodology Chapter 3Research Methodology Chapter 3
Research Methodology Chapter 3
 
research.pptx
research.pptxresearch.pptx
research.pptx
 
Measurement, scaling and sampling
Measurement, scaling and samplingMeasurement, scaling and sampling
Measurement, scaling and sampling
 
Ssr test construction admin and scoring
Ssr test construction admin and scoringSsr test construction admin and scoring
Ssr test construction admin and scoring
 

Mehr von Mahesh Chand

IV cannulation
IV cannulationIV cannulation
IV cannulationMahesh Chand
 
How to develop self reported scale
How to develop self reported scaleHow to develop self reported scale
How to develop self reported scaleMahesh Chand
 
Intercostal drainage tube insertion
Intercostal drainage tube insertionIntercostal drainage tube insertion
Intercostal drainage tube insertionMahesh Chand
 
Care of patient in a hospital settings
Care of patient in a hospital settingsCare of patient in a hospital settings
Care of patient in a hospital settingsMahesh Chand
 
Legal and ethical issues in icu
Legal and ethical issues in icuLegal and ethical issues in icu
Legal and ethical issues in icuMahesh Chand
 
Infection control in intensive care unit
Infection control in intensive care unitInfection control in intensive care unit
Infection control in intensive care unitMahesh Chand
 
Burn first aid ppt
Burn first aid pptBurn first aid ppt
Burn first aid pptMahesh Chand
 
Anesthesia ppt
Anesthesia pptAnesthesia ppt
Anesthesia pptMahesh Chand
 
Legal aspects in OT
Legal aspects in OTLegal aspects in OT
Legal aspects in OTMahesh Chand
 
Sterilization and disinfection
Sterilization and disinfectionSterilization and disinfection
Sterilization and disinfectionMahesh Chand
 
Pain management techniques
Pain management techniquesPain management techniques
Pain management techniquesMahesh Chand
 
Cardiac assessment
Cardiac assessmentCardiac assessment
Cardiac assessmentMahesh Chand
 
Organization and functions of nursing srvices and education
Organization and functions of nursing srvices and educationOrganization and functions of nursing srvices and education
Organization and functions of nursing srvices and educationMahesh Chand
 
Thyroidstorm ppt
Thyroidstorm pptThyroidstorm ppt
Thyroidstorm pptMahesh Chand
 
Acute respiratory failure ppt
Acute respiratory failure pptAcute respiratory failure ppt
Acute respiratory failure pptMahesh Chand
 
Pulmonary tuberculosis
Pulmonary tuberculosisPulmonary tuberculosis
Pulmonary tuberculosisMahesh Chand
 
Chronic Obstructive pulmonary diasese
Chronic Obstructive pulmonary diaseseChronic Obstructive pulmonary diasese
Chronic Obstructive pulmonary diaseseMahesh Chand
 
Pleural effusion
Pleural effusionPleural effusion
Pleural effusionMahesh Chand
 
Acute pyelonephritis
Acute pyelonephritisAcute pyelonephritis
Acute pyelonephritisMahesh Chand
 

Mehr von Mahesh Chand (20)

IV cannulation
IV cannulationIV cannulation
IV cannulation
 
How to develop self reported scale
How to develop self reported scaleHow to develop self reported scale
How to develop self reported scale
 
Intercostal drainage tube insertion
Intercostal drainage tube insertionIntercostal drainage tube insertion
Intercostal drainage tube insertion
 
Care of patient in a hospital settings
Care of patient in a hospital settingsCare of patient in a hospital settings
Care of patient in a hospital settings
 
Legal and ethical issues in icu
Legal and ethical issues in icuLegal and ethical issues in icu
Legal and ethical issues in icu
 
Infection control in intensive care unit
Infection control in intensive care unitInfection control in intensive care unit
Infection control in intensive care unit
 
Burn first aid ppt
Burn first aid pptBurn first aid ppt
Burn first aid ppt
 
Anesthesia ppt
Anesthesia pptAnesthesia ppt
Anesthesia ppt
 
Fumigation
FumigationFumigation
Fumigation
 
Legal aspects in OT
Legal aspects in OTLegal aspects in OT
Legal aspects in OT
 
Sterilization and disinfection
Sterilization and disinfectionSterilization and disinfection
Sterilization and disinfection
 
Pain management techniques
Pain management techniquesPain management techniques
Pain management techniques
 
Cardiac assessment
Cardiac assessmentCardiac assessment
Cardiac assessment
 
Organization and functions of nursing srvices and education
Organization and functions of nursing srvices and educationOrganization and functions of nursing srvices and education
Organization and functions of nursing srvices and education
 
Thyroidstorm ppt
Thyroidstorm pptThyroidstorm ppt
Thyroidstorm ppt
 
Acute respiratory failure ppt
Acute respiratory failure pptAcute respiratory failure ppt
Acute respiratory failure ppt
 
Pulmonary tuberculosis
Pulmonary tuberculosisPulmonary tuberculosis
Pulmonary tuberculosis
 
Chronic Obstructive pulmonary diasese
Chronic Obstructive pulmonary diaseseChronic Obstructive pulmonary diasese
Chronic Obstructive pulmonary diasese
 
Pleural effusion
Pleural effusionPleural effusion
Pleural effusion
 
Acute pyelonephritis
Acute pyelonephritisAcute pyelonephritis
Acute pyelonephritis
 

KĂźrzlich hochgeladen

What is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPWhat is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPCeline George
 
Q4 English4 Week3 PPT Melcnmg-based.pptx
Q4 English4 Week3 PPT Melcnmg-based.pptxQ4 English4 Week3 PPT Melcnmg-based.pptx
Q4 English4 Week3 PPT Melcnmg-based.pptxnelietumpap1
 
ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4MiaBumagat1
 
How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPCeline George
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for BeginnersSabitha Banu
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Celine George
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
Science 7 Quarter 4 Module 2: Natural Resources.pptx
Science 7 Quarter 4 Module 2: Natural Resources.pptxScience 7 Quarter 4 Module 2: Natural Resources.pptx
Science 7 Quarter 4 Module 2: Natural Resources.pptxMaryGraceBautista27
 
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxMULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxAnupkumar Sharma
 
Choosing the Right CBSE School A Comprehensive Guide for Parents
Choosing the Right CBSE School A Comprehensive Guide for ParentsChoosing the Right CBSE School A Comprehensive Guide for Parents
Choosing the Right CBSE School A Comprehensive Guide for Parentsnavabharathschool99
 
THEORIES OF ORGANIZATION-PUBLIC ADMINISTRATION
THEORIES OF ORGANIZATION-PUBLIC ADMINISTRATIONTHEORIES OF ORGANIZATION-PUBLIC ADMINISTRATION
THEORIES OF ORGANIZATION-PUBLIC ADMINISTRATIONHumphrey A BeĂąa
 
DATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersDATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersSabitha Banu
 
Grade 9 Q4-MELC1-Active and Passive Voice.pptx
Grade 9 Q4-MELC1-Active and Passive Voice.pptxGrade 9 Q4-MELC1-Active and Passive Voice.pptx
Grade 9 Q4-MELC1-Active and Passive Voice.pptxChelloAnnAsuncion2
 
ENGLISH6-Q4-W3.pptxqurter our high choom
ENGLISH6-Q4-W3.pptxqurter our high choomENGLISH6-Q4-W3.pptxqurter our high choom
ENGLISH6-Q4-W3.pptxqurter our high choomnelietumpap1
 
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYKayeClaireEstoconing
 
4.18.24 Movement Legacies, Reflection, and Review.pptx
4.18.24 Movement Legacies, Reflection, and Review.pptx4.18.24 Movement Legacies, Reflection, and Review.pptx
4.18.24 Movement Legacies, Reflection, and Review.pptxmary850239
 

KĂźrzlich hochgeladen (20)

What is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPWhat is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERP
 
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
Q4 English4 Week3 PPT Melcnmg-based.pptx
Q4 English4 Week3 PPT Melcnmg-based.pptxQ4 English4 Week3 PPT Melcnmg-based.pptx
Q4 English4 Week3 PPT Melcnmg-based.pptx
 
ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4
 
How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERP
 
Raw materials used in Herbal Cosmetics.pptx
Raw materials used in Herbal Cosmetics.pptxRaw materials used in Herbal Cosmetics.pptx
Raw materials used in Herbal Cosmetics.pptx
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for Beginners
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
 
FINALS_OF_LEFT_ON_C'N_EL_DORADO_2024.pptx
FINALS_OF_LEFT_ON_C'N_EL_DORADO_2024.pptxFINALS_OF_LEFT_ON_C'N_EL_DORADO_2024.pptx
FINALS_OF_LEFT_ON_C'N_EL_DORADO_2024.pptx
 
Science 7 Quarter 4 Module 2: Natural Resources.pptx
Science 7 Quarter 4 Module 2: Natural Resources.pptxScience 7 Quarter 4 Module 2: Natural Resources.pptx
Science 7 Quarter 4 Module 2: Natural Resources.pptx
 
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxMULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
 
Choosing the Right CBSE School A Comprehensive Guide for Parents
Choosing the Right CBSE School A Comprehensive Guide for ParentsChoosing the Right CBSE School A Comprehensive Guide for Parents
Choosing the Right CBSE School A Comprehensive Guide for Parents
 
THEORIES OF ORGANIZATION-PUBLIC ADMINISTRATION
THEORIES OF ORGANIZATION-PUBLIC ADMINISTRATIONTHEORIES OF ORGANIZATION-PUBLIC ADMINISTRATION
THEORIES OF ORGANIZATION-PUBLIC ADMINISTRATION
 
DATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersDATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginners
 
Grade 9 Q4-MELC1-Active and Passive Voice.pptx
Grade 9 Q4-MELC1-Active and Passive Voice.pptxGrade 9 Q4-MELC1-Active and Passive Voice.pptx
Grade 9 Q4-MELC1-Active and Passive Voice.pptx
 
ENGLISH6-Q4-W3.pptxqurter our high choom
ENGLISH6-Q4-W3.pptxqurter our high choomENGLISH6-Q4-W3.pptxqurter our high choom
ENGLISH6-Q4-W3.pptxqurter our high choom
 
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
 
4.18.24 Movement Legacies, Reflection, and Review.pptx
4.18.24 Movement Legacies, Reflection, and Review.pptx4.18.24 Movement Legacies, Reflection, and Review.pptx
4.18.24 Movement Legacies, Reflection, and Review.pptx
 

How to develop self reported scale ppt

  • 1. HOW TO DEVELOP SELF REPORTED SCALE Presenter: Mahesh Chand Nursing Tutor
  • 2. Introduction Researchers sometimes are unable to identify an appropriate instrument to operationalize a construct. This may occur when the construct is new, but often it is due to limitations of existing instruments. Because this situation occurs fairly often, this chapter provides an overview of the steps involved in developing high-quality self- report scales.
  • 3. Steps 1. Conceptualization and item generation 2. Preliminary evaluation of items 3. Field testing the instruments 4. Analysis of scale development data 5. Scale refinement and validation 6. Interpretation of scale score 7. Critiquing
  • 4.  Conceptualization the construct  Deciding the type of scale  Generating an item pool  Making decision about item features  Wording the items CONCEPTUALIZATION AND ITEM GENERATION
  • 5. You will not be able to quantify an attribute adequately unless you thoroughly understand the underlying construct you wish to capture. • Underlying construct: You cannot develop items to produce the right score, and you cannot expect good content and construct validity, if you are unclear about the construct. Thus, the first step in scale development is to become an expert on the construct. • Complex construct: Complex constructs have a number of different facets or dimensions, and it is important to identify and understand each one. Conceptualizing the Construct
  • 6. Deciding on the Type of Scale • Before items can be generated, you need to decide on the type of scale you wish to create because item characteristics vary by scale type. E.g. Traditional summated rating (Likert type) scale
  • 7. Generating an Item Pool • An early step in scale construction is to develop a pool of possible items for the scale. • Items—which collectively constitute the operational definition of the construct—need to be carefully crafted to reflect the latent variable they are designed to measure. • Here are some possible sources for generating an item pool- i. Existing instruments ii. The literature iii. Concept analysis iv. In-depth qualitative research v. Clinical observations
  • 8. Making Decisions about Item Features You need to make decisions about such issues as the number of items to develop, the number and form of the response options, whether to include positively and negatively worded items, and how to deal with time. • Number of Items: There is no magic formula for how many items should be developed, but our advice is to generate a very large pool of items. As you proceed, many items will be discarded. Longer scales tend to be more reliable, so starting with a large number of items. E.g.: In your final scale (e.g., 30 to 40 items for a 10- item scale), but the minimum should be 50% more (e.g., 15 items for a 10-item scale).
  • 9. • Response Options: Scale items involve both a stem (declarative statement) and response options. Likert scales involve response options on a continuum of agreement, such as frequency (never/always), importance (very important /unimportant), quality (excellent/very poor), and likelihood (highly likely/impossible). The midpoint can be labeled with such phrases as “neither agree nor disagree,” “undecided,” “agree and disagree equally,” or simply “?”. Strongly disagree, disagree, undecided, agree, strongly agree Never, almost never, sometimes, often, almost always Very important, important, somewhat important, of little importance, unimportant
  • 10. • Positive and Negative Stems: Psychometricians advised scale developers to deliberately include both positively and negatively worded statements and to reverse score negative items. E.g. Two items for a scale of depression: “I frequently feel sad,” and “I don’t often feel sad.” Research suggests that acceptance can be minimized by putting the most positive response options (e.g. strongly agree) at the end of the list rather than at the beginning.
  • 11. • Item Intensity: The intensity of the statements should be similar and fairly strongly worded. If items are worded such that almost anyone would agree with them. • For E.g. “Good health is important” would generate almost universal agreement
  • 12. • Item Time Frames: A time frame should not emerge as a consequence of item development. You should decide in advance, based on your conceptual understanding of the construct and the needs for which the scale is being constructed, how to deal with time. • 2-week period was chosen because it parallels the duration of symptoms required for a diagnosis of major depressive episode. Postpartum Depression Screening Scale asks respondents to indicate their emotional state in the previous 2 weeks— for example, over the last 2 weeks I: “ . . . had trouble sleeping even when my baby was asleep” and “ . . . felt like a failure as a mother”
  • 13. Wording the Items • Items should be worded in such a manner that every respondent is answering the same question. a) Clarity b) Jargon (technical or special words) c) Length d) Double negatives e) Double-barelled items.(Insects and Snakes)
  • 14. PRELIMINARY EVALUATION OF ITEMS Preliminary evaluation of the items can be made under following: • Internal review • Input for target population • External review by expert • Internal Review: Once a large item pool has been generated, it is time for critical appraisal. It is vital importance to assess the scale’s readability, unless the scale is intended for a highly educated population. There are different approaches for assessing the reading level of written documents.
  • 15. Set A Set B • I am frequently exhausted. I am often tired. • I invariably get insufficient sleep. I don’t get enough sleep. Set A have a Flesch-Kincaid grade level of 12.0 and a Flesch reading ease score of 4.8. Set B, by contrast, has a grade level of 1.8 and a reading ease score of 89.4. • It is clear from the foregoing analysis that the second set of items would be superior for a population that includes people with limited education.
  • 16. • Input from the Target Population: In the next step, the initial pool of items is pretested. a small sample of people (20 to 40 or so) representing the target population is invited to complete the items. • In cognitive interviews, people are asked to reflect upon their interpretation of the items and their answers so that the underlying process of response selection is better understood. There are two basic approaches to cognitive interviewing-
  • 17. • The think-aloud method, wherein respondents are asked to explain step by step how they processed the question and arrived at an answer. • A second approach is to conduct an interview in which the interviewer uses a series of targeted probes that encourage reflection about underlying cognitive processes. • As an alternative or supplement to pretests, focus groups can also be used at this stage in scale development.
  • 18. External Review by Experts: External review of the revised items by a panel of experts should be undertaken to assess the scale’s content validity. It is advisable to undertake two rounds of review.  First to refine faulty items or to add new items to cover the domain adequately.  Second to formally assess the content validity of the items and scale. It can be done under following steps-
  • 19. • Selecting and Recruiting the Experts: In the initial phase of a two-part review, we advise having an expert panel of 8 to 12 members, with a good mix in terms of roles and disciplines. • The second panel for formally assessing the content validity of a more refined set of items should consist of three to five experts in the content area. E.g.: For a scale designed to measure fear of dying in the elderly, the experts might include nurses, gerontologists, and psychiatrists.
  • 20. • Preliminary Expert Review: Content Validation of Items: The experts’ job is to evaluate individual items and the overall scale using guidelines established by the scale developer. The first panel of experts is usually invited to rate each item in terms of several characteristics, such as clarity of wording; relevance of the item to the construct; and appropriateness for the target population. The relevance is most often rated as follows: 1 = not relevant, 2 = somewhat relevant, 3 = quite relevant, 4 = highly relevant.
  • 21. • The scale items shown below have been developed to measure one dimension of the construct of safe sexual behaviors among adolescents Item s Not relevant Somewhat relevant Quite relevant Highly relevant 1 I asked my partner about his/her sexual history before having intercourse 1 2 3 4 2 I don’t have sex without asking the person if he/she has been tested for HIV/AIDS 1 2 3 4 3 When I am having sex with someone for the first time insist that we use a condom 1 2 3 4 An expert panel should be asked to consider whether the items taken as a whole adequately cover the construct domain.
  • 22. Standard method for computing an item-level content validity index (I-CVI) = Number giving a rating of 3 or 4 on the 4-point scale --------------------------------------------------------- Divided by the number of experts. For example, if five experts rated an item as 3 and one rated the item as 2, the I-CVI would be .83. This means that there must be 100% agreement among raters when there are four or fewer experts. When there are five to eight experts, one rating of “not relevant” can be tolerated. Items with lower than desired I-CVIs need careful scrutiny.
  • 23. • Content Validation of the Scale: In the second round of content validation, a smaller group of experts (three to five) can be used to evaluate the relevance of the revised set of items and to compute the scale content validity. • First, experts who rated every item as “highly relevant” (or “not relevant”) may not be sufficiently discriminating. • Second, an expert who gave high ratings to items that were judged by most others to not be relevant might be problematic. • Third, the proportion of items judged relevant should be computed for all judges.
  • 24. • For example, if an expert rated 8 out of 10 items as relevant, the proportion for that judge would be .80. The pattern across experts can be examined for “outliers.” If the average proportion across raters is, for example, .80, you might consider not inviting back for a second round of experts whose average proportion was either very low (e.g., .50) or very high (e.g., 1.0). • On a 10- item scale, for example, if the I-CVIs for 5 items were .80 and the I-CVIs for the remaining 5 items were 1.00, then the S-CVI/Ave would be .90. An S-CVI/Ave of .90 or higher is desirable.
  • 25. FIELD TESTING THE INSTRUMENT At this point, you will have whittled down and refined your items based on your own and others’ careful scrutiny. It can be done under- • Developing sampling plan • Developing a data collection plan • Preparing for data collection
  • 26. • Developing a Sampling Plan: The sample for testing the scale should be representative of the population for whom the scale has been devised and should be large enough to support complex analyses. • Recommendations range from 3-4 people per item to 40-50 per item, with 10 per item often being recommended. That means that if you have 20 items, your sample should be at least 200.
  • 27. • Developing a Data Collection Plan: Decisions have to be made concerning how to administer the instrument (e.g., by mailed or distributed questionnaires, over the Internet) and what to include in the instrument. • The instrument should include the scale items and basic demographic information.
  • 28. • Preparing for Data Collection: As in all data collection efforts, care should be taken to make the instrument attractive, professional looking, and easy to understand. Colleagues, mentors, or family members should be asked to evaluate the appearance of the instrument before it is reproduced. • Instructions for completing the instrument should be clear, and a readability assessment of the instructions is useful. There should be no ambiguity about what is expected of respondents.
  • 29. ANALYSIS OF SCALE DEVELOPMENT DATA Analysis of developed data can be done under following steps: • Basic item analysis • Exploratory factor analysis • Internal consistency analysis • Test – retest reliability analysis
  • 30. • Basic Item Analysis: Inter-item correlations between .30 and .70 are often recommended, with correlations lower than .30 suggesting little compatibility with the underlying construct and ones higher than .70 suggesting over no helpful. However, the evaluation depends on the number of items in the scale. An average inter-item correlation of .57 is needed to achieve a coefficient alpha of .80 on a 3-item scale, but an average of only .29 is needed for a 10-item scale.
  • 31. • Exploratory Factor Analysis: Factor analysis complex interrelationships among items and identifies items that “go together” as unified concepts. E.g.: 50 items measuring women’s attitudes toward menopause. We could form a scale by adding together scores from several individual items, but which items should be combined? Would it be reasonable to combine all 50 items? Probably not, because the 50 items are not all tapping the same thing—there are various dimensions to women’s attitude toward menopause. One dimension may relate to aging, reproductive ability, sexuality, avoidance of monthly menstruation. Women’s attitude on one dimension may be independent of their attitude on another.
  • 32. • Internal Consistency Analysis: After a final set of items is selected based on the item analysis and factor analysis, an analysis should be undertaken to calculate coefficient alpha. Alpha, it may be recalled, provides an estimate of a key measurement property of a multi-item scale, internal consistency. They also calculate the value of coefficient alpha for the scale and for a hypothetical scale with each individual item removed. Sometimes, removal of a faulty item actually increases alpha.
  • 33. • Test–Retest Reliability Analysis: Test–retest reliability a particularly important indicator of a scale’s quality. When the time interval is too brief, carryover effects can lead to artificially high estimates of reliability. • Some experts advise that the time interval between measurements should be around 1 to 2 weeks.
  • 34. SCALE REFINEMENT AND VALIDATION Scale refinement and validation can be made under following steps: • Revising the scale • Scoring the scale • Conducting a validation study
  • 35. • Revising the Scale: The analyses undertaken in the development study often suggest the need to revise or add items. For example, if subscale alpha coefficients are lower than .80 or so, consideration should be given to adding items for subsequent testing. • Before deciding that your scale is finalized, it is a good idea to examine the content of the items in the scale.
  • 36. • Scoring the Scale: Scoring the scale is often easy with Likert type items: items scores are typically just added together with reverse scoring of items. A system of weighting the items might be considered. E.g. A scale to assess a person’s risk of a disease or condition( cardiovascular disease) might benefit from weighting some items( high BP) more heavily than others.
  • 37. • Conducting a validation study: Designing a validation study entails much of the same issues (and advice) as designing a development study, in terms of sample composition, sample size, and data collection strategies. Thus, we focus here on analyses undertaken in a validation study.
  • 38. • Confirmatory factory analysis: CFA involves testing a measurement model, which specifies the hypothesized relationships among underlying latent and manifest variables. The measurement model specifies the hypothesized factor structure. The latent variables provide a method for evaluating relationships between observed variables (the items) and unobserved variables(factors)
  • 39. 1 2 3 4 5 Physical Fatigue Mental Fatigue 6 7 8 9 10 e1 e2 e3 e4 e5 e6 e7 e8 e9 e10 The analysis would also indicate whether the overall model fit is good, based on a goodness of fit statistic. If the hypothesized model is not a good fit to the data, the measurement model could be re-specified and retested.
  • 40. INTERPRETATION OF SCALE SCORES After the scale refinement and validation next steps comes to interpretation of score. It is done in form of: • Percentile • Standard score • Norms • Cutoff marks
  • 41. • Percentiles: A percentile indicates the percentage of people who score below a particular score value. Percentiles can range from the 0th to the 99th percentile, and the 50th percentile corresponds to the median. Percentile values are most useful when they are determined on the basis of a large, representative sample.
  • 42. • Standard scores: – Standard scores are expressed in terms of their relative distance from the mean, in standard deviation units. – A standard score of 0.0 corresponds to a raw score exactly at the scale’s mean. – A standard score of 1.0 corresponds to a score 1 SD above the mean, and −1.0 corresponds to below the mean. – Standard scores can be readily calculated from raw scores once the mean and SD have been computed.
  • 43. • Norms: – Norms are often established for key demographic characteristics, such as age and gender. – Norms are often expressed in terms of percentiles. For example, an adult male with a score of 72 on the scale might be at the 80th percentile, but a female with the same score might be at the 85th percentile. • Guidelines for norming instruments have been developed by Nunnally and Bernstein(1994).
  • 44. • Cutoff Points: Interpretation of scores can often be facilitated when the instrument developer establishes cut points for classification purposes. Cut points are defined in terms of percentiles E.g.: Children’s weights, those below the 5th percentile are often interpreted as underweight whereas those above the 95th percentile are considered overweight.
  • 45. CRITIQUING SCALE DEVELOPMENT STUDIES • Articles on scale development appear regularly in many nursing journals. If you are planning to use a scale in a substantive study, you should carefully review the methods used to construct the scale and test its measurement properties. You should scrutinize whether the evidence regarding the scale’s psychometric adequacy is sufficiently sound to merit its use.
  • 46. Guidelines for Critiquing Scale Development and Assessment Reports • Does the report offer a clear definition of the construct? Did it provide sufficient context for the study through a summary of the literature and discussion of relevant theory? Is the population for whom the scale intended adequately described? • Does the report indicate how items were generated? Do the procedures seem sound? Was information provided about the reading level of scale items? • Does the report describe content validation efforts, and was the description thorough? Is there evidence of good content validity?
  • 47. • Were appropriate efforts made to refine the scale (e.g., through pretests, item analysis)? • Was the development or validation sample of participants appropriate in terms of representativeness, size, and heterogeneity? • Was factor analysis used to examine or validate the scale’s dimensionality? If yes, did the report offer evidence to support the factor structure and the naming of factors? • Were appropriate methods used to assess the scale’s internal consistency and reliability? Were estimates of reliability and internal consistency sufficiently high?
  • 48. References • Denise F Polit and Beck, Nursing research 10th edition, Wolters Kluwer publication, page No. 595-637, 499-505, 505-507 • KP Neeraja Textbook of Nursing Education, 1st edition, Jaypee publication page No. 432-440. • Polit, D. F. (2014). Getting serious about test- retest reliability: A critique of retest research and some recommendations. Quality of Life Research, 1713–1720. • Suresh K Sharma Nursing research and statistics, 2nd edition .for reference