Measuring Variables2

Embed Size (px)

Citation preview

  • 8/10/2019 Measuring Variables2

    1/24

    Defining and Measuring

    Variables

  • 8/10/2019 Measuring Variables2

    2/24

    1. An overview of measurement

    two aspects of measurement are particularlyimportant in planning a research study orreading a research report:

    often there is not a one-to-one relationship betweenthe variable measured and the measurementobtained (knowledge, performance and exam grade)

    there are usually several different options formeasuring any particular variable (types of exams

    and questions on exams) Direct measurement (height, weight) vs indirectmeasurement (motivation, knowledge, memory,marital satisfaction)

  • 8/10/2019 Measuring Variables2

    3/24

    2. Constructs and operationaldefinitions

    Theories summarize our observations, explainmechanisms underlying a particular behavior andmake predictions about the behavior.

    many research variables, particularly variables ofinterest to behavioral scientists, are hypotheticalattributes or mechanisms explaining and predictingsome behavior in a theory are called constructs

    externalstimulus construct behaviorfactor

    reward motivation performance

  • 8/10/2019 Measuring Variables2

    4/24

    constructs can not be directly observed or measured however, researchers can measure external, observable

    events as an indirect method of measuring the constructitself operational definition

    is a procedure for measuring and defining a construct, indirectmethod of measuring something that can not be measured directly

    an operational definition specifies a measurement procedure formeasuring an external, observable behavior and uses the resultingmeasurements as a definition and a measurement of thehypothetical construct

    e.g. IQ test is an operational definition for the construct intelligence

    - provide and example of a theoretical construct and itsoperational definition

    You dont always have to come up with your ownoperational definition of the construct, you can use someconventional measurement procedure from previousstudies

  • 8/10/2019 Measuring Variables2

    5/24

  • 8/10/2019 Measuring Variables2

    6/24

    Validity of measurement

    Validity of measurement concerns the truth of the measurement

    it is the degree to which the measurementprocess measures the variable it claims tomeasure

    Is the IQ score truly measuring intelligence?What about size of the brain and bumps onthe scull?

  • 8/10/2019 Measuring Variables2

    7/24

    Different kinds of validity

    face validity

    the simplest and least scientific definition of validity it is demonstrated when a measure superficially appears to

    measure what it claims to measure

    Based on subjective judgment and difficult to quantify e.g. intelligence and reasoning questions on the IQ test Problem - participants can use the face validity to change

    their answers

    concurrent validity (criterion validity)

    is demonstrated when scores obtained from a new measureare directly related to scores obtained from a moreestablished measure of the same variable

    e.g. new IQ test correlates with an older IQ test

  • 8/10/2019 Measuring Variables2

    8/24

    Different kinds of validity (cont.)

    Different kinds of validity predictive validity

    when scores obtained from a measure accurately predict

    behavior according to a theory e.g. high scores on need for achievement test predict

    competitive behavior in children (ring toss game)

    construct validity is demonstrated when scores obtained from a measure are

    directly related to the variable itself Reflects how close the measure relates to the construct

    (height and weight example)

    in one sense, construct validity is achieved by repeatedlydemonstrating every other type of validity

  • 8/10/2019 Measuring Variables2

    9/24

    Different kinds of validity (cont.)

    Different kinds of validity convergent validity

    is demonstrated by a strong relationship between the scoresobtained from two different methods of measuring the same

    construct e.g. an experimenter observing aggressive behavior in childrencorrelated with teachers ratings of their behavior

    divergent validity is demonstrated by using two different methods to measure

    two different constructs convergent validity must be shown for each of the twoconstructs and little or no relationship exists between thescores obtained from the two different constructs when theyare measured by the same method

    e.g. aggressive behavior and general activity level in children

  • 8/10/2019 Measuring Variables2

    10/24

  • 8/10/2019 Measuring Variables2

    11/24

    Reliability of measurement

    Reliability of measurement a measurement procedure is said to be reliable if

    repeated measurements of the same individual underthe same conditions produce identical (or nearly

    identical) values reliability is the stability or the consistency of

    measurement

    measured score = true score + error

    IQ score = true IQ score + mood, fatigue etc.

  • 8/10/2019 Measuring Variables2

    12/24

    Reliability and error ofmeasurement

    Inconsistency (lack of reliability) of measurement comesfrom error

    The higher the error the more unreliable themeasurement

    Sources of error observer error

    the individual who makes the measurements can introduce simplehuman error into the measurement process

    environmental changes small changes in the environment from one measurement to another

    (e.g. time of the day, distraction in the room, lighting) participant changes

    participants change between measurements (mood, hunger,motivation)

  • 8/10/2019 Measuring Variables2

    13/24

    Types and measures of reliability

    successive measurements Obtaining scores from two successive measurements and calculating

    a correlation between them the same group, the same measurement at two different times test-retest reliability

    simultaneous measurements obtained by direct observation of behaviors (two or more separateobservers at the same time), consistency across raters

    inter-rater reliability

    internal consistency degree of consistency of scores from separate items on a test or

    questionnaire consisting of multiple items you want all the items or groups of items tapping the sameprocesses

    researchers commonly split the set of items in half, compute aseparate score of each half, and then evaluate the degree ofagreement between the two scores

    split-half reliability

  • 8/10/2019 Measuring Variables2

    14/24

    The relationship between reliability

    and validity

    they are partially related and partiallyindependent

    reliability is a prerequisite for validity

    (measurement procedure can not be validunless it is reliablee.g. IQ, huge variance ofrepeated measurements is impossible if weare truly measuring intelligence)

    it is not necessary for a measurement to bevalid for it to be reliable (e.g. height as ameasure of intelligence)

  • 8/10/2019 Measuring Variables2

    15/24

    4. Scales of measurement

    Scales define the type categories we use in measurementand the selection of a scale has direct impact on ourability to describe relationships between variables

    the nominal scale

    simply represents qualitative difference in the variable measured can only tell us that a difference exists without the possibility

    telling the direction or magnitude of the difference

    e.g. majors in college, race, gender, occupation

    the ordinal scale the categories that make up an ordinal scale form an ordered

    sequence

    can tell us the direction of the difference but not the magnitude

    e.g. coffee cup sizes, socioeconomic class, T-shirt sizes, foodpreferences

  • 8/10/2019 Measuring Variables2

    16/24

    Scales of measurement (cont.)

    the interval scale categories on an interval scale are organizedsequentially, and all categories are the same size

    we can determine the direction and the magnitude ofa difference

    May have an arbitrary zero (convenient point ofreference)

    e.g. temperature in Farenheit, time in seconds

    the ratio scale

    consists of equal, ordered categories anchored by azero point that is not arbitrary but meaningful(representing absence of a variable

    allows us to determine the direction, the magnitude,and the ratio of the difference

    e.g. reaction time, number of errors on a test

  • 8/10/2019 Measuring Variables2

    17/24

    5. Modalities of measurement

    One can measure a construct by selectinga measure from three main categories

    There are three basic modalities ofmeasurement: self-report

    physiological measurement

    behavioral measurement

    behavioral observation content analysis and archival research

  • 8/10/2019 Measuring Variables2

    18/24

    Self-report measures

    you ask a participant to describe his behavior,to express his opinion or characterize hisexperience in an interview or by using aquestionnaire with ratings

    Positive aspects Only the individual has direct access to informationabout his state of mind

    More direct measure

    Negative aspects Participants may distort the responses to create abetter self-image or to please the experimenter

    The response can also be influenced by wording ofthe questions and other aspects of the situation

  • 8/10/2019 Measuring Variables2

    19/24

    Physiological measures

    Physiological manifestations of the underlyingconstruct

    e.g. EEG, EKG, galvanic skin response,perspiration, PET, fMRI

    advantages provides accurate, reliable, and well-defined

    measurements that are not dependent on subjectiveinterpretation

    disadvantages equipment is usually expensive or unavailable Presence of monitoring devices may create unnatural

    situation

    question: Are these procedures a valid measure of theconstruct (e.g. increase in heart rate to fear, arousal)

  • 8/10/2019 Measuring Variables2

    20/24

    Behavioral measures

    behaviors that can be observed and measured (e.g.reaction time, reading speed, focus of attention,disruptive behavior, number of words recalled on a

    memory test) How to select the right behavioral measure? Depends on the purpose of the study

    In clinical setting the same disorder can reveal itself throughdifferent symptoms

    In studying memory we want to have the same measure for allsubjects to be able to compare them

    Beware of situational changes in behavior (e.g.disruptive behavior in school vs when observed) anddifferent behavioral indicators of a construct

  • 8/10/2019 Measuring Variables2

    21/24

    6. Other aspects of measurement

    multiple measures sometimes you can use two (or more) different

    procedures to measure the same variable (e.g. heartrate and questionnaire as a measure of fear)

    problems (the two variables may not behave in thesame way)

    e.g. a specific therapy for treating fear may have large effecton behavior but no effect on heart rate

    the lack of agreement between two measures is calleddesynchrony

    One measure can be more sensitive than other Different measures may indicate different dimensions of the

    variable and change at different times during the treatment

  • 8/10/2019 Measuring Variables2

    22/24

    Sensitivity and range effects

    are the measures sensitive enough to respond to thetype and magnitude of the changes that areexpected? (e.g. seconds vs. milliseconds, difficult oreasy exams)

    range effects a ceiling effect (the clustering of scores at the high end of a

    measurement scale, allowing little or no possibility ofincreases in value, e.g. test that is too easy)

    a floor effect (the clustering of scores at the low end of a

    measurement scale, allowing little or no possibility ofdecreases in value, e.g. test that is too difficult)

    Range effects are usually a consequence of using a measurethat is inappropriate for a particular group (e.g. 4-grade testfor college students)

  • 8/10/2019 Measuring Variables2

    23/24

    Participant reactivity andexperimenter bias

    participant reactivity is the way how participant reacts tothe experimental situation (e.g. overly cooperative, overlydefensive, or hostile) To avoid these problems one can try to disguise the true purpose of

    the experiment or observe individuals without their awareness

    (beware ethical issues) experimenter bias is the way experimenter influences

    results (e.g. by being warm and friendly with one group ofparticipants vs. cold and stern with other group)

    to avoid participant reactivity and experimenter bias we

    use: standardized procedures (e.g. instructions recorded on a tape) a research study is single blind if the researcher does not know the

    predicted outcome a research study is double blind if both the researcher and the

    participants are unaware of the predicted outcome

  • 8/10/2019 Measuring Variables2

    24/24

    Participant reactivity andexperimenter bias

    to avoid participant reactivity andexperimenter bias we use blind experiments

    a research study is single blind if the researcherdoes not know the predicted outcome

    a research study is double blind if both theresearcher and the participants are unaware of thepredicted outcome