20
Ignorance or Integration: The Cognitive Processes Underlying Choice Behavior SHAHAR AYAL 1,y and GUY HOCHMAN 2y 1 Center for Behavioral Economics, Fuqua School of Business, Duke University, USA 2 Max Wertheimer Minerva Center for Cognitive Studies, Faculty of Industrial Engineering and Management,The Technion Ɇ Israel Institute of Technology, Haifa, Israel ABSTRACT The fast-and-frugal heuristic framework assumes noncompensatory tools for human preferences (i.e., priority heuristic) and inferences (i.e., take the best heuristic). According to this framework, these heuristics predict choice behavior as well as model the cognitive processes underlying such behavior. The current paper presents two studies that juxtapose predictions derived from these two simple heuristics with alternative predictions derived from compensatory principles. Dependent measures that included reaction time, choice pattern, confidence level, and accuracy were better predicted by compensatory indices than by noncompensatory indices. These findings suggest that people do not rely on limited arguments only, but tend to integrate all acquired information into their choice processes. This tendency was replicated even when the experimental task facilitated the use of noncompensatory principles. We argue that the fast and frugal heuristics can predict the final outcome only under certain conditions, but even in these particular situations they are not applicable to the processes underlying choice behavior. An integrative model for choice behavior is proposed that better represents the data. Copyright # 2009 John Wiley & Sons, Ltd. key words compensatory models; noncompensatory models; take the best heuristic; priority heuristic; choice behavior INTRODUCTION Decision-making models have been traditionally classified into compensatory versus noncompensatory theoretical frameworks (e.g., Einhorn, 1971; Elrod, Johnson, & White, 2004; Ford, Schmitt, Schechtman, Hults, & Doherty, 1989; Payne, Bettman, & Johnson, 1993). Generally speaking, compensatory models (e.g., Expected Utility Theory, by von Neumann & Morgenstern, 1944; Prospect Theory by Kahneman & Tversky, 1979; Tversky & Kahneman, 1992) assume that choices are governed by tradeoffs between conflicting Journal of Behavioral Decision Making J. Behav. Dec. Making, 22: 455–474 (2009) Published online 12 March 2009 in Wiley InterScience (www.interscience.wiley.com) DOI: 10.1002/bdm.642 *Correspondence to: Shahar Ayal, Duke University, Fuqua School of Business, 1 Towerview Drive, Durham, NC 27708, USA. E-mail: [email protected] y Both authors contributed equally to the article. Copyright # 2009 John Wiley & Sons, Ltd.

Ignorance or integration: the cognitive processes underlying choice behavior

  • Upload
    idc-ac

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Journal of Behavioral Decision Making

J. Behav. Dec. Making, 22: 455–474 (2009)

Published online 12 March 2009 in Wiley InterScience

(www.interscience.wiley.com) DOI: 10.1002/bdm.642

*Correspondence to: Shahar Ayal, DE-mail: [email protected] authors contributed equally to

Copyright # 2009 John Wiley &

Ignorance or Integration: The CognitiveProcesses Underlying Choice Behavior

SHAHAR AYAL1�,y and GUY HOCHMAN2y

1Center for Behavioral Economics, Fuqua School of Business, Duke University, USA2MaxWertheimerMinerva Center for Cognitive Studies, Faculty of IndustrialEngineering andManagement,TheTechnion�Israel Institute of Technology, Haifa,Israel

ABSTRACT

The fast-and-frugal heuristic framework assumes noncompensatory tools for humanpreferences (i.e., priority heuristic) and inferences (i.e., take the best heuristic).According to this framework, these heuristics predict choice behavior as well as modelthe cognitive processes underlying such behavior. The current paper presents twostudies that juxtapose predictions derived from these two simple heuristics withalternative predictions derived from compensatory principles. Dependent measuresthat included reaction time, choice pattern, confidence level, and accuracy were betterpredicted by compensatory indices than by noncompensatory indices. These findingssuggest that people do not rely on limited arguments only, but tend to integrate allacquired information into their choice processes. This tendency was replicated evenwhen the experimental task facilitated the use of noncompensatory principles. We arguethat the fast and frugal heuristics can predict the final outcome only under certainconditions, but even in these particular situations they are not applicable to theprocesses underlying choice behavior. An integrative model for choice behavior isproposed that better represents the data. Copyright # 2009 John Wiley & Sons, Ltd.

key words compensatory models; noncompensatory models; take the best heuristic;

priority heuristic; choice behavior

INTRODUCTION

Decision-making models have been traditionally classified into compensatory versus noncompensatory

theoretical frameworks (e.g., Einhorn, 1971; Elrod, Johnson, & White, 2004; Ford, Schmitt, Schechtman,

Hults, & Doherty, 1989; Payne, Bettman, & Johnson, 1993). Generally speaking, compensatory models (e.g.,

Expected Utility Theory, by von Neumann &Morgenstern, 1944; Prospect Theory by Kahneman & Tversky,

1979; Tversky & Kahneman, 1992) assume that choices are governed by tradeoffs between conflicting

uke University, Fuqua School of Business, 1 Towerview Drive, Durham, NC 27708, USA.

the article.

Sons, Ltd.

456 Journal of Behavioral Decision Making

attributes (e.g., reasons, cues). For example, when decision makers have to choose between two stochastic

alternatives, they use weighting and summing processes of some sort, and choose the alternative with the

highest subjective expected utility. Thus, in a compensatory process, high values on some attributes can

compensate for low values on others, and the magnitude of the difference between the expected values of the

given alternatives is a crucial element in the decision process (cf. Busemeyer & Townsend, 1993).

In contrast, noncompensatory models (e.g., Elimination by Aspects, by Tversky, 1972; Lexicographic

Theory by Fishburn, 1974) rely on the assumption that different attributes of the alternatives are used in the

order of their validity or importance. If the most valid attribute (i.e., the ‘‘best’’) differentiates between

alternatives, then the alternative it supports is chosen, and the information search is terminated. Otherwise,

the next best argument is chosen in an argument-wise search pattern. Thus, according to noncompensatory

models, there is no tradeoff between conflicting attribute values. The results of these processes are twofold.

First, decisions can be based on one argument while ignoring all others. Second, these processes circumvent

the need for more complex processes of information integration, such as weighting and summing

(Brandstatter, Gigerenzer, & Hertwig, 2006).

Recently, Gigerenzer and colleagues (e.g., Brandstatter et al., 2006; Gigerenzer, Todd, & the ABC

research group, 1999) put forward the fast-and-frugal heuristics as noncompensatory process models of

human preferences (e.g., the priority heuristic, by Brandstatter et al., 2006) and inferences (e.g., the take the

best heuristic, by Gigerenzer & Goldstein, 1996). As Brandstatter et al. (2006) argue, ‘‘The priority heuristic

is intended to model both choice and process: It not only predicts the outcome but also specifies the order of

priority, a stopping rule, and a decision rule’’ (p. 427).

While many studies confirmed that under some conditions fast-and-frugal heuristics predict choice

behavior (e.g., Broder, 2000, 2003; Broder & Schiffer, 2003a; Rieskamp& Hoffrage, 1999, 2005), there is an

ongoing debate as to whether these heuristics can also serve as process models (e.g., Broder, 2002; Glockner,

2007; Glockner & Betsch, 2008a; Johnson, Schulte-Mecklenbeck, &Willemsen, 2008). In the current paper,

we intend to demonstrate that process models of individuals’ choice behavior cannot be accounted for by

simple noncompensatory principles, but rather have to integrate compensatory and noncompensatory

components. To do so, we derive predictions about indirect measures of process from two prototypical fast-

and-frugal heuristics, and juxtapose them with contradicting predictions from simple compensatory

principles.

The process of the priority heuristicThe priority heuristic (henceforth PH) (Brandstatter et al., 2006) is a simple lexicographic model that

describes human preferences. According to this fast-and-frugal heuristic, the choice between alternatives that

yield different monetary prospects (e.g., the choice between ‘‘Alternative A—gain $3 with a probability of

0.25, otherwise nothing’’ and ‘‘Alternative B—gain $4 with a probability of 0.2, otherwise nothing’’) is

governed by a hierarchy of three rules:

Rule 1: If the difference between the minimum payoffs is larger than 10% of the maximum payoff (i.e., the

aspiration level), choose the gamble with the higher minimum payoff.

Rule 2: Otherwise, if the difference between probabilities of the minimum payoff is larger than 0.1, choose

the gamble with the smaller probability of the minimum payoff.

Rule 3: Otherwise, choose the gamble with the highest maximum payoff.

According to Brandstatter et al. (2006), this fast-and-frugal heuristic can account for a number of

empirical phenomena, such as the Allais paradox (Allais, 1953), the certainty and the possibility effects

(Kahneman & Tversky, 1979), as well as for the majority of choices documented in previous research in the

field (e.g., the data sets from Erev, Roth, Slonim, & Barron, 2002 and from Lopes &Oden, 1999). In addition,

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 457

according to Brandstatter et al. (2006), the PH has the highest predictive value relative to other frugal

heuristics (e.g., the minimax heuristic) or even relative to complex modifications of expected utility (e.g.,

Cumulative Prospect Theory by Tversky & Kahneman, 1992).

In order to support the PH as a process model, Brandstatter et al. (2006) used two indirect measures:

Reaction time (RT) and choice proportion (proportion of choices predicted by the model). Their results were

in line with the PH predictions. First, in an original set of 40 problems, RTwas faster for problems in which

only one PH rule was required to make a choice as compared to problems in which three PH rules were

required. Second, in a reanalysis of two other sets of problems (Kahneman & Tversky, 1979; Lopes & Oden,

1999), Brandstatter et al. found that choice proportion was correlated with the number of PH rules required to

make a choice in each problem. Namely, problems that required examination of only one rule had the highest

choice proportion, two-rule problems had lower proportion, and three-rule problems had the lowest

proportion.

Two recent studies on the choice between two risky alternatives challenged the PH as a process model.

Using a mouse-lab task in addition to RT and choice proportion measures, Johnson et al. (2008) showed that

the search for information does not adhere to the fast-and-frugal principles. Instead, decision makers

exhibited more evidence for integrating data (e.g., probability-payoff transitions within alternative) than for

relying on limited data and ignoring the rest (e.g., comparison of outcomes across the two alternatives).

Glockner and Betsch (2008a) replicated these results and strengthened their validity by using an open mouse-

lab methodology (for a detailed discussion of mouse-lab methods, see Glockner & Betsch, 2007).

In addition, Glockner and Betsch (2008a) show that the RTand choice proportion measures support the PH

predictions only in those cases in which these predictions correspond to the predictions of Cumulative

Prospect Theory (Tversky & Kahneman, 1992). These results suggest that the original support for the PH as a

process model might stem from an overlap between predictions of noncompensatory and compensatory

models. It is thus essential to disentangle this confound by focusing on situations in which the predictions of

these two models are contrasting.

The process of the take the best heuristicThe take the best (henceforth TTB) heuristic (Gigerenzer & Goldstein, 1999; Gigerenzer et al., 1999) is a

simple noncompensatory tool for probabilistic inferences. According to the fast-and-frugal framework, TTB

is implemented when people are required to choose the correct answer for a given question and they are

uncertain about the answer (e.g., which city has more residents? A. Phoenix; B. Dallas). According to

Gigerenzer et al. (1999), in these cases people tend to base their decision on a hierarchy of probabilistic cues.

First, cues will be ordered according to the probability that the cue will identify the correct answer. This

hierarchy determines the lexicographic order in which information (i.e., cues) will be searched. Second, TTB

works by activating three hierarchical rules:

Rule 1: Take the cue with the highest validity, and look up the cue values for the two alternatives.

Rule 2: If the cue is diagnostic (that is, if the cue is positive for one alternative and negative for the other),

stop the search and go to Step 3. Otherwise, go back to Step 1 and search for the next cue.

Rule 3: Choose the alternative with the positive cue value for the decision criterion.

According to Gigerenzer and Goldstein (1996), in some environments this fast-and-frugal heuristic

outperforms more complex compensatory strategies, such as the linear weighted additive (henceforth

WADD) strategy. The WADD strategy computes the sum of all cue values multiplied by the validity of the

cue, and favors the alternative with the largest sum. Indeed, several studies show that when information

acquisition is costly (Broder, 2000), under time pressure (Rieskamp & Hoffrage, 1999, 2005), or when

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

458 Journal of Behavioral Decision Making

information must be retrieved from memory (Broder & Schiffer, 2003a), the TTB model predicts people’s

inferences quite accurately. In addition, it has been shown that when the cost of information acquisition is

high, the pattern of information search fits the TTB predictions (Newell & Shanks, 2003; Newell, Weston, &

Shanks, 2003). However, the WADD model was found to be more suitable than the TTB in predicting

people’s inferences when information acquisition costs are low, when there is no time pressure, and when

information is provided simultaneously (cf. Rieskamp&Otto, 2006). In addition, Glockner (2007) shows that

even though simple noncompensatory strategies such as the TTBmight sometimes lead to accurate decisions,

the majority of people do not tend to adopt them.

Note, however, that focusing solely on predicting the outcomes, the TTB can be considered a submodel of

the WADD (i.e., considering different transformation functions for cue weights, the WADD can always

produce the same choice as TTB), so choice data that support the TTB cannot rule out the possibility that a

WADD strategywas used (Glockner, in press; Lee&Cummins, 2004). Thus, as in the case of the PH, support for

the TTB as a process model might stem from the overlap between predictions of noncompensatory and

compensatory models. Therefore, it is important to focus on situations in which the predictions of these two

models contrast each other rather than to merely rely on their overall predictive power (cf. Broder, 2002).

Indirect measures for processPrevious research shows that people employ both compensatory and noncompensatory strategies, depending

on task demands and environmental factors (Ford et al., 1989; Payne et al., 1993). Since the choice

predictions of different models overlap, it is essential to use multiple means of investigation simultaneously

in the attempt to identify the nature of a specific cognitive process. In the current study, several indirect

measures are offered for the differentiation between the compensatory and the noncompensatory strategies

for choice behavior, including reaction time, choice proportion, and confidence level.

Reaction time (RT)

Fast-and-frugal heuristics assume a sequential limited search of arguments (i.e., information), with an

explicit stopping rule according towhich search is terminated as soon as a discriminative piece of information

is acquired. Therefore, noncompensatory models predict that the fewer arguments (e.g., cues, rules) are

required to make a choice, the faster it will take people to respond. Alternatively, compensatory principles

assume that decision makers integrate relevant arguments, outcomes, and probabilities. Thus, compensatory

models predict that RT will be faster the less integration is required (Payne et al., 1993), or when the

integration of arguments is summed up by larger differences between the expected utilities of the alternatives

(Bergert & Nosofsky, 2007).

Choice proportion

Choice proportion describes the proportion of individuals’ choices that are in line with the prediction of a

specific decision strategy. If the predictions of these two models are sufficiently different, then choice

proportions in each problem can be used to efficiently identify which decision strategy has been adopted

(Broder & Schiffer, 2003b). However, when the model predicts majority of choices, such as in the case of the

PH, choice proportion cannot serve as a means by which to assess the model (Brandstatter et al., 2006). In

these cases a rank order of choice proportion could be used to evaluate models with the assumption that

people make processing errors in each step of their chosen strategy. According to the noncompensatory

principles, this assumption implies that the fewer arguments (e.g., cues, reasons) required to make a choice

the higher the choice proportion will be (for an elaborate explanation, see Brandstatter et al., 2006, p. 428).

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 459

Alternatively, according to compensatory principles, choice proportion depends on the magnitude of the

perceived distance between the expected utilities of the two alternatives. That is, a higher choice proportion

will be obtained when the perceived difference between the two alternatives is larger.

Confidence level

Noncompensatory tools postulate a lexicographic search pattern of information according to a natural

validity hierarchy, so that one-reason (rule) decision-making will be based on the most valid diagnostic

argument. This implies that basing the decision on arguments that have a higher natural priority or importance

will lead to the highest level of confidence in any particular judgment. Take for example a situation in which

the capital city cue (i.e., whether the city is the capital? yes/no) is a better predictor of the city size than the

soccer team cue (i.e., whether the city has a team in the first league? yes/no). In this situation, relying on the

capital city cue is supposed to lead to greater confidence in determining that City A is bigger than City B, as

compared to relying on the soccer team cue. Importantly, this predicted correlation between the priority of

information (e.g., cues, rules) and the level of confidence in a choice has not been explicitly stated for the PH.

However, it seems reasonable to predict such a correlation on the basis of previous assumptions postulated

within the fast-and-frugal heuristics framework. Specifically, the linear relationship between priorities of

arguments (e.g., cues) and level of confidence was one of the building blocks of the probabilistic mental

model and the TTB (Gigerenzer, Hoffrage, & Kleinbolting, 1991; Gigerenzer & Hug, 1992). According to

these noncompensatory models, a confidence rating is determined by the subjective validity of the cue on

which a choice is based (Broder, 2000; Gigerenzer et al., 1991).

Compensatory principles, on the other hand, suggest that an adaptive level of confidence should rely on all

acquired information as well as on its integration. Thus, compensatory models predict that confidence level

will be higher the larger the difference perceived between the alternatives (cf. Glockner & Betsch, 2008b).

Accordingly, participants can base their confidence level either on the magnitude of distance between the

expected utilities of the alternatives, or on the magnitude of difference between the overall weighted sums of

cues supporting each alternative (Budescu, 2006; Erev, Wallsten, & Budescu, 1994; Winkler & Poses, 1993).

STUDY 1: EXAMINING THE PROCESSES UNDERLYING RISKY CHOICES

In this study we examine preferences in two datasets. The first set (called here the Priority Replication set)

replicates the two-outcome choice problems used by Brandstatter et al. (2006) (see Appendix 1). The second

set is based on previous data from the collapse condition of Wakker, Erev, and Weber (1994) (called here the

Wakker et al. set, see Appendix 2). As in the Brandstatter et al. (2006) experiment, the two sets were classified

into a 2 (one rule or three rules required)� 2 (gambles of similar or dissimilar expected value) mixed-

factorial design. The level of similarity in the two datasets was defined according to Brandstatter et al.’s

(2006) classification. That is, any choice problem in which the difference between the expected values (EVs)

of the gambles did not exceed 1% of the maximal EV was classified as similar. All the other choice problems

were classified as dissimilar.

As can be seen in Table 1, this classification allows the formulation of two competing predictions for each

of the aforementioned measures—one that fits the PH and another that follows the compensatory principle of

expected utility maximization.

MethodParticipants

The participants of the priority replication set were 48 undergraduate students from the Faculty of Industrial

Engineering and Management at the Technion—Israel Institute of Technology. The sample included

29 women and 19 men, with a mean age of 23.54 years (SD¼ 1.90).

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

Table 1. Competing predictions of the PH and COMP models for each of the indirect measures (Study 1)

Indirect measure PH prediction COMP prediction

Reaction time (RT) RT will be faster in the one-rule-requiredchoice problems than in the three-rules-required choice problems, independentlyof the level of similarity between theexpected values

RT will be faster in the dissimilar-expected-value choice problems thanin the similar-expected-value choiceproblems, independently of thenumber of rules required

Choice proportion Choice proportion aligned with the PHwill be higher in the one-rule-requiredchoice problems than in the three-rules-required choice problems, independentlyof the level of similarity between theexpected values

Choice proportion aligned with thecompensatory model (EV maximization)will be higher in the dissimilar-expected-value choice problems than in the similar-expected-value choice problems,independently of the number of rulesrequired

�Confidence level Confidence-level will be higher in theone-rule-required choice problems thanin the three-rules-required choice problems,independently of the level of similaritybetween the expected values

Confidence-level will be higher in thedissimilar-expected-value choice problemsthan in the similar-expected-value choiceproblems, independently of the numberof rules required

PH¼ priority heuristic; COMP¼ compensatory model. Choice problems are classified according to (a) the number of PH rules requiredto reach a decision (one vs. three) (b) the level of similarity between the expected values (similar vs. dissimilar).�Note: As already mentioned, it has been postulated that there is a direct relationship between the level of confidence in a choice and thecue validity hierarchy (Broder, 2000; Gigerenzer et al., 1991). According to the fast-and-frugal framework, including the PH, the searchfor information is terminated when a certain aspiration level is met (Brandstatter et al., 2006; Gigerenzer, 1997; Gigerenzer & Goldstein,1996; Hoffrage, 2005). This implies that we can generalize the proposed relationship between confidence and validity of cues to a similarrelationship between confidence and priority of reason.

460 Journal of Behavioral Decision Making

The Wakker et al. (1994) dataset consisted of data from 22 psychology students who had participated in

the study as part of their Introductory Psychology course at the University of North Carolina at Chapel Hill.

Design and procedure

The priority replication set. The experiment was programmed in Microsoft Visual Basic 6 and run on a

desktop PC. Participants were presented with a series of 20 two-alternative forced-choice decisions between

gambles (e.g., choose between Alternative A—gain $2000 with probability 0.6, otherwise $500; and

Alternative B—gain $2000 with probability 0.4, otherwise $1000). All gambles were from the gain domain

(i.e., all outcomes are positive or zero) and they were presented in a random order. The participants were

asked to choose the most attractive alternative as fast as possible. RTwas measured by the software from the

time the choice problem appeared on the screen to the time of decision. At the end of the RT section,

participants were asked to rate how confident they were about each choice they had made, on a confidence

scale ranging from 50 to 100% in steps of 10%.

The Wakker et al. set. Participants’ task was to choose the preferred option in each pair of gambles. For a full

description of the design and procedure, see Wakker et al. (1994). We reanalyzed the original data, which

consisted of RTand choice proportion measures. Our analysis included all the gain gambles that should have

been solved according to 1 or 3 priority rules, thus excluding two of the original choice problems.

Results and discussionAnalysis of RT

The RT results of the priority replication set are shown in the left panel of Figure 1. When the two options had

similar EVs, the mean RT for the one-rule-required choice problems was 12.31 seconds (SD¼ 8.34), and

when the two options had dissimilar EVs the mean RT decreased to 10.03 (SD¼ 5.40). Similarly, for the

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

Figure 1. Mean reaction time (RT) as a function of the similarity between the expected values (EVs) and the number ofreasons examined. Error bars represent standard errors

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 461

three-rules-required choice problems the mean RT for similar EVs was 10.9 (SD¼ 6.18) and the mean for the

dissimilar EV choices was 8.78 (SD¼ 4.34). A 2� 2 repeated measures analysis of variance (ANOVA) was

conducted to test the effects of the number of rules required and the level of similarity between the expected

values on RT as the dependent variable. This analysis revealed a significant main effect both for the level of

similarity between the EVs (F (1, 47)¼ 11.16, p< 0.001) and for the number of rules required (F (1,

47)¼ 4.18, p< 0.05). No significant interaction between the two variables was found.

The RT results of theWakker et al. set are shown in the right-hand panel of Figure 1. When the two options

had similar EVs, the mean RT for the one-rule-required choice problems was 13.12 (SD¼ 7.24) and when the

two options had dissimilar EVs the mean RT decreased to 10.2 (SD¼ 3.89). Similarly, for the three-rules-

required choice problem the mean RT for similar EVs was 14.4 (SD¼ 7.04) and the mean for dissimilar EV

choices was 12.52 (SD¼ 4.72). A 2� 2 ANOVA was conducted to test the effects of the number of rules

required and the level of similarity between the expected values on RT as the dependent variable. This

analysis revealed a significant main effect of the level of similarity between the expected values (F (1,

17)¼ 9.64, p< 0.01). Neither a main effect of the number of rules required (F (1, 17)¼ 1.00, ns) nor an

interaction between the two variables was found.

These results strongly support the RT prediction of the compensatory model. In both datasets RT was

significantly faster for the choice problems with dissimilar EVs than for problems with similar EVs,

regardless of the number of rules required. On the other hand, the RT prediction of the PH was not confirmed,

as the relevant pattern of results was inconsistent across datasets. In the first set, there was a significant effect

for the number of rules but it was in the opposite direction to the one predicted by the PH and in contrast to the

results reported by Brandstatter et al. (2006) with the same set of choice problems. However, in the second

set, the number of rules did not have any effect on RT. One possible explanation for the inconsistency between

the current results and the results of Brandstatter et al. (2006) could be based on a sampling bias. Namely, in

the current study, participants were engineering students while in the Brandstatter et al. study participants

were psychology students. Since engineering students are more mathematically oriented, it is possible that

they are more familiar with the concept of EV and can thus utilize it more easily, whereas the effect of the

number of rules became marginal and reflected random errors.

Taken together, the RT results suggest that participants were much more sensitive to the magnitude of

difference between the EVs of the alternatives than to the numbers of rules required according to the PH

(which yielded inconsistent results). Thus, under the current settings, the RT data indicated that participants

made tradeoffs between probabilities and outcomes, and employed complex compensatory processes for

calculating and comparing EVs.

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

Figure 2. Left-hand panel. Mean choice proportion (CP) as a function of the similarity between the expected values(EVs) and the number of reasons examined. Right-hand panel. Choice proportion (CP) predicted by the PH and the

COMP as a function the equality between the predictions of the models. Error bars represent standard errors

462 Journal of Behavioral Decision Making

Analysis of choice proportion

The choice proportion results of the Priority Replication set are shown in the left panel of Figure 2. Across

participants, the mean proportion of choices that were in line with PH for the one-rule-required choice

problems was 0.55 (SD¼ 0.22) when the two options had similar EVs and this mean increased to 0.85

(SD¼ 0.19) for dissimilar EV problems. For the three-rules-required choice problem the mean proportion of

choices was 0.6 (SD¼ 0.23) for similar EVs and 0.89 (SD¼ 0.18) for dissimilar EV problems. A 2� 2

within-participants repeated measures ANOVA was conducted to test the effects of the number of rules

required and the level of similarity between the expected values on choice proportion as the dependent

variable. This analysis revealed a significant main effect of the level of similarity between the expected values

(F (1, 47)¼ 119.91, p< 0.001). Neither a main effect of the number of rules required (F (1, 47)¼ 2.71, ns),

nor an interaction between the two variables was found.

A closer examination of the priority replication set revealed that in all the dissimilar EV choice problems the

PH predictions coincided with EV maximization. In contrast, in the Wakker et al. set, there were not only

14 dissimilar problems in which this was the case, but also 6 choice problems in which the PH predictions

contradicted the compensatory predictions (see Appendix 2). An analysis of the dissimilar problems only and a

comparison of choice proportion under the equal and the non-equal prediction choice problems allowed for a

more thorough examination of the two competing models. The results of this analysis (presented in the right-

hand panel of Figure 2) show that when the two models share the same predictions, the mean proportion of

choices that aligns with this prediction is 0.697 (SD¼ 0.16). However, when the two models had contradicting

predictions, choice proportion according to the PH prediction decreased to 0.52 (SD¼ 0.18) and was not

significantly above chance level (t (21)¼ 0.51, ns), whereas choice proportion according to the compensatory

prediction remained robust at 0.61 (0.19), and above chance level (t (21)¼ 2.7, p< 0.03). A paired-samples t-test

analysis revealed a significant decrease in choice proportion according to the PH prediction (t (21)¼ 2.93,

p< 0.005), and no decrease in choice proportion according to the compensatory predictions (t (21)¼ 1.35, ns). In

addition, when the two models had contradicting predictions, choice proportion according to the compensatory

prediction was significantly higher than choice proportion according to the PH (t (21)¼�1.91, p< 0.05).

The choice proportion results confirm the compensatory prediction for choice proportion, and not the PH

prediction. The pattern of results in the first set converges with the pattern of results for the RT measure.

Choice proportion depended on the level of similarity between EVs, with higher proportions obtained for

choice problems with dissimilar EVs as compared to choice problems with similar EVs. In addition, the

choice proportion was not sensitive to the number of rules required. Moreover, the choice proportion results

of the second set suggest that when the two models share the same choice predictions, both PH and

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 463

compensatory models have high predictive power. These cases cannot discriminate between the two models

and thus do not contribute to our understanding of the underlying process (cf. Glockner and Betsch, 2008a).

When focusing on diagnostic cases in which the two models have contradicting predictions, it turns out that

only the compensatory model retained its predictive power and could correctly predict choice behavior in

those situations, while the PH predictive power was at chance level.

Finally, it is important to note that the choice proportion was not at chance level in the similar EV choice

problems in which the compensatory model predicted indifference (for example, in the priority replication set

the aggregated mean of choice proportion across the two conditions with similar EVs was 0.57, t (47)¼ 3.48,

p< 0.001). It could thus be argued that when integration does not prevail, participants employ additional

strategies, possibly noncompensatory in nature, in order to reach to a decision.

Analysis of confidence level

This variable was measured only in the Priority Replication set, since it was not a part of the Wakker et al.

design. The confidence level results are shown in Figure 3. For the one-rule-required choice problems the

mean confidence level was 85.73% (SD¼ 11.29) when the two options had similar EVs and this mean

increased to 91.82% (SD¼ 8.07) for the dissimilar EV choice problems. For the three-rules-required choice

problems the mean confidence level was 85.02% (SD¼ 10.42) for similar EVs and 90.3% (SD¼ 8.75) for

the dissimilar EV choices. A 2� 2 within participants repeated measures ANOVAwas conducted to test the

effects of the number of rules required and the level of similarity between the expected values on confidence

level as the dependent variable. This analysis revealed a significant main effect for the level of similarity

between the expected values (F (1, 47)¼ 27.84, p< 0.001). Neither a main effect for the number of rules

required (F (1, 47)¼ 2.57, ns), nor an interaction between the two variables was found.

These results support the compensatory prediction for confidence-level, not the PH prediction. Participants

tended to exhibit higher levels of confidence in their preferences for choice problemswith dissimilar EVs relative

to choice problems with similar EVs, regardless of the number of rules required. The confidence for one-rule-

required problems was slightly higher than the confidence for the three-rules-required problems, but this

difference was not significant. Similar to the RT and the choice proportion results, the confidence level results

suggest that participants tend to integrate outcome and probability information in a compensatory manner.

Figure 3. Mean level of confidence (LOC) as a function of the similarity between the expected values (EVs) and thenumber of reasons examined. Error bars represent standard errors

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

464 Journal of Behavioral Decision Making

In summary, the findings of the three different measures in Experiment 1 support the idea that people tend

to integrate information regarding both outcome and probabilities when facing preference decision tasks,

rather than rely on limited information in a noncompensatory manner. Furthermore, these results demonstrate

that the predictive power of noncompensatory models cannot be generalized to diagnostic cases, when the

two models have contradicting predictions.

Note, however, that these results provide evidence for the cognitive processes underlying choice behavior

when all reasons are displayed simultaneously, and all the required information is available and fully

displayed at no cost (i.e., description based decisions; Hertwig, Barron, Weber, & Erev, 2004). In contrast, in

situations in which information needs to be searched (either in memory or in information stores), and is

costly, it is more likely that the cognitive processes will be more noncompensatory and lexicographic in

nature (e.g., Gigerenzer et al., 1999). Study 2 was designed to test this possibility.

STUDY 2: EXAMINING THE PROCESSES UNDERLYING PROBABILISTIC INFERENCES

In this study we explore the cognitive processes underlying the inferences that people makewhen trying to select

the correct answer in two-alternative forced-choice questions (e.g., Which team won the 2007 European-Cup

final? A. Milan; B. Liverpool). In order to examine the processes underlying these situations, we constructed a

binary choice task in which information was provided in a cue-wise manner, and cue acquisition was costly. This

paradigmwas adapted from previous studies (e.g., Broder, 2000; Newell & Shanks, 2003; Rieskamp&Hoffrage,

1999), in an attempt to create an optimal environment for using the TTB heuristic (cf. Rieskamp & Otto, 2006).

Individuals’ choice strategies were analyzed using a Maximum Likelihood strategy classification method

(also called Bayesian strategy classification, Broder & Schiffer, 2003b; Broder, in press). Participants are

thereby classified according to their choices as TTB orWADD users, using maximum likelihood estimation.1

In addition, participants’ level of confidence in their choice was documented. These classifications made it

possible to formulate two competing predictions for each of the aforementioned measures, one for the TTB

and one for the WADD strategy (Table 2).

MethodParticipants

Sixty undergraduate students from the faculty of Industrial Engineering and Management at the Technion—

Israel Institute of Technology participated in the experiment. The sample included 28 women and 32 men,

with a mean age of 23.9 years (SD¼ 2.10).

1Maximum likelihood estimation was calculated with the formula developed by Broder and Schiffer (2003b, p. 201). The conditionallikelihood Lk for a data set (individual set of choices) under the condition that a certain decision strategy k was applied with a constanterror rate ek is given by

Lk ¼ pðnjk

��k; "kÞ ¼YJ

j¼1

nj

njk

� �ð1� "kÞnjk"k

ðnj�njkÞ;

with J being the total number of categories of decision tasks, nj being the number of choices in category j, and njk being the number ofchoices in line with the prediction of strategy k in category j. The error rate ek for each decision strategy is estimated by

"k ¼XJ

j¼1

ðnj � njkÞ" #

�XJ

j¼1

nj

" #:

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

Table 2. Competing predictions of the TTB and the COMP models for each of the indirect measures (Study 2)

Indirect measures TTB prediction COMP prediction

Choice proportion The majority of participants will usethe TTB strategy

The majority of participants will usethe WADD strategy

Confidence level Confidence level will depend on thecue precedence, independently of thedifference in weighted cue values

Confidence level will depend on thedifference in weighted cue values forthe options, independently of cueprecedence

TTB¼Take The Best model;WADD¼Weighted Additive model; COMP¼Compensatory model. Cue precedence is determined by thelocation of the cue within the validity hierarchy. Weighted cue values are the sum of gathered cues multiplied by their validity (i.e.,weight).

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 465

Design and stimulus

The experiment was programmed in Visual Basic 6 and was run on a desktop PC. Participants were presented

with a series of 40 two-alternative forced-choice questions regarding two soccer teams playing a match.

Matches were sampled from the inventory of actual games played in the Italian soccer league in 1998–1999

that did not end in a tie. However, to ensure that answers are based solely on gathered information, a randomly

generated fictitious namewas assigned to each team to prevent familiarity and identification. In addition, each

team appeared in the sample no more than four times on average, and the questions were presented in a

random order between participants. An example of a question could be: ‘‘Which of the two teams won the

game played on February 14, 1999? (A) Fedele; (B) Cocci.’’ Three binary cues provided information in the

form of a yes/no question concerning the teams in question. The three cues were:

(1) T

Copy

he previous game cue—Did the team win the previous match between the two teams in question?

(2) T

he top team cue—Was the team considered a top team at the time (i.e., one of the six leading teams)?

(3) T

he big city cue—Does the team come from a big city (i.e., a city with more than 250 000 residents)?

The three cues differed in their objective validity, with the previous-game cue being the most valid

(p¼ 0.78), the top-team cue the second most valid (p¼ 0.72), and the big-city cue the least valid (p¼ 0.65).

Validity was defined as the conditional probability that the cue would correctly identify the winning team for

all possible comparisons of two teams that differed on this cue. Following each forced choice question,

participants were asked to state how confident they were about their choice, on a confidence level scale

(ranging from 50 to 100% in steps of 10%) presented separately for each question.

Procedure

Participants were told that on each trial they would be asked to determine which of two teams that played in

the Italian soccer league in the years 1998–1999 had won the match. To help them make these choices, they

were encouraged to buy up to three cues for each specific game. To create optimal conditions for TTB

adoption, each cue was priced 1/2 NIS (approximately 0.15 USD). Cues could be bought online and

participants were given the value (YES or NO) of each cue they bought. In order to ensure interest and

credibility, an additional monetary incentive of 50 NIS (approximately 15 USD) was offered for at least

19 correct answers, and 30 NIS (approximately 9 USD) for 16–18 correct answers. The cost of the cues was

subtracted from the potential reward, to ensure that the cost would not exceed the gain. Otherwise, 15 NIS

(approximately 4.5 USD) were paid as a show-up fee.

Participants were randomly assigned to one of two conditions. Under the Subjective condition,

participants were first instructed to determine the cue validity hierarchy by rating the three cues according to

perceived validity. Then, cues were provided according to this order and each participant had the opportunity

right # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

466 Journal of Behavioral Decision Making

to gather information in a lexicographical order according to his/her subjective hierarchy. Under the

Objective condition, participants were directly informed that information would be provided according to the

accuracy probability of the cues. Then, participants were given the opportunity to gather information through

these three cues according to this objective order (i.e., first the previous-game cue, then the top-team cue, and

finally the big-city cue).

Participants were instructed to provide their answers as quickly and as accurately as possible. After

providing each answer, participants were asked to rate how confident they were about their choice.

Strategy classification and model prediction calculations

For each decision, questions were first classified into three categories: (a) non-differentiating cases—in

which both strategies make the same predictions; (b) differentiating cases—in which different predictions

are made; and (c) random choice cases—in which at least one strategy does not lead to a decision and

predicts random choices. Error probabilities in applying the respective strategy (i.e., answering against the

model’s predictions) were determined for each individual across all decisions (e¼wrong predictions divided

by number of decisions with prediction). The conditional likelihood of each person’s choice pattern was

estimated given the application of a certain strategy with a constant error rate.2 Participants were classified as

users of the strategy for which the highest likelihood was found. A Bayes-ratio was calculated as a measure of

the classification certainty by dividing the likelihood for the application of the identified strategy by the

likelihood of the other strategy (Broder, in press).

The predictions for the WADD strategy were determined by comparing WADD scores for both options

(weighted sum of cue validity factor wcue and cue value; negative cue values were coded as �1, positive as

þ1). Cue validity factors were derived by correcting for the fact that cues with a 0.50 validity are

uninformative (i.e., wcue¼ pcue—0.50) (cf. Glockner & Betsch, 2008b). In addition, for each problem, we

determined the prediction of the TTB and the number of cues that were necessary to apply a TTB strategy.

Results and discussionAs in Study 1, the focus here was on situations that could distinguish between the noncompensatory and the

compensatory predictions. To provide a fair evaluation, only trials in which information was acquired were

considered. Thus, 17% of the trials (24% of the trials in the objective condition and 11% of the trials in the

subjective condition) were excluded from the analyses because they were based on pure guessing without

acquiring any piece of information.

Manipulation and paradigm check

To check whether the manipulation yielded different information search patterns, we examined the chosen

hierarchies of the subjective condition. As seen in Table 3, 12 of the 30 participants (40%) intuitively adopted

the objective validity hierarchy (i.e., previous game-top team-big city), whereas 18 participants (60%)

adopted a different hierarchy. Of these 18, 15 participants (50%) adopted a subjective hierarchy in which the top-

team cuewas considered themost valid cue and the big city cuewas considered the least valid cue (i.e., top team-

previous game-big city). The remaining 3 participants (10%) adopted different cue validity hierarchies.

2The equal weighted model (EQW) (Dawes, 1979) was not tested separately, since in the majority of cases it includes similarcompensatory processes and thus makes the same predictions as the WADD. It should be noted, however, that the WADD classificationsmight therefore partially contain some EQW adherents.

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

Table 3. Proportion of correct responses (hit rate) as a function of cue hierarchy (Study 2)

Experimental condition Cue hierarchy N Hit rate

Objective PG—TT—BC 30 0.725 (0.097)Subjective PG—TT—BC 12 0.720 (0.121)

TT—PG—BC 15 0.710 (0.077)BC—PG—TT 2 0.718 (0.056)PG—BC—TT 1 0.692

N¼ number of participants employing the specific cue hierarchy (order of search pattern); PG¼ previous game; TT¼ top team; BC¼ bigcity. The right column contains proportions of hit rate across trials. Standard deviations are in brackets.

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 467

To provide additional evidence for the information search pattern and determine whether our experimental

paradigm replicated previous findings in the field, we evaluated the number of TTB violations, both in the

subjective and in the objective condition. Calculation of TTB violations relied on the parameters used by

Newell, Weston, and Shanks (2003). Our results appear to correspond to previous empirical examinations of

the TTB (e.g., Broder, 2000; Newell & Shanks, 2003; Newell et al., 2003), and support what was termed by

Broder (2000) the weak hypothesis, according to which only some but not all participants acquired

information and decided according to the prediction of the TTB strategy (for full information about

individual TTB violations in each of the experimental condition, see http://tx.technion.ac.il/�hochmang/

Online.htm).

Analysis of choice strategy classification

The analysis of individual choice patterns revealed a clear dominance of the WADD strategy. Namely, the

maximum likelihood estimations showed that 72% of the participants were classified as users of the WADD

strategy, whereas only 18% used the TTB. Choice strategy could not be determined for the remaining 10% of

the sample. In addition, we analyzed Bayes-ratios to determine the quality of the strategy classification. The

median of the Bayes ratio was 11305.84 for the WADD classifications and 8963.58 for the TTB

classification.3 Thus, in both cases the classification seems to be very robust.

These results confirm the compensatory prediction for choice proportion, not the TTB prediction.

Examining choice patterns that differentiated between compensatory and noncompensatory predictions,4 we

found that the majority of participants based their decisions on integration of information rather than on

limited information.

Analysis of confidence level

To examine the relative importance of the predictors of confidence level, we applied a dominance analysis

(Azen & Budescu, 2003; Budescu, 1993). Dominance analysis is a general method for comparing the relative

importance of predictors in a multiple regression, by comparing the R2 values from all possible subset

models. An evaluation of the relative contribution of each predictor to the R2 can determine the contribution

of each predictor in explaining the dependent variable (e.g., confidence level in our case).

The results of this analysis suggest that the WADD index completely dominated the precedence of cue, as

the additional contribution of the WADD index was greater than the contribution of the cue precedence

3Bayes-ratios reflect how likely it is that participants whowere classified as users of a certain choice strategy in fact used this strategy andnot the other one. Thus, these results suggest, for example, that it is 11306 times more likely that a participant who was classified as aWADD user was in fact a WADD user and not a TTB user (Broder, in press).4The average number of cases in which the predictions of the choice pattern differed between the two models across participants was22.05 (SD¼ 12.27) out of 40. This average is dependent upon actual behavior of participants (e.g., number of cues acquired).

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

Table 4. Dominance analysis of TTB and WADD as predictors of confidence-level (Study 2)

Subset model (X) r2X�LOC

Additional contribution

TTB WADD

All sampleTTB 0.03 0.13WADD 0.16 0.03

TTB adherentsTTB 0.038 0.036WADD 0.074 0.006

The column labeled r2X�LOC represents the variance in confidence-level explained by the model appearing in the corresponding row.Columns labeled TTB andWADD contain the additional contributions to the explained variance gained by adding the column variable tothe row model. Blank cells indicate that data are not applicable.

468 Journal of Behavioral Decision Making

variable (see the upper panel in Table 4). A regression analysis in which theWADD index and the precedence

of cue predicted confidence level together was significant (R2¼ 0.16, F (1, 405)¼ 76.46, p< 0.0001).

However, in this regression theWADD index was significant (b¼ 18.13, t (405)¼ 8.74, p< 0.0001), whereas

the precedence of cue was not significant (t (611)¼�1.53, p¼ 0.13). These results suggest that the

confidence level could be better explained by compensatory principles than by noncompensatory principles.

Note that it could be argued that the confidence level results are a byproduct of the low percentage of TTB

users in the current dataset. To examine this possibility, we conducted the same analysis only for those cases

in which the choice pattern was predicted by the TTB (i.e., the search was stopped for the first discriminating

cue and decision was based on this cue alone). The lower panel of Table 4 shows that this analysis yielded a

similar pattern of results. Again, the WADD index was found to dominate completely the precedence of cue.

A regression analysis revealed a significant overall prediction (R2¼ 0.071, F (1261)¼ 20.02, p< 0.0001).

Yet, while the WADD index was significant (b¼ 14.86, t (261)¼ 4.48, p< 0.0001), the precedence of cue

was not significant (t (261)¼�1.64, p¼ 0.10). These results suggest that even when the choice pattern is

captured by a noncompensatory strategy, there is evidence for underlying compensatory processes.

One-reason decision making or integration

Lastly, although the results of both the strategy classification and the confidence level analysis clearly

demonstrate that people tend to gather more information than predicted by noncompensatory models, one can

argue that individuals’ final decision is based on one reason, rather than on the integration of all the available

information. To examine this possibility, we analyzed the proportions of correct responses (hit rate).

Traditionally, the hit rate is used to assess the objective accuracy of the model. However, due to our cue

validity hierarchy manipulation, along with the environment that facilitated TTB use, the hit rate could

further indicate which process was used. More specifically, if individuals were basing their decision on the

most valid cue, while ignoring all other information (as implied by the TTB), the hit rate should be higher

when only few cues are acquired. Of course, this trend is expected only under the objective cue hierarchy but

not under the subjective cue hierarchy. On the other hand, if people integrate all the information they gather in

a compensatory manner, the hit rate should increase rather than decrease the more cues are acquired.

The results of the hit rate analysis showed that across choice problems the hit rate increased as a function

of the number of cues, regardless of the cue validity hierarchy. A 3� 2 ANOVA was conducted to test the

effect of the number of cues (one, two, or three) and the cue validity hierarchy (objective, subjective) on hit

rate as the dependent variable. This analysis revealed a significant main effect of the number of cues (F (2,

74)¼ 6.213, p< 0.005). Neither a main effect of the type of cue validity hierarchy was found, nor an

interaction between the two variables.

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 469

This pattern of results converges with the strategy classification and the confidence level results. The fact

that under both conditions more cues increased the hit rate rather than decreased, it suggests that people tend

to integrate acquired information in a compensatory manner.

In summary, since not all available information was acquired (in several cases no information was

acquired at all), the results of Study 2 imply that decisionmakers tend to save resources (e.g., processing time,

costs) by neglecting some of the available information, as suggested by the fast-and-frugal framework (e.g.,

Gigerenzer et al., 1999). At the same time, these results indicate that people do not neglect information that

has already been gathered, as seen by the fact that such information affects their choice patterns, their

confidence judgments, and their hit rates.

GENERAL DISCUSSION

The theoretical appeal of noncompensatory models lies in their parsimony and in their potentially impressive

predictive power under certain situations. Traditionally, these models were designed to predict choice

behavior (i.e., paramorphic models), and they reflected, to some extent, a tradeoff between simplicity and

descriptive accuracy. For example, in an early noncompensatory model proposed by Tversky (i.e.,

Elimination by Aspects, 1972) the author stated that ‘‘. . .there may be many contexts in which it (the model)

provides a good approximation to much more complicated compensatory models and could thus serve as a

useful simplification procedure. . .’’ (p. 298). However, the fast-and-frugal framework undermines this

traditional convention, suggesting that a new characterization of the mechanism underlying choice behavior

is in order, and that this new framework should be strictly noncompensatory in nature, and free of the classic

compensatory principles of integration.

The results of the current study indicate that the fast-and-frugal framework might be too simplistic, as it

cannot capture the full essence and complexity of the processes underlying choice behavior. Five hypotheses

concerning indirect measures of processes were derived from the PH and the TTB models and were

juxtaposed with contradicting compensatory hypotheses to conduct an empirical examination of the

mechanisms underlying choice behavior. The pattern of results suggests that the compensatory model is

better at predicting decision behavior than is the noncompensatory model.

Specifically, when information was provided simultaneously, and at no cost (i.e., in the preference task in

Study 1), RT, choice proportion, and confidence level in one’s judgment were all better explained by

compensatory principles, such as summing and weighting. These results converge with recent findings (e.g.,

Glockner & Betsch, 2008a; Johnson et al., 2008) according to which people exhibit much more evidence for

integrating data in risky choice than for relying on limited data and ignoring the rest. Moreover, the choice

pattern analysis suggests not only that noncompensatory models cannot capture the underlying processes of

human preferences, but also that they are poorly qualified as as-if models (cf. Birnbaum, 2008; Glockner &

Betsch, 2008a; Rieger &Wang, 2008). The same pattern of results was replicated in an environment in which

information was provided sequentially and at high costs (i.e., the inference task in Study 2), an environment

traditionally considered to encourage the employment of noncompensatory choice strategies. As in Study 1,

the process measures of choice proportion and level of confidence, as well as a measure of accuracy, were

better explained by compensatory principles. It could be argued, however, that the manipulation in Study 2

provided no opportunity to learn the structure of the environment. Thus, participants failed to learn that the

TTB is an appropriate strategy (cf. Rieskamp & Otto, 2006). This account can explain why the TTB choice

proportions were low. In addition, instead of arguing that compensatory processes govern these tasks, it could

simply be argued that in unfamiliar environments people might tend to employ compensatory strategies (e.g.,

WADD) (see for example Broder, 2003). It is important to point out, however, that the level of confidencewas

predicted by the compensatory index alone even in those cases in which participants were classified as TTB

adherents exclusively (i.e., when the TTB accurately predicted choice pattern). Namely, even when the TTB

predicted choice behavior, it could not account for the underlying processes.

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

470 Journal of Behavioral Decision Making

In summary, the results from different indirect measures (i.e., RT, choice proportion, confidence level, and

hit rate) that were tested in diverse environments converge and point to the conclusion that compensatory

processes of integration dominate choice behavior. Although the use of noncompensatory tools in certain

situations could not be ruled out, it seems that people do not forsake more complex compensatory processes

of integration, and prefer to attend to previously gathered information.

Note that Brandstatter et al. (2006) were familiar with the conclusion that people might employ

compensatory tools, and suggested that the PH model might have been stronger had it assumed that people

computed EVs, took their ratio into account, and chose the gamble with the highest EV when the ratio

exceeded 2. The results of the current study support this conclusion, and posit that the coexistence of

compensatory and noncompensatory principles may lead to the formation of better decision-making models

(see also Payne et al., 1992). Importantly, Brandstatter et al. (2006) suggested that measures of reaction time

and choice proportions could examine the process underlying choice. However, this line of research has led to

inconsistent conclusions in the past (i.e., support of noncompensatory principles in Brandstatter et al. and of

compensatory principles in Glockner & Betsch, 2008a). The current study sheds light on this puzzle and

suggests that the predictive accuracy of the noncompensatory models primarily depends on their agreement

with compensatory predictions, and thus it is reduced dramatically in diagnostic situations (cf. Birnbaum,

2008; Glockner & Betsch, 2008a).

The use of confidence level in the comparison of compensatory and noncompensatory hypotheses is quite

novel (but see also Glockner, 2006; Glockner & Betsch, 2008b). The importance of measuring subjective

levels of confidence in a choice stems not only from the direct connection between this confidence level and

basic assumptions of noncompensatory models, but also from the fact that confidence level may reflect a

subjective evaluation of the processes that led to a certain decision. Indeed, the fact that evidence for

compensatory processes was obtained both through a meta-cognitive measure (confidence level) and through

different controlled and uncontrolled measures (i.e., RT, choice proportion, and hit rate) strengthens the

validity of the present results. In addition, these results highlight the importance of usingmultiple measures to

examine and understand the processes underlying choice behavior, and to distinguish between competing

theories.

The current results lead us to suggest a model according towhich choice behavior integrates compensatory

and noncompensatory principles. Under such a model, decision makers start a choice task with a preliminary

intuitive or automated compensatory process of integrating all (or most of) the available information. For

example, this integration is used to evaluate the expected value of the alternatives, or to evaluate the validity

of the available cues. If the results of this preliminary process are satisfactory (e.g., the difference between the

EVs is large enough) (cf. Decision Field Theory of Busemeyer & Townsend, 1993), the decision process is

terminated and a decision is made. In all other cases, decision makers use a rational compensatory selection

among the many tools available at their disposal. Presumably, most of these tools are noncompensatory in

nature.

Our model explains why previous research reported high predictive power for noncompensatory models

(e.g., Brandstatter et al., 2006; Gigerenzer & Goldstein, 1996; Gigerenzer et al., 1999) but failed to support

process predictions (e.g., Glockner & Betsch, 2008a; Johnson et al., 2008). The noncompensatory models’

predictive power is evident in two specific situations. In the first situation, the preliminary compensatory

model governs choice but the final outcome also fits with noncompensatory predictions (i.e., non diagnostic

situations). In the second situation, the preliminary compensatory screening does not provide a clear answer

(e.g., choice between two alternatives with similar expected utility), and thus the final decision is based on a

noncompensatory heuristic. In any case, the model assumes that the default initial process underlying choice

behavior is compensatory.

Thus, contrary to the conclusions made by Gigerenzer and colleagues (e.g., Brandstatter et al., 2006;

Gigerenzer et al., 1999), the claim that decision makers have noncompensatory heuristics in their toolbox

highlights the importance of the compensatory processes, rather than makes them dispensable.

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

APPENDIX 1: THE CHOICE PROBLEMS OF THE PRIORITY REPLICATION SET (STUDY 1)

One reason examined Three reason examined

EV similar EV dissimilar EV similar EV dissimilar

(2000,0.60; 500, 0.40) (3000,0.60; 1500, 0.40) (2000,0.10; 500, 0.90) (5000,0.10; 500, 0.90)(2000, 0.40; 1000, 0.60) (2000, 0.40; 1000, 0.60) (2500, 0.05; 550, 0.95) (2500, 0.05; 550, 0.95)

(5000, 0.20; 2000, 0.80) (6000, 0.20; 3000, 0.80) (4000, 0.25; 3000, 0.75) (7000, 0.25; 3000, 0.75)(4000, 0.50; 1200, 0.50) (4000, 0.50; 1200, 0.50) (5000, 0.20; 2800, 0.80) (5000, 0.20; 2800, 0.80)

(4000, 0.20; 2000, 0.80) (5000, 0.20; 3000, 0.80) (6000, 0.30; 2500, 0.70) (9000, 0.30; 2500, 0.70)(3000, 0.70; 1000, 0.30) (3000, 0.70; 1000, 0.30) (8200, 0.25; 2000, 0.75) (8200, 0.25; 2000, 0.75)

(900, 0.40; 500, 0.60) (1900, 0.40; 1500, 0.60) (3000, 0.40; 2000, 0.60) (6000, 0.40; 2000, 0.60)(2500, 0.20; 200, 0.80) (2500, 0.20; 200, 0.80) (3600, 0.35; 1750, 0.65) (3600, 0.35; 1750, 0.65)

(1000, 0.50; 0, 0.50) (2000, 0.50; 1000, 0.50) (2500, 0.33; 0, 0.67) (5500, 0.33; 0, 0.67)(500, 1.00) (500, 1.00) (2400, 0.34; 0, 0.66) (2400, 0.34; 0, 0.66)

All the choice problems were taken from the problems used in Brandstatter et al. (2006) original priority heuristic article.The choice problems are classified here according to (a) the number of priority heuristic rules that are required to be examined (one vs.three), and (b) the level of similarity between the excepted values (similar vs. dissimilar).

APPENDIX 2: THE CHOICE PROBLEMS OF THE WAKKER ET AL. SET (STUDY 1)

One reason examined Three reason examined

EV similar EV dissimilar EV similar EV dissimilar

(4.5, 0.25; 9, 0.20; 6.5, 0.55) (3, 0.40; 2, 0.50; 0, 0.10) (4.5, 0.25; 9, 0.20; .5, 0.55) (0.5, 0.40; 7.5, 0.20; 1.5, 0.40)(6, 0.25; 7, 0.20; 6.5, 0.55) (2, 1.00) (6, 0.25; 7, 0.20; 0.5, 0.55) (0.5, 0.40; 6, 0.20; 2.5, 0.40)

(4.5, .25; 9, 0.20; 9.5, 0.55) (3, 0.40; 0, 0.60) (4.5, 0.25; 9, 0.20; 3.5, 0.55) (3, 0.40; 7.5, 0.20; 1.5, 0.40)(6, .25; 7, 0.20; 9.5, 0.55) (2.5, 0.50; 1.5, 0.50) (6, 0.25; 7, 0.20; 3.5, 0.55) (3, 0.40; 6, 0.20; 2.5, 0.40)

(3.5, 0.10; 6, 0.70; 12.5, 0.20) (6, 0.15; 0.5, 0.65; 3, 0.20) (5.5, 0.40; 7.5, 0.20; 1.5, 0.40)(5.5, 0.10; 6, 0.70; 10.5, 0.20) (5.5, 0.15; 0.5, 0.65; 3.5, 0.20) (5.5, 0.40; 6, 0.20; 2.5, 0.40)

(3.5, 0.10; 9.5, 0.70; 12.5, 0.20) (6, 0.15; 2.5, 0.65; 3, 0.20) (8, 0.40; 7.5, 0.20; 1.5, 0.40)(5.5, 0.10; 9.5, 0.70; 10.5, 0.20) (5.5, 0.15; 2.5, 0.65; 3.5, 0.20) (8, 0.40; 6, 0.20; 2.5, 0.40)

(3.5, 0.10; 13, 0.70; 12.5, 0.20) (6, 0.15; 4.5, 0.65; 3, 0.20) (3.5, 0.10; 2.5, 0.70; 12.5, 0.20)(5.5, 0.10; 13, 0.70; 10.5, 0.20) (5.5, 0.15; 4.5, 0.65; 3.5, 0.20) (5.5, 0.10; 2.5, 0.70; 10.5, 0.20)

(3, 0.40; 0, 0.10; 2, 0.50) (6, 0.15; 6.5, 0.65; 3, 0.20) (3, 0.40; 0, 0.60)(2, 1.00) (5.5, 0.15; 6.5, 0.65; 3.5, 0.20) (2, 0.50; 0, 0.50)

(3, 0.40; 0, 0.10; 4, 0.50) (2, 0.20; 0, 0.80) (2, 0.60; 5, 0.40)(2, 0.50; 4, 0.50) (1.5, 0.25; 0, 0.75) (2, 0.50; 4, 0.50)

(3, 0.40; 0, 0.10; 6, 0.50)(2, 0.50; 6, 0.50)

(4, 0.50; 2, 0.10; 5, 0.40)(4, 1.00)

(6, 0.50; 2, 0.10; 5, 0.40)(6, 0.50; 4, 0.50)

(8, 0.50; 2, 0.10; 5, 0.40)(8, 0.50; 4, 0.50)

(4, 0.10; 0, 0.80; 1, 0.10)(1, 0.20; 2, 0.80)

(2, 0.80; 0, 0.20)(1.5, 1.00)

All these choice problems were presented in the original article ofWakker et al. (1994). The choice problems are classified here accordingto (a) the number of priority heuristic rules that are required to be examined (one vs. three), and (b) the level of similarity between theexcepted values (similar vs. dissimilar).

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 471

472 Journal of Behavioral Decision Making

REFERENCES

Allais, M. (1953/1979). La psychologie de l’home rationnel devant le risque: Critique des postulats et axiomes de l’ecoleAmericaine. JT Econometrica, 21, 503–546. Translated and reprinted in Allais & Hagen, 1979 (in French).

Azen, R., & Budescu, D. V. (2003). The dominance analysis approach for comparing predictors in multiple regression.Psychological Methods, 8, 129–148.

Bergert, F. B., & Nosofsky, R. M. (2007). A response-time approach to comparing generalized rational and take-the-bestmodels of decision making. Journal of Experimental Psychology: Learning, Memory, and Cognition, 33, 107–129.

Birnbaum, M. H. (2008). Evaluation of the priority heuristic as a descriptive model of risky decision making: Commenton Brandstatter, Gigerenzer, and Hertwig (2006). Psychological Review, 115, 253–262.

Brandstatter, E., Gigerenzer, G., & Hertwig, R. (2006). The priority heuristic: Making choices without trade-offs.Psychological Review, 113, 409–432.

Broder, A. (2000). Assessing the empirical validity of the ‘‘take-the-best’’ heuristic as a model of human probabilisticinference. Journal of Experimental Psychology: Learning, Memory, and Cognition, 26, 1332–1346.

Broder, A. (2002). Take the best, Dawes’ rule, and compensatory decision strategies: A regression-based classificationmethod. Quality and Quantity, 36, 219–238.

Broder, A. (2003). Decision making with the ‘‘adaptive toolbox’’: Influence of environmental structure, intelligence,and working memory load. Journal of Experimental Psychology: Learning, Memory, and Cognition, 29, 611–625.

Broder, A., & Schiffer, S. (2003a). ‘‘Take the best’’ versus simultaneous feature matching: Probabilistic inferences frommemory and effects of representation format. Journal of Experimental Psychology: General, 132, 277–293.

Broder, A., & Schiffer, S. (2003b). Bayesian strategy assessment in multi-attribute decision making. Journal ofBehavioral Decision Making, 16, 193–213.

Broder, A. (in press). Outcome-based strategy classification. In A. Glockner , & C. L. M. Witteman (Eds.), Tracingintuition: Recent methods in measuring intuitive and deliberate processes in decision making. London: PsychologyPress & Routledge.

Budescu, D. V. (1993). Dominance analysis: A new approach to the problem of relative importance of predictors inmultiple regression. Psychological Bulletin, 114, 542–551.

Budescu, D. V. (2006). Confidence in aggregation of opinions from multiple sources. In K. Fiedler , & P. Juslin (Eds.),Information sampling and adaptive cognition (pp. 327–352). New York, NY; Cambridge University Press.

Busemeyer, J. R., & Townsend, J. T. (1993). Decision field theory: A dynamic cognition approach to decision making.Psychological Review, 100, 432–459.

Dawes, R. M. (1979). The robust beauty of improper linear models in decision making. American Psychologist, 34, 571–582.

Einhorn, H. J. (1971). Use of nonlinear, noncompensatory models as a function of task and amount of information.Organizational Behavior and Human Performance, 6, 1–27.

Elrod, T., Johnson, R. D., & White, J. (2004). A new integrated model of noncompensatory and compensatory decisionstrategies. Organizational Behavior and Human Decision Processes, 95, 1–19.

Erev, I., Roth, A. E., Slonim, R. L., & Barron, G. (2002). Combining a theoretical prediction with experimental evidenceto yield a new prediction: An experimental design with a random sample of tasks. Unpublished manuscript, ColumbiaUniversity and Faculty of Industrial Engineering and Management, The Technion, Haifa, Israel.

Erev, I., Wallsten, T. S., & Budescu, D. V. (1994). Simultaneous over- and underconfidence: The role of error in judgmentprocesses. Psychological Review, 101, 519–527.

Fishburn, P. C. (1974). Lexicographic order, utilities and decision rules. Management Science, 20, 1442–1471.Ford, J. K., Schmitt, N., Schechtman, S. L., Hults, B. M., & Doherty, M. L. (1989). Process tracing methods:

Contributions, problems and neglected research questions. Organizational Behavior and Human Decision Processes,43, 75–117.

Gigerenzer, G. (1997). Bounded rationality: Models of fast and frugal inference. Swiss Journal of Economics andStatistics, 133, 201–218.

Gigerenzer, G., & Goldstein, D. G. (1996). Reasoning the fast and frugal way: Models of bounded rationality.Psychological Review, 103, 650–669.

Gigerenzer, G., & Goldstein, D. G. (1999). Betting on one good reason: The take the best heuristic. In G. Gigerenzer ,P. M. Todd , & the ABC Research Group (Eds.), Simple heuristics that make us smart (pp. 75–95). Oxford, England:Oxford University Press.

Gigerenzer, G., & Hug, K. (1992). Domain-specific reasoning: Social contracts, cheating, and perspective change.Cognition, 43, 127–171.

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

S. Ayal and G. Hochman The Cognitive Processes Underlying Choice Behavior 473

Gigerenzer, G., Hoffrage, U., & Kleinbolting, H. (1991). Probabilistic mental model: A Brunswikian theory ofconfidence. Psychological Review, 98, 506–528.

Gigerenzer, G., Todd, P. M. & the ABC Research Group. (1999). Simple heuristics that make us smart. Oxford, England:Oxford University Press.

Glockner, A. (2006). Automatische Prozesse bei Entscheidungen (Automatic processes in decision making). Hamburg:Kovac. (in German).

Glockner, A. (2007). Does intuition beat fast and frugal heuristics? A systematic empirical analysis. In H. Plessner , C.Betsch , & T. Betsch (Eds.), Intuition in judgment and decision making (pp. 309–326). New York: Lawrence ErlbaumAssociates.

Glockner, A. (in press). Multiple measure strategy classification: Outcomes, decision times and confidence ratings. In A.Glockner , & C. L. M. Witteman (Eds.), Tracing intuition: Recent methods in measuring intuitive and deliberateprocesses in decision making. London: Psychology Press & Routledge.

Glockner, A., & Betsch, T. (2007). Evidence for multiple-reason decision making under time limits. Manuscript submittedfor publication.

Glockner, A., & Betsch, T. (2008a). Do people make decisions under risk based on ignorance? An empirical test of thepriority heuristic against cumulative prospect theory. Organizational Behavior and Human Decision Processes, 107,75–95.

Glockner, A., & Betsch, T. (2008b). Multiple-reason decision making based on automatic processing. Journal ofExperimental Psychology: Learning, Memory, and Cognition, 34, 1055–1075.

Hertwig, R., Barron, G., Weber, E. U., & Erev, I. (2004). Decision from experience and the effect of rare events.Psychological Science, 15, 534–539.

Hoffrage, U. (2005). Heuristics: Fast and frugal. In B. C. Everitt , & D. C. Howell (Eds.), Encyclopedia of statistics inbehavioral science (Vol. 2, pp. 795–799). Chichester, UK: Wiley.

Johnson, E. J., Schulte-Mecklenbeck, M., &Willemsen, M. C. (2008). Process models deserve process data: A commenton Brandstatter, Gigerenzer, and Hertwig (2006. Psychological Review, 115, 263–273.

Kahneman, D., & Tversky, A. (1979). Prospect theory: An analysis of decision under risk. Econometrica, 47, 263–291.

Lee, M. D., & Cummins, T. D. R. (2004). Evidence accumulation in decision making: Unifying the ‘take the best’ and‘rational’ models. Psychonomic Bulletin and Review, 11, 343–352.

Lopes, L. L., & Oden, G. C. (1999). The role of aspiration level in risky choice: A comparison of cumulative prospecttheory and SP/A theory. Journal of Mathematical Psychology, 43, 286–313.

Newell, B. R., & Shanks, D. R. (2003). Take the best or look at the rest? Factors influencing ‘‘one-reason’’ decisionmaking. Journal of Experimental Psychology: Learning, Memory, and Cognition, 29, 53–65.

Newell, B. R., Weston, N. J., & Shanks, D. R. (2003). Empirical tests of a fast and frugal heuristic: Not everyone ‘‘takes-the-best.’’ Organizational Behavior and Human Decision Processes, 91, 82–96.

Payne, J. W., Bettman, J. R., Coupey, E., & Johnson, E. J. (1992). A constructive process view of decision making:Multiple strategies in judgment and choice. Acta Psychologica, 80, 107–141.

Payne, J. W., Bettman, J. R., & Johnson, E. J. (1993). The adaptive decision maker. New York: Cambridge UniversityPress.

Rieger, M. O., & Wang, M. (2008). What is behind the priority heuristic? A mathematical analysis and comment onBrandstatter, Gigerenzer, and Hertwig (2006. Psychological Review, 115, 274–280.

Rieskamp, J., & Hoffrage, U. (1999). When do people use simple heuristics and how can we tell? In G. Gigerenzer , &P. M. Todd , & the ABC Research Group (Eds.), Simple heuristics that make us smart (pp. 141–167). Oxford, England:Oxford University Press.

Rieskamp, J., & Hoffrage, U. (2005). The use of simple heuristics: Inferences and preferences under time pressure.Manuscript submitted for publication.

Rieskamp, J., & Otto, P. E. (2006). SSL: A theory of how people learn to select strategies. Journal of ExperimentalPsychology: General, 135, 207–236.

Tversky, A. (1972). Elimination by aspects: A theory of choice. Psychological Review, 79, 281–299.Tversky, A., & Kahneman, D. (1992). Advances in prospect theory: Cumulative representation of uncertainty. Journal of

Risk and Uncertainty, 5, 297–323.von Neumann, J., & Morgenstern, O. (1944). Theory of game and economic behavior. Princeton: Princeton UniversityPress.

Wakker, P., Erev, I., & Weber, E. U. (1994). Comonotonic independence: The critical test between classical and rank-dependent utility theories. Journal of Risk and Uncertainty, 9, 195–230.

Winkler, R. L., & Poses, R. M. (1993). Evaluating and combining physicians’ probabilities of survival in an intensive careunit. Management Science, 39, 1526–1543.

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm

474 Journal of Behavioral Decision Making

Authors’ biographies:

Shahar Ayal is a postdoctoral fellow at Fuqua School of Business, Duke University. His research interests are heuristicsand biases, risk perception, unethical behavior, and the effect of inter-group relations on decision making.

Guy Hochman is a psychology doctoral student at the Technion—Israel Institute of Technology. His research interestsare heuristics and biases, process models, and physiological measures and their relation to decision-making processes.

Authors’ addresses:

Shahar Ayal, Duke University, The Fuqua School of Business, 1 Towerview Drive, Durham, NC 27708, USA

Guy Hochman, William Davidson Faculty of Industrial Engineering and Management. Technion - Israel Institute ofTechnology, Technion City, Haifa 32000, Israel.

Copyright # 2009 John Wiley & Sons, Ltd. Journal of Behavioral Decision Making, 22, 455–474 (2009)

DOI: 10.1002/bdm