Next Article in Journal
A Scent of Lemon—Seller Meets Buyer with a Noisy Quality Observation
Next Article in Special Issue
Bounded Memory, Inertia, Sampling and Weighting Model for Market Entry Games
Previous Article in Journal
Do I Really Want to Know? A Cognitive Dissonance-Based Explanation of Other-Regarding Behavior
Previous Article in Special Issue
Intergroup Prisoner’s Dilemma with Intragroup Power Dynamics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Loser Can Be a Winner: Comparison of Two Instance-based Learning Models in a Market Entry Competition

Dynamic Decision Making Laboratory, Carnegie Mellon University, Pittsburgh, PA 15213, USA
*
Author to whom correspondence should be addressed.
Games 2011, 2(1), 136-162; https://doi.org/10.3390/g2010136
Submission received: 21 December 2010 / Revised: 1 March 2011 / Accepted: 14 March 2011 / Published: 16 March 2011
(This article belongs to the Special Issue Predicting Behavior in Games)

Abstract

: This paper presents a case of parsimony and generalization in model comparisons. We submitted two versions of the same cognitive model to the Market Entry Competition (MEC), which involved four-person and two-alternative (enter or stay out) games. Our model was designed according to the Instance-Based Learning Theory (IBLT). The two versions of the model assumed the same cognitive principles of decision making and learning in the MEC. The only difference between the two models was the assumption of homogeneity among the four participants: one model assumed homogeneous participants (IBL-same) while the other model assumed heterogeneous participants (IBL-different). The IBL-same model involved three free parameters in total while the IBL-different involved 12 free parameters, i.e., three free parameters for each of the four participants. The IBL-different model outperformed the IBL-same model in the competition, but after exposing the models to a more challenging generalization test (the Technion Prediction Tournament), the IBL-same model outperformed the IBL-different model. Thus, a loser can be a winner depending on the generalization conditions used to compare models. We describe the models and the process by which we reach these conclusions.

1. Introduction

A choice prediction competition was organized by Erev, Ert, and Roth [1]. This modeling competition focused on decisions from experience in market entry games (hereafter, market entry competition, MEC, http://sites.google.com/site/gpredcomp/). The market entry games involved four interacting players who had to decide between entering a market (a risky alternative) and staying out of the market (a safe alternative) in a number of trials. The payoffs of entering the market decreased with the number of entrants, and were also subject to probabilistic influences on the outcomes. Human data from an estimation set was made available for researchers, who used it to calibrate their models. These models were then submitted to compete over the best predictive value in a new dataset called the competition set.

Our team (the co-authors of this paper) submitted two versions of the same cognitive model to the competition. The cognitive model submitted was developed according to the Instance-Based Learning Theory (IBLT [2]). One version of the IBL model assumed that the four players in the game had identical characteristics. As it will be explained later, this model, called IBL-same, included the same set of parameter values for each of the four players in the game. The other version of the IBL model, called IBL-different, assumed heterogeneity of the four players in the game, and included different sets of parameter values for each of the four players in the game. The IBL-different model won the runner-up prize of the competition among 25 other submissions, while IBL-same model achieved the 11th place.

The current paper reports the process and main lessons learned through the submission of the two versions of the same IBL model. First, we describe the MEC and behavioral methods used in the estimation and competition sets. Next, we summarize IBLT and describe the IBL model developed for the MEC. Next, we discuss the process by which the model parameters were determined in both the IBL-same and IBL-different models. We discuss the calibration (or fit) of each model to the estimation set, and present our a-priori expectations of the performance of the two models in the competition set of the MEC. Then, we discuss the actual performance of the two models in the competition set. We follow with discussing the unexpected results of the MEC, and show that the IBL-same (the loser) model outperforms the IBL-different (the winner) model under more challenging generalization conditions in a different dataset. To conclude, we discuss the main lessons learned from our participation in the MEC.

2. Market Entry Competition and Behavioral Methods

Each game in the MEC consists of four players who make individual choices in a number of trials. In each trial of a game, each of the four players decides individually between entering a risky market (risky alternative) or staying out (safe alternative). The payoff for player i if entering the market at trial t is

10 k ( E ) + G t

Where k is a parameter drawn (with equal probability) and with replacement from the set {2, 3, 4, 5, 6, 7}.

E is the number of entrants in trial t, and Gt is a binary gamble that yields “H with probability Ph; and L otherwise.” H is a positive number and L is a negative number, determined according to the algorithm described in Appendix 1 of Erev et al. [1].

The payoff for a player i if staying out in trial t is

round ( G t / S ) with p = 0.5 ; and round ( G t / S ) otherwise

Where S is a parameter drawn (with equal probability) and with replacement from the set {2, 3, 4, 5, 6}.

Thus, the payoff for a player depends on the player's own choice (to enter or stay out), the choices of the other players (E, such that the more people enter the market the lower the payoff from entry), and the trial's outcome of a gamble (Gt).

There were a total of 40 games used in the estimation set. These games were determined by a random selection of the parameters: k, S, H, Ph, and L (using the algorithm described in Appendix 1 in Erev et al. [1]). One hundred and twenty students participated in the estimation set. The set involved 8 sessions, each of which included between 12 and 20 participants. Each session used 10 of the 40 games, so that each subset of 10 games was run twice in a counterbalanced order. In each session, each participant was randomly matched with three other participants, and each of the 10 games was played for 50 trials. Participants did not receive a description of the payoff calculation, but they obtained feedback after each trial. Feedback included the payoff from their own choice and their “forgone” payoff (i.e., the payoff they would have obtained had they selected the other alternative).

Results were grouped for each game and the 50 trials played in each game were separated into a first and a second block of 25 trials each. The dependent measures of performance used in Erev et al. [1] are the average of the following for each of the two blocks:

  • Entry rate: the proportion of entry decisions

  • Efficiency: the mean observed payoffs

  • Alternation rate: the proportion of times players changed their choices (from entering to staying out or from staying out to entering) between trials

Thus, six statistics were used as dependent measures.

2.1. Competition Criteria and Dataset

The human data for the estimation set was made available to researchers ahead of time. They were allowed to analyze the data, study the observed behavior, and build their own models. After the submission deadline, the competition set was run using the same behavioral procedures used in the estimation set. The competition set involved the selection of 40 games different from those used in the estimation set, but the games were also determined by the same selection algorithm (Appendix 1 in [1]). The MEC focused on the models' predictions of the six statistics described above in the new 40 competition set problems. Currently, both the estimation and competition studies are publicly available, posted by the organizers in the MEC web page ( http://sites.google.com/site/gpredcomp/study-results/competition-study).

Twenty-five models were submitted to the competition and each of the models was evaluated using the mean squared deviation (described in detail in Erev et al. [1]) between the models' predictions and the observed performance in the competition set. Each model obtained a final score, the nMSD (normalized mean squared deviation). The nMSD is the mean of the six normalized mean squared deviations (MSD) for each of the six statistics described above. Each of the MSDs for each of the dependent measures was calculated in the following three steps: (1) compute the squared deviation between the model's prediction and the observed statistic in each of the 40 games; (2) compute the mean squared deviation over all the 40 games; and (3) normalize each score by the variable's estimated error variance. The model with the lowest nMSD score won the MEC.

3. An Instance-Based Learning Model

The cognitive model submitted to the MEC is based on a cognitive theory of decisions from experience, Instance-Based Learning Theory (IBLT), originally developed to explain and predict learning and decision making in dynamic decision-making environments [2].

IBLT proposes a key representation of cognitive information: an instance. An instance is a representation of each decision alternative, often consisting of three parts: a situation (a set of attributes that define the alternative), a decision, and an outcome resulting from making that decision in that situation. The theory also proposes a generic decision-making process that starts by recognizing and generating instances through interaction with an environment, and finishes with the reinforcement of the instances that led to good decision outcomes through feedback from the environment. The general decision-making process is explained in detail in Gonzalez et al. [2], and it involves the following steps: The recognition of a situation from an environment (a task) and the creation of decision alternatives; the retrieval of similar experiences from the past to make decisions, or the use of decision heuristics in the absence of similar experiences; the selection of the best alternative; and, the process of reinforcing positive experiences through feedback.

At each decision stage, IBLT selects an instance that has the highest utility (blended value, explained below). The different parts of an instance and the selection of an alternative are built through the general IBLT decision process: creating a situation from attributes in the task, creating an expected utility for making a decision, and updating the utility value according to the outcomes observed from an alternative. Instances corresponding to outcomes accumulate over time and their blended values depend on the availability of those instances in memory. This availability is measured by a statistical mechanism called Activation, originally developed in the ACT-R cognitive architecture [3].

3.1. The IBL Model for the MEC

IBL models are particular representations of IBLT for specific tasks. Many IBL models have been developed in a wide variety of tasks, including dynamically-complex tasks [2-5], training paradigms of simple and complex tasks [6,7], simple stimulus-response practice and skill acquisition tasks [8], repeated binary-choice tasks [9,10] among others. Although most of the IBL models developed are task specific, a recent IBL model showed that it generalizes well to multiple repeated-choice tasks that share the same task structure. The IBL model reported in Lejarraga, Dutt, and Gonzalez was built to predict performance in individual binary-choice tasks, and generalized accurately to choices in a repeated-choice task, probability-learning tasks, and repeated-choice tasks with changing probability of outcomes as a function of trials [10]. The IBL model for the market entry task is an extension of the model reported in Lejarraga et al. [10].

Instances in the repeated choice and MEC model are much simpler than in other IBL models, as the structure of these tasks is simple. Each instance consists of a label that identifies an alternative in the task and the outcome obtained (i.e., a button label and its observed outcome). For example (Enter, $4), is an instance in which the decision was to enter the market and the outcome as a result of that choice was $4. In the MEC, since participants also observed forgone payoffs, another similar instance was created for the obtained foregone payoff in each decision made, for example (Stay Out, $3).

In each trial t of a market-entry game, the process of selecting alternatives in the model starts with an inertia rule (Equation 1 below), which determines whether the previous choice in the task is repeated according to the surprise-triggers-change hypothesis by Nevo and Erev [12]. If the previous choice is not repeated, then the alternative with the highest blended value is selected (Equation 5 below). The blended value of an alternative is calculated in each trial t of the game and it is a transient value that depends on the outcome stored in an instance and the probability of retrieval of that instance from memory (Equation 6 below). Furthermore, the probability of retrieval of an instance from memory is a function of its activation in memory (activation is a function of the recency, frequency and noise in the retrieval of an instance) (Equation 7 below).

3.2. Inertia Mechanism

Erev et al. [1] report several behavioral regularities found in the estimation set in the MEC: Surprise-triggers-change from choosing one alternative to the other, and the presence of strong inertia in repeated choice. These two effects relate to sequential dependencies between choices in human data, measured with the Alternation rate. These sequential dependencies in choices over time have also been demonstrated by Biele, Erev, and Ert [11] and Nevo and Erev [12].

Our model in Lejarraga et al. [10] was not built to account for alternations, but rather to account for the proportion of risky choices in repeated-choice tasks. Given that many influential models of repeated choice have found a weak relationship between the generic measures of performance and sequential dependencies [13,14], we investigated how IBL models are able to account for sequential dependencies and the tradeoffs between the proportion of risky choices and the proportion of alternations in repeated-choice tasks [15]. To capture sequential dependencies in the MEC data, we built on the surprise-triggers-change hypothesis by Nevo and Erev [12] and proposed a new inertia mechanism that considers blended values instead of running averages. This mechanism is determined at the moment of making a choice in trial t+1 by a simple rule:

If the draw of a random value in the uniform distribution

U ( 0 , 1 ) < ( pInertia ) Surprise ( t )

Then

Repeat the choice as in the previous trial

Else

Select the alternative with the highest blended value as per Equation 2 (below)

The pInertia or the Probability of Inertia is a free parameter between 0 and 1 and initially defined at 0.30, according to Biele et al. [11]. The value of the Surprise(t) is assumed to depend on the gap (absolute difference) between an expectation of the outcome and the outcome actually received. In our inertia mechanism this is the absolute difference between the observed outcome and the blended value for that alternative. Since forgone payoffs are observed in the market entry games, the gap is calculated for the two alternatives, enter and stay out, as follows:

Gap ( t ) = ½ × ( | outcome Enter ( t 1 ) V Enter ( t 1 ) | + | outcome Stay out ( t 1 ) V Stay out ( t 1 ) | )

The outcome Enter (t–1) is the observed or foregone outcome obtained upon entering the market in the last trial and outcome Stay out (t–1) is the observed or foregone outcome obtained upon staying out. The V Enter (t–1) and V Stay out (t–1) are the blended values of the two alternatives obtained in the last trial (the calculation of blended values is defined below).

The surprise in trial t is normalized by the mean gap (in the first t−1 trials):

Surprise ( t ) = Gap ( t ) / [ Mean ( Gap ( t ) ) + Gap ( t ) ]

The Mean(Gap(t)) is defined over 50 trials of a market entry game as:

Mean ( Gap ( t ) ) = Mean ( Gap ( t 1 ) ) ( 1 1 / 50 ) + Gap ( t ) ( 1 / 50 )

Erev et al. [1] justified the gap-based formulation of surprise by the observation that the activity of certain dopamine-related neurons is correlated with the difference between average past payoff (or the blended value) and the present outcome. This assumption is the only extension to the model reported in Lejarraga et al. [10] needed to account for the sequential dependencies reflected in the proportion of alternations in the market entry games. Naturally, the higher the value of pInertia, the more the IBL model will repeat its choice.

3.3. The General IBLT Mechanisms

In making a choice, IBLT selects the alternative with the highest blended value, V [16] resulting from all instances belonging to an alternative. The blended value of alternative j is defined as

V j = i = 1 n p i x i

Where xi is the value of the observed (obtained or foregone) outcome in the outcome slot of an instance i corresponding to the alternative j, and pi is the probability of that instance's retrieval from memory (for the MEC as noted in Equation 2, the value of j is either to enter or to stay out). The blended value of an alternative (its utility) is the sum of all observed outcomes xi in the outcome slot of corresponding instances in memory, weighted by their probability of retrieval. In any trial t, the probability of retrieval of instance i from memory is a function of that instance's activation relative to the activation of all other instances corresponding to that alternative, given by

P i , t = e A i , t / τ j e A j , t / τ

Where τ is random noise defined as = σ × 2, and σ is a free noise parameter (more details below). Noise in equation 2 captures the imprecision of recalling instances from memory.

The activation of each instance in memory depends upon the Activation mechanism originally proposed in the ACT-R architecture [3]. A simplified version of the activation mechanism that relied on recency and frequency of use of instances in memory was sufficient to capture human choice behavior in several repeated-choice and probability-learning tasks [10]. For each trial t, Activation Ai,t of instance i is:

A i , t = ln ( t i { 1 , , t 1 } ( t t i ) d ) + σ . ln ( 1 γ i , t γ i , t )

Where d is a free decay parameter, and ti is the time period of a previous trial where the instance i was created or its activation was reinforced due to an outcome in the task corresponding to the instance in memory. The summation will include a number of terms that coincides with the number of times that an outcome has been observed in previous trials and that the corresponding instance i's activation has been reinforced in memory. Therefore, the activation of an instance corresponding to an observed outcome increases with the frequency of observation (i.e., by increasing the number of terms in the summation) and with the recency of those observations (i.e., by small differences in t - ti of outcomes that correspond to that instance in memory). The decay parameter d affects the activation of the instance directly, as it captures the rate of forgetting. In ACT-R, the d parameter is almost always set to 0.5. The higher the value of the d parameter, the faster the decay of memory, and the harder it is for the model to recall distant memories of its instances with outcomes.

The γi,t term is a random draw from a uniform distribution bounded between 0 and 1, and the σ . ln ( 1 γ i , t γ i , t ) term represents Gaussian noise important for capturing variability of human behavior. The σ is a free noise parameter that has no default value in ACT-R, but that it has been found to have a mean of 0.45 in many ACT-R studies [17]. Higher σ values imply greater noise in the retrieval of instances from memory.

3.4. Special Treatment of the First Trial

In the first trial of a game, the model has no past instances in its memory from which to calculate blended values. Therefore, the model makes a selection between two instances in memory for the first trial by assuming some initial blended values. Each initial blended value corresponds to one of the two alternatives, entering or staying out. The blended value of the pre-populated instances may represent the expectations that participants bring to the laboratory [10]. The choice of the blended values in the two instances was motivated from the observed entry rate in the first trial of the estimation set. We found that the observed entry rate was about 73% (>50%) in the first trial of the estimation set, i.e., more than 50% of the participants entered the market in the first trial. We speculate that one reason for the higher entry rate could be that the experiment was framed as a market entry competition. Due to this observation and the fact that the ratio of the blended values assigned to the two instances determines the entry rate in the first trial in the model, we assigned a +94 value as the blended value of the instance corresponding to the enter alternative and a +34 value to the blended value of the instance corresponding to the stay out alternative. As seen, the ratio of the value assigned to the instance corresponding to the “enter” alternative to the sum of the values assigned to both instances, i.e., 94/(94 + 43) computes to 73%, i.e., the observed entry rate in the first trial in the estimation set. In the first trial, a decision to enter or stay out is also based solely upon these blended values for each pre-populated instance. The inertia mechanism is used from the second trial onwards. As the value of V Enter (0) and V Stay out (0) do not exist, Gap (1) is calculated by replacing V Enter (0) and V Stay out (0) by 94 and 34 in Equation 2. Also, the Mean(Gap(1)) is taken to be a very small number close to 0 (= .00001). Our assumption on Mean(Gap(1)) is similar to that by Erev et al. [1].

4. Two Versions of the IBL Model: IBL-same and IBL-different in the Estimation Set

As described above, the IBL model for the MEC consists of three free parameters that define a decision maker: d, σ, and pInertia. We created two different submissions for the MEC: the IBL-same and IBL-different. Each of the two submissions included four identical copies of the same IBL model, where each copy represented one of the four simulated players in a market-entry game. Furthermore, each copy (or simulated player) had three free parameters: d, σ, and pInertia.

The only difference between the IBL-same and the IBL-different submissions is that the IBL-different model used a different set of model parameters for each of the four copies (or simulated players) of the IBL model. In contrast, the IBL-same used the same values of the three parameters in each of the four copies. Thus, the IBL-different allowed for 4 different values of d, 4 different values of σ, and 4 different values of pInertia. The IBL-same, in contrast, assumed the same d, σ, and pInertia values for all the four simulated participants in the model. The values of the parameters for both submissions were determined through an optimization process to fit the observed data in the estimation set.

4.1. Optimization of Parameters through a Genetic Algorithm

The parameter values of the IBL-different and IBL-same models were obtained through an optimization process that maximized the fit of the model's data to human behavior in the estimation set. The process involved an optimization of the parameter values for each model involving a genetic algorithm. The goal of the generic algorithm was to find the set of parameters that minimized the error between the model's predictions and the observed behavior. Specifically, the genetic algorithm attempted to minimize the normalized mean squared deviation (nMSD).

The genetic algorithm tests different combinations of parameters in a model to minimize the nMSD between predictions and observed behavior across the 40 problems in the estimation set of the MEC. First, different parameter combinations (N) are selected and tested. This first group of parameters combinations is the referred to as the first “generation.” Each test of a combination of parameters involves running the model multiple times (i.e., multiple simulated participants) and obtaining the mean prediction, which is then compared to the mean behavior across the six measures that determine the nMSD. The parameter combinations are then ranked from best fitting to worst fitting based upon the calculated nMSD values. After ranking, the best half of the parameter combinations are kept (N/2), and the rest (N/2) are discarded. The parameter combinations that are kept are then duplicated, bringing the number of parameter combinations back to the original amount (N). Then, the N parameter combinations are paired off with each other at random (thus forming N/2 pairs). Now, each parameter combination exchanges some of its adjustable parameter values with the corresponding parameter value of its partner (this is called “reproduction”). For example, suppose the following two three-parameter combinations have been paired off: (a1, b1, c1) and (a2, b2, c2). Then, due to the exchange of the adjustable parameter values a1 and a2 in the pair, the resulting parameter combinations will be (a2, b1, c1) and (a1, b2, c2). The exchange of parameters defines a new generation that is different from the previous generation but maintains characteristics (i.e., the parameter values) of the previous generation's best cases. After the exchange, a new set of N parameter combinations is tested in the model and new nMSDs are obtained. The process is repeated for 10,000 generations. This value is extremely large and thus ensures a very high level of confidence in the optimized parameter values obtained. We simulated 100 four-participant teams for each combination of parameters in the model during optimization to derive an nMSD. Once the optimization was completed we increased the number of four participant teams in the model from 100 to 1,000. This value ensures that a model's prediction for the dependent measures is stable and does not change much from generation-to-generation for the same parameter combination used in the model.

For the purpose of optimizing the two IBL models using the genetic algorithm, the d and σ parameters were varied between 0.0 and 10.0, and the pInertia parameter was varied between 0.0 and 1.0. The assumed range of variation of d and σ parameters in the models is large and ensures that the optimization does not miss the minimum nMSD value on account of a small range of parameter variation.

The fitted values of the three parameters for the IBL-same model were:

d = 1.97, σ = 1.17, and pInertia = 0.23

The fitted values of the three parameters for the four simulated participants in the IBL-different submission were:

Player 1: d = 3.00, σ = 1.08, and pInertia = 0.13

Player 2: d = 1.73, σ = 1.44, and pInertia = 0.63

Player 3: d = 1.22, σ = 1.16, and pInertia = 0.02

Player 4: d = 2.93, σ = 1.26, and pInertia = 0.22

Because a larger number of free parameters allows models greater flexibility, the IBL-different model fitted the human data on the estimation set slightly better (nMSD = 1.153) than the IBL-same model (nMSD = 1.308).

These predictions of the IBL models were obtained for their best fitting parameters (determined above) for a set of 1,000 simulated four-player teams. The choices of all participants were averaged to obtain the six dependent measures used to evaluate the competing models: Entry rate, Efficiency, and Alternation rate in the first block (B1, the first 25 trials) and the second block (B2, last 25 trials) of each game. The MSDs and the nMSD value of each model were obtained by the three-step procedure detailed above.

Table 1 summarizes the MSDs for both models in the estimation set. Detailed values of the six statistics for the IBL-same, IBL-different, and observed values per problem are included in Table A1 in the Appendix.

4.2. Expectations of IBL-same and IBL-different Performance in the MEC

Our principal motivation for the submission of two identical models that differed only in how the value of parameters for each simulated participant was treated (same or different per participant) was to explore the tradeoffs of complexity (in terms of number of parameters) and generalization [18].

Generalization is the process of predicting new findings from an existent model [19]. The MEC focused on the prediction of the six statistics a priori, implying that researchers could not use any information concerning the observed behavior in the competition set, since this was unavailable at the time of submission of models.

Although generalizing ability of a model was rewarded in the MEC, as well as in other recent model competitions [20], parsimony was not. For example, in a recent modeling competition of binary choice (i.e., Technion Prediction Tournament, TPT), the winner model in the “sampling paradigm” was a complex model made of 4 sub-models and 40 different free parameters [20]. Similarly, the winning model in the MEC involves 7 parameters where 6 out of the 7 parameters take a uniformly distributed range of values around a parameter mean. Both the MEC and the TPT assumed that by following the Generalization Criterion Method [19] as an evaluation procedure, parsimonious models (i.e., with a lesser number of parameters) would rank high. The authors of the Generalization Criterion Method suggest that, because the estimation and generalization sets are different “conditions,” simple models would generalize better than complex ones, an advantage that is absent in other evaluation methods like cross-validation which uses the same dataset split in different ways for calibration and testing of a model. We believe, and show evidence, that the estimation and competition sets in the MEC are not sufficiently different to produce the effect observed by Wang and Busemeyer [19].

Evidence suggests that models tradeoff complexity—which leads to accurate fit in the estimation set—with generalizing capacity (i.e., accurate predictions) [18], and often these dimensions tradeoff in non-linear ways [21,22]. Therefore, comparing models across these two dimensions is challenging.

Jae Myung, Mark Pitt, and colleagues have studied these tradeoffs in many different ways [21-23]. Some of the main conclusions that we can summarize from their work are that:

  • A complex model with many parameters can fit data better than a simple model with fewer parameters through over-fitting. Over-fitting occurs when a model captures not only the underlying phenomenon but also the noise and variability of a particular dataset. A model that captures noise of a data set would make poor predictions in unknown, generalization conditions.

  • Generalization involves predictive accuracy and the ability of a model to predict statistics of future, unseen data samples, while using the parameters derived in an original calibration data sample.

Given that the IBL-same and IBL-different models were exactly the same on their cognitive assumptions and principles of how decisions makers learn and make decisions in a market entry game, the only difference in the models is that different participants may recruit cognitive processes differently, rather than being homogenous participants. Because the IBL-different model treated the four simulated participants in a market-entry game as different individuals (with different parameters), it is possible that this model captured not only the underlying learning and decision-making process, but also the noise and variability of different individuals in the estimation set. Thus, it was expected that IBL-different model would fit the estimation set better than IBL-same model.

However, our main question was whether the generalization procedure used in the MEC would favor the simple IBL-same model or the more complex IBL-different model. On the one hand and given the diverse values of the parameters per player, the IBL-different model may have over-fitted the estimation set and thus predict the data in the competition set worse than the IBL-same model. On the other hand, the estimation and competition sets were similar in many aspects that would question how challenging a generalization would be: The problems in the competition set had the same structure as those in the estimation set; the problems were obtained with the same selection algorithm in both studies; and the participants, although different, were drawn from the same population in both studies. Thus, it is likely that there exist systematic sources of variation and correlated noise across the estimation and competition set, which would favor the more complex IBL-different over the parsimonious IBL-same model in the competition set.

5. Results of MEC: Competition Set

The IBL-same and IBL-different models were submitted to the MEC. The models were run using the parameters found in the estimation set (described in the previous section) and compared to the data obtained in the competition set. Following the same procedure as in the estimation set, the predictions for the competition set were obtained by averaging the choices for a set of 1,000 simulated four-player teams for the parameters determined in the estimation set. The evaluation of the models in the competition set was done by scoring the models in the same six dependent measures as those used to calibrate the models in the estimation set.

Table 2 reports a summary of the MSD scores and the nMSDs of IBL-different and IBL-same model on the competition set. Detailed scores of the six statistics for the IBL-same, IBL-different, and observed values per problem in the competition set are included in Table A2 in the Appendix.

The IBL-different nMSD value of 1.078 outperformed the nMSD value of the IBL-same model of 1.218. The nMSD values of the top 15 models tested in the competition data set are reported in ( http://sites.google.com/site/gpredcomp/8-competition-results-and-winners). The IBL-different model ranked in 3rd place while IBL-same ranked in 11th place.

5.1. Why Did the IBL-same Performed Worse than IBL-different?

The observation that the parsimonious IBL-same model was outperformed by the more complex IBL-different model suggests that the competition set was not sufficiently different to the estimation set to favor parsimony over complexity (or flexibility). We argue that the generalization condition of the MEC had the characteristics of a traditional cross-validation rather than a generalization as in Wang and Busemeyer [19].

According to the cross-validation method [24], a dataset is divided into two samples, one sample is used as the estimation set and the other set is used as the prediction (competition) set. The MEC did not follow this procedure but rather ran two different studies, at different times, with different participants, and thus, in this sense, the MEC competition set is not strictly cross-validation. The MEC competition set, however, is not strictly a generalization set either. A generalization test indicates that the sampling for the calibration (estimation) set should be restricted to exclusively new experimental conditions [18,19]. As discussed above, both the estimation and competition sets followed the same problem structure, same problem selection algorithm, and similar population of participants. In fact, upon analysis of human data in the MEC, we found no differences in the entry rate between the estimation set (55%) and the competition set (54%) (t(78) = 0.225, ns, r = 0.03). Similarly, we found no differences in the efficiency between the estimation set (−0.17) and the competition set (−0.21) (t(78) = 0.099, ns, r = 0.01). Furthermore, the alternation rate between the estimation set (22%) and competition set (25%) did not differ (t(78) = –1.479, ns, r = 0.17). These results suggest that the MEC's estimation and competition sets were similar on all three dependent measures. We, therefore, extend the tests of the IBL-same and IBL-different models with a more challenging generalization process: We use problems that are structurally different from the problems used in the MEC.

6. A Challenging Generalization: the Technion Prediction Tournament

Prior to the MEC, the Technion Prediction Tournament (TPT) was organized by Erev and colleagues [20]. In the TPT, competing models were evaluated following a similar generalization criterion method as in the MEC [19]. An IBL model that was similar to the IBL-same model in this paper has been shown to predict observed behavior in the TPT's “E-repeated” condition better than all models submitted as contestants to the TPT [10]. Given these results and the observation that the IBL-same model did not perform as well as the IBL-different model in the MEC's competition set, we generalized both models, IBL-same and IBL-different, to the “E-repeated” condition of the TPT.

The TPT consisted of 60 problems in an estimation set and 60 problems in a competition set. The problems involved a choice between a safe alternative that offered a medium (M) payoff with certainty and a risky alternative that offered a high (H) payoff with some probability (pH) and a low (L) payoff with the complementary probability. The M, H, pH and L were generated randomly, and a selection algorithm assured that the 60 problems in each study were different and differed in the domain (positive, negative, and mixed outcomes) and probability (high, medium, and low pH). The resulting set of problems in the three domains and the three probability values was large and representative of the diversity in the problems. In the “E-repeated” condition, participants made 100 repeated choices per problem from experience where each choice in a problem affected their payoff [25].

The E-repeated condition of the TPT involved different binary-choice problems that were similar in their structure to the MEC's games: each game for each decision maker involved two alternatives, where one is safe and the other is risky. However, the TPT's E-repeated games were dynamically different from the MEC's games. In the TPT problems, the risky alternative is consistent across all trials, i.e., the outcomes are drawn from the same probability distributions consistently across trials; whereas, in the MEC, the outcomes in the risky alternative for a player are a function of the decision choices of other players (the TPT involved a single decision maker, whereas the MEC involves four interacting players). Also, the outcomes from each alternative in the TPT could be positive, negative, or mixed and the number of outcomes was a maximum of two per alternative, whereas, in the MEC, the possible outcomes on alternatives were mixed and the number of outcomes could be up to eight per alternative. Finally, participants in the TPT did not observe foregone payoffs as participants in the MEC did. These differences between the MEC and TPT problems make the TPT a challenging generalization test for the IBL-different and IBL-same model developed for the MEC.

6.1. Adapting the IBL Model of the MEC to the TPT

The similarity of the problem structure between the MEC and TPT allows us to generalize the models from the MEC to the TPT without significant changes in the working of the model. In order to execute both versions of the IBL model in the TPT, we adapted the number of problems (from 40 in the MEC to 60 in the TPT per set), the number of trials (from 50 in the MEC to 100 in the TPT per game in each of the two sets, estimation and competition). Although the problems in the MEC involve foregone outcomes, and these are absent in the TPT, our IBL remains unchanged: Outcomes in IBL are processed in the same way, whether obtained or forgone. The difference in the availability of foregone outcomes in MEC and TPT is captured by the IBL model in the following way. Simulated participants for TPT generate fewer instances—and reinforce them fewer times—than simulated participants in the MEC. Thus, behavior emerging from the IBL model comes from the same cognitive processes assumed in information processing. The only adaptation necessary to evaluate the IBL model of the MEC in the TPT was the calculation of the gap in the inertia rule (Equation 2). Given the absence of forgone payoffs in the TPT, the calculation of the Gap(t) in Equation 2 changes to:

Equation 2—used in TPT:

Gap ( t ) = ( | outcome Enter ( t 1 ) V Enter ( t 1 ) | )

There are other ways in which we could modify the model to fit the TPT data better. For example, the propensities used in the MEC for the first choice would not apply to TPT where the games involved no market context (see section 3.4). Similarly, the inertia mode could be modified to capture the sensitivity of inertia to the lack of foregone payoffs in the TPT. However, the nature of generalization involves testing models in new problems without modifications. Thus, although we acknowledge that our IBL model could perform better in the TPT if we made some changes, we only pursued those changes that were strictly necessary to run the model in the TPT.

6.2. Results from the Generalization from MEC to TPT

We ran both versions of the IBL model in the TPT problems, pooling the 60 problems in the TPT's estimation set and the 60 problems in the TPT's competition set. The IBL-same and IBL-different models were run using the same parameters as the MEC estimation set (detailed above). The IBL-different model was run four times on each problem, i.e., each time with a specific set of parameters representing each of its 4 players.

We compared the two versions of the IBL model using two different statistics across the 120 problems: the average proportion of risky choices (R-rate) across the 120 problems and the average proportion of alternations (A-rate) across the 120 problems (the average in both the measures was taken over all trials and participants in each of the 120 problems). The R-rate and A-rate in the TPT are similar to the “Entry rate” and “Alternation rate” in the MEC.

Table 3 shows the MSD values obtained by the IBL-different and IBL-same models using the R-rate and A-rate measures across the 120 problems in the TPT. The table also shows the values of the MSDs obtained for each set of parameters representing each of the four players in the IBL-different model. The Total MSD is the sum of the MSD determined upon the R-rate and the MSD determined upon the A-rate. The “average of all four players” is not the average of the MSDs calculated for the four players in the first four rows of the table; rather, it is the MSD obtained by averaging the R-rate and A-rate over the four players and then calculating the MSD between the averaged R-rate and A-rate and the corresponding R-rate and A-rate observed in human data. Results shed light onto the parsimony-flexibility trade-off. The simpler IBL-same outperformed the more complex IBL-different in this generalization: The MSD upon the R-rate, MSD upon the A-rate, and the Total MSD for IBL-same are smaller than the corresponding MSDs for each of the players of the IBL-different model and also lower than the average MSD values of the four players. Detailed values of the R-rate and A-rate statistics for the IBL-same, IBL-different, and observed human values per problem are included in Table A3 in the Appendix.

7. Discussion

Cognitive models are particular examples of representing human behavior in a particular task. Modeling competitions provide multiple advantages to account for the observed behavior, including precise and quantitative predictions provided by cognitive models in tasks of interest, and submission of models based upon different mathematical and cognitive approaches [20]. Moreover, because the organizers of the competitions make results publicly available, the contribution of these competitions is of transcendental scientific value. Researchers are able to continue the study of choice behavior independently of the competition's deadline.

Cognitive models of decisions from experience are becoming more popular in many fields including cognitive science, behavioral economics, and social sciences in general. The proliferation of different cognitive models that attempt to explain human behavior in simple choice tasks such as repeated binary choice, probability learning, and others has highlighted one major problem: many of these cognitive models are developed specifically for the tasks where they are tested. These models tend to lack common psychological theoretical ground and thus show difficulty in predicting behavior across different tasks [10]. Unifying theories of cognition still needs to be recognized as one important aim of science [26], and model competitions that stress model generalization should pursue this objective.

In this research, we present a cognitive model submitted to the choice prediction competition in market-entry games [1]. This model has been built based on IBLT, a theory of decisions from experience in dynamic tasks [2]. The important characteristic of the IBL model presented in this paper is that it is not an isolated attempt to predict choice in market-entry games, only. The IBL model presented in this study has been used to predict behavior in multiple choice tasks including binary-choice tasks, probability-learning tasks, tasks with changing probabilities as a function of trials, choice tasks with more than two alternatives, and choice through sampled and repeated presentation of experience [6,10,15].

Two versions of the same IBL model were submitted to the MEC. IBL-same assumed that the participants in the market-entry games were homogeneous, in that all participants were defined by the same set of parameter values in the IBL model (i.e., memory decay, noise, and inertia). In contrast, IBL-different assumed that the four participants of the market-entry games were unique, in that each participant was defined by an independent set of parameters. The results of fitting the model parameters to the estimation set indicate that the parameters in the IBL-same model are distinct and different from the parameters found in any of the four types of individuals in the IBL-different model.

The combination of parameters in the IBL-same model indicate a high value of d (1.97 compared to the 0.5 default value in ACT-R), a high value of noise (1.17 compared to a common value of 0.45 in ACT-R), and a relatively low value of inertia (0.23 compared to the 0.30 value found from human data in Biele et al. [11]). This set of parameters suggests a faster memory decay and thus reliance of recent experiences, variability in the retrieval of information from memory, and lower repetition of past choices over repeated trials of a game. The four individuals in the IBL-different model vary from one another. For example, Player 1 is a type of individual that relies much more upon recent memories and repeats the past choices much less than the average individual in the IBL-same model. Player 2, in contrast, although relies in recent past choices to the same extent as the average individual in the IBL-same model, it tends to show stronger inertia and repeats the previous choice more often (0.63) than the average individual in the IBL-same model.

We were curious to find out which of the two versions of the IBL model would provide a better representation of human behavior in the generalization conditions of the MEC. On the one hand, IBL-same is a more parsimonious model than the IBL-different (i.e., in terms of the number and homogeneity of free parameters). The assumption of homogeneous participants is a simplification that is common in cognitive modeling. In fact, most cognitive models aim at predicting the “average behavior” of a group of individuals, rather than at predicting behavior of a single individual. Yet, although the IBL-different model may be a more complex model, it may also be more realistic, because it assumes that participants in the market-entry games are not necessarily equal.

The IBL-different model resulted in superior match to human behavior in the generalization conditions of the MEC than the IBL-same model. However, many questions emerged from this result. The superior performance of the IBL-different model may be due to the similarity between conditions in the estimation and competition studies. Therefore, we pursued a stronger generalization test for the IBL-same and IBL-different models: We tested both versions of the IBL model in the problems used for the TPT modeling competition [20]. The problems used in the TPT are structurally different from the market entry games, and therefore present a challenging generalization scenario. The results of this generalization show the superior performance of the IBL-same over the IBL-different model.

We derive three main lessons from our participation in the MEC:

  • Lesson 1: Parsimonious models are favored over complex models when the estimation and competition conditions are sufficiently different.

  • Lesson 2: Overly complex models may appear to generalize well in generalization conditions that maintain the same task structure as the estimation conditions.

  • Lesson 3: Simpler models that generalize better than complex ones may be limited at capturing individual differences in human behavior.

Additionally, we believe there are still questions regarding the predictions and dynamics of behavior in the IBL model. For example, we observed interesting interactions among the six different statistics of the entry rate, efficiency (averaged observed payoffs), and alternation rate. In different studies, we have found that there are important tradeoffs to consider between fitting human behavior according to the proportion of entry decisions and the rate of alternations [15]: the rate of alternation generally falls sharply over trials depicting a rapid learning, yet the overall entry rate might remain the same or decrease slightly.

As discussed above, the MEC data sets are important sources for analyses of human behavior and model evaluation. Future research should explore how the aggregate learning behavior contrasts individual learning behavior. This approach would shed light on the tradeoffs faced when modeling individual or average behavior. This approach would help us understand the advantages of the IBL-different and the IBL-same versions of our model more accurately.

The results presented in this study lead us to we foresee future model competitions that involve generalization across different tasks, and that evaluate models by weighting individual and average performance across multiple tasks.

Table 1. Summary of fit results in the Estimation set.
Table 1. Summary of fit results in the Estimation set.
ModelMSD Entry rateMSD EfficiencyMSD Alternation ratenMSD



1st half2nd half1st half2nd half1st half2nd half
IBL-different0.9741.3020.9471.2391.3201.1341.153
IBL-same1.0741.4171.1361.5081.2291.4871.308
Table 2. Summary of results in the Competition set.
Table 2. Summary of results in the Competition set.
ModelMSD Entry rateMSD EfficiencyMSD Alternation ratenMSD



1st half2nd half1st half2nd half1st half2nd half
IBL-different1.4671.3120.9680.9280.5561.2381.078
IBL-same1.6261.3821.0610.9470.5891.7051.218
Table 3. Performance of the IBL-different and IBL-same models in the TPT.
Table 3. Performance of the IBL-different and IBL-same models in the TPT.
ModelParameters1MSD R-rateMSD A-rateTotal MSD
IBL–differentPlayer 1d=3.00,
σ=1.08,
pInertia=0.13
0.0750.0050.080
Player 2d=1.73,
σ=1.44,
pInertia=0.63
0.0330.0050.039
Player 3d=1.22,
σ=1.16,
pInertia=0.02
0.0210.4340.460
Player 4d=2.93,
σ=1.26,
pInertia=0.22
0.0490.0040.053
Average of all 4 Players0.0230.0030.025
IBL–sameIBL–loserd=1.97,
σ=1.17,
pInertia=0.23
0.0180.0020.021

Note 1: These parameters were determined in the estimation set of the MEC; thus, the two models are generalized using the MEC parameters in the TPT.

Acknowledgments

This research is supported by the Defense Threat Reduction Agency (DTRA) grant number: HDTRA1-09-1-0053 to Cleotilde Gonzalez and Christian Lebiere. The authors would like to thank Hau-yu Wong (Dynamic Decision Making Lab) for help with proofreading.

Appendix

Table A1. The 40 market entry games in the estimation set. Problem parameters, observed choice rates, and predictions by the IBL-same and IBL-different.
Table A1. The 40 market entry games in the estimation set. Problem parameters, observed choice rates, and predictions by the IBL-same and IBL-different.
StudyprobkphhlsfObserved choicesIBL-sameIBL-different

Ent1Ent2Eff1Eff2Alt1Alt2Ent1Ent2Eff1Eff2Alt1Alt2MSDEnt1Ent2Eff1Eff2Alt1Alt2MSD
Est.120.0470–350.710.802.772.660.160.160.760.792.632.670.170.100.690.750.782.672.720.180.110.58
Est.220.2330–940.550.622.642.750.250.230.610.582.462.460.260.220.650.600.582.512.520.260.220.52
Est.320.671–230.880.942.392.240.100.040.890.962.322.140.100.030.090.880.952.362.190.110.050.03
Est.420.7330–8040.710.642.582.570.280.270.660.652.312.360.260.240.600.650.622.372.400.270.250.58
Est.520.820–8050.660.672.502.670.290.270.700.692.312.360.230.211.250.700.672.362.420.230.230.91
Est.620.834–2030.730.822.452.500.240.180.780.792.402.380.190.160.750.770.782.442.420.200.170.58
Est.720.946–9050.860.872.342.380.130.110.810.832.272.260.140.110.500.810.812.332.350.150.120.71
Est.820.951–2050.860.912.482.310.120.080.850.902.362.230.130.070.070.840.892.412.300.140.080.13
Est.920.964–9030.870.902.362.340.140.080.840.872.262.220.120.090.280.830.852.312.290.120.100.51
Est.1030.170–840.420.481.221.110.290.250.470.420.981.070.250.201.270.460.441.101.220.240.191.11
Est.1130.99–8040.800.73–0.330.290.180.250.750.74–0.25–0.210.190.161.530.740.72–0.120.040.190.181.20
Est.1230.917–7060.760.830.10–0.410.190.120.750.75–0.23–0.260.180.150.950.750.73–0.15–0.040.190.171.58
Est.1340.0660–420.420.410.520.840.220.150.460.410.100.430.250.190.870.450.420.240.550.240.180.50
Est.1440.240–1040.480.46–0.340.040.310.310.450.40–0.240.020.280.231.350.460.43–0.130.060.270.240.98
Est.1540.3120–940.490.44–0.070.300.340.380.510.45–0.46–0.140.290.252.690.510.46–0.37–0.120.290.262.22
Est.1640.64–620.560.58–0.27–0.260.220.260.550.51–0.39–0.120.280.241.030.550.52–0.39–0.140.290.261.08
Est.1740.640–6030.580.55–0.96–0.200.280.250.560.54–1.12–0.820.290.270.700.570.53–1.06–0.680.300.280.58
Est.1840.733–820.570.55–0.290.090.240.200.560.52–0.42–0.120.270.240.500.560.52–0.38–0.090.280.250.58
Est.1940.820–8020.640.63–1.30–1.210.280.270.640.60–1.79–1.340.250.230.670.630.60–1.64–1.170.250.240.47
Est.2040.91–960.530.480.120.630.210.160.540.50–0.020.500.230.140.180.530.500.080.470.230.170.16
Est.2140.963–7030.650.62–0.84–0.380.230.180.620.60–0.93–0.590.220.150.320.630.61–0.89–0.610.220.160.19
Est.2250.0280–230.360.310.240.640.170.170.400.33–0.340.430.220.121.280.390.33–0.170.440.220.140.85
Est.2350.0790–730.390.24–0.810.340.190.130.350.29–0.65–0.290.230.171.400.360.30–0.50–0.130.220.171.30
Est.2450.5380–9050.650.58–3.41–2.440.270.360.520.50–2.35–1.940.300.285.220.520.51–2.31–1.940.300.284.87
Est.2550.81–420.450.42–0.310.110.200.180.450.40–0.540.090.230.160.310.450.41–0.490.040.240.190.32
Est.2650.884–3030.520.49–0.95–0.570.220.210.500.46–1.22–0.580.250.190.460.500.46–1.11–0.560.250.200.33
Est.2750.935–7040.570.57–1.63–1.430.270.200.540.50–1.70–1.100.230.171.050.540.50–1.65–1.020.240.180.95
Est.2860.190–1050.260.27–0.130.070.220.190.330.27–1.51–0.950.230.174.350.350.30–1.39–0.880.230.194.05
Est.2960.1930–730.390.32–1.35–0.450.270.260.380.32–1.71–1.070.270.220.930.390.33–1.63–1.010.270.220.69
Est.3060.2950–2030.470.48–2.74–2.430.380.360.420.38–2.44–1.830.290.254.510.430.39–2.33–1.830.290.264.00
Est.3160.467–660.380.34–0.90–0.380.230.240.390.34–1.32–0.730.260.220.600.400.34–1.29–0.710.270.230.59
Est.3260.576–840.440.39–1.56–0.590.260.270.400.35–1.37–0.780.270.220.660.400.35–1.33–0.780.280.240.54
Est.3360.8220–9030.630.55–5.33–3.140.260.210.600.56–5.04–4.100.260.231.580.600.55–4.79–3.780.250.231.12
Est.3460.888–6040.570.50–3.30–1.960.160.190.490.46–2.61–1.870.260.202.690.510.46–2.71–1.840.260.212.42
Est.3570.0690–640.310.35–1.40–1.430.290.210.280.21–1.62–0.930.210.144.150.290.23–1.43–0.860.210.153.31
Est.3670.2130–830.390.31–2.20–1.040.300.230.360.30–2.41–1.580.270.210.770.360.30–2.28–1.530.270.220.58
Est.3770.580–8050.510.55–4.18–4.780.340.320.490.46–4.86–4.090.300.272.620.490.46–4.70–3.980.300.292.37
Est.3870.699–2050.460.34–2.62–0.880.250.230.400.35–2.40–1.550.270.221.120.410.36–2.33–1.560.280.231.19
Est.3970.817–3020.410.34–2.25–0.930.220.210.400.36–2.29–1.570.270.220.960.400.36–2.28–1.480.280.230.97
Est.4070.911–1020.340.27–0.71–0.300.190.170.330.25–1.040.080.210.120.760.330.26–0.920.050.210.130.46
Average0.560.54–0.390.040.230.210.540.51–0.52–0.120.240.191.310.540.52–0.44–0.060.240.201.15
Table A2. The 40 market entry games in the competition set. Problem parameters, observed choice rates, and predictions by the IBL-same and IBL-different.
Table A2. The 40 market entry games in the competition set. Problem parameters, observed choice rates, and predictions by the IBL-same and IBL-different.
StudyprobkphhlsfObserved choicesIBL–sameIBL-different


Ent1Ent2Eff1Eff2Alt1Alt2Ent1Ent2Eff1Eff2Alt1Alt2MSDEnt1Ent2Eff1Eff2Alt1Alt2MSD
Comp.120.0470–330.690.782.852.750.170.150.760.792.632.670.170.110.750.750.772.672.720.180.110.51
Comp.220.189–230.810.822.482.570.190.210.850.902.442.370.130.082.440.840.882.462.420.140.091.73
Comp.320.240–1020.530.502.622.510.290.280.590.572.442.430.270.231.140.580.572.482.500.270.230.98
Comp.420.336–360.750.812.752.570.160.140.820.862.492.450.150.110.990.820.852.532.490.160.120.68
Comp.520.43–250.900.952.292.180.130.080.880.962.342.170.110.030.240.870.952.372.200.110.040.21
Comp.620.952–4030.870.932.242.230.110.060.840.882.312.240.130.090.490.830.862.362.300.140.100.80
Comp.720.972–6050.880.932.332.280.110.040.850.902.312.190.120.060.260.840.882.372.270.120.080.50
Comp.830.0390–330.470.531.431.500.280.260.570.541.151.400.230.152.270.560.541.241.450.230.151.96
Comp.930.19–120.650.701.060.880.170.160.680.670.680.860.220.160.470.670.670.770.900.220.170.47
Comp.1030.332–150.690.690.700.810.260.200.690.700.670.760.200.111.020.690.700.730.780.200.130.76
Comp.1130.3690–5030.400.371.271.260.280.300.490.470.730.860.300.272.220.490.480.800.910.300.282.27
Comp.1230.4710–920.560.571.261.390.300.260.620.580.660.790.280.260.960.610.580.710.830.290.270.84
Comp.1330.57–750.630.611.011.080.260.250.640.610.620.720.260.240.260.630.610.720.760.270.250.17
Comp.1440.0740–330.420.450.740.590.280.210.480.430.060.500.240.181.010.480.430.180.570.240.180.83
Comp.1540.449–720.540.54–0.35–0.120.330.360.540.50–0.53–0.270.290.260.990.540.50–0.45–0.190.300.280.77
Comp.1640.467–650.560.50–0.510.390.270.260.540.51–0.45–0.190.280.240.380.540.50–0.40–0.160.290.260.37
Comp.1740.4710–950.450.510.370.150.350.340.540.50–0.57–0.320.290.262.390.540.50–0.52–0.290.300.282.07
Comp.1840.537–860.530.540.01–0.140.270.280.550.51–0.52–0.300.280.250.450.550.51–0.55–0.280.290.270.41
Comp.1940.829–4020.710.65–1.98–1.140.250.260.620.59–1.35–1.070.260.231.500.630.59–1.36–0.910.260.241.36
Comp.2040.8610–6020.720.69–2.08–1.720.250.230.640.62–1.66–1.390.240.211.310.660.62–1.71–1.190.230.221.28
Comp.2140.888–6040.770.76–2.72–2.690.240.210.670.64–1.89–1.580.230.194.090.660.64–1.73–1.380.230.205.03
Comp.2250.295–260.400.37–0.100.250.230.260.440.39–0.610.010.240.181.010.440.39–0.51–0.020.250.200.74
Comp.2350.3380–4050.420.41–0.92–1.160.330.320.450.42–1.65–1.200.300.260.880.450.42–1.53–1.180.300.270.66
Comp.2450.3690–5060.460.36–1.30–0.540.340.310.460.43–1.71–1.310.300.271.450.460.43–1.63–1.270.300.271.36
Comp.2550.427–560.450.46–0.41–0.390.230.220.450.40–0.93–0.460.270.230.620.460.41–0.95–0.490.280.250.75
Comp.2650.62–320.390.370.100.220.280.210.450.40–0.570.020.250.190.900.450.40–0.55–0.040.250.210.92
Comp.2750.674–830.500.44–1.04–0.410.260.240.460.42–0.88–0.400.270.220.250.470.42–0.85–0.380.270.240.18
Comp.2850.918–8060.570.56–1.75–1.490.200.220.600.56–2.68–2.060.230.181.230.610.57–2.63–2.020.230.181.15
Comp.2960.0860–550.270.30–0.11–0.040.200.180.350.28–1.33–0.690.230.172.230.350.29–1.17–0.540.230.171.81
Comp.3060.1250–760.410.30–1.40–0.210.270.220.350.28–1.46–0.890.240.181.100.350.30–1.29–0.750.240.190.75
Comp.3160.460–4050.500.46–3.03–2.450.350.360.460.42–2.98–2.330.300.271.130.470.44–3.01–2.500.300.280.79
Comp.3260.5670–9020.580.57–4.49–3.980.370.300.520.49–3.87–3.140.300.282.410.520.49–3.67–3.120.300.282.53
Comp.3360.636–1050.390.40–0.84–0.870.310.270.410.36–1.50–0.810.270.220.850.420.36–1.46–0.810.270.230.69
Comp.3470.280–2050.390.34–3.14–1.900.240.230.360.32–2.82–2.040.270.220.270.370.34–2.64–2.090.270.230.31
Comp.3570.370–3050.430.48–2.72–3.470.320.320.410.37–3.36–2.630.290.252.650.420.38–3.28–2.660.290.262.20
Comp.3670.3320–1060.400.39–2.59–1.920.320.330.380.32–2.55–1.700.280.231.510.380.33–2.53–1.780.280.250.99
Comp.3770.445–460.360.30–1.03–0.200.230.160.330.27–1.31–0.490.230.170.360.330.27–1.23–0.520.230.180.33
Comp.3870.580–8030.520.49–4.86–3.630.360.360.490.46–4.76–3.950.300.281.110.500.46–4.83–3.880.310.280.85
Comp.3970.881–750.330.28–1.01–0.140.240.190.320.25–0.960.230.200.091.130.320.26–0.860.140.200.120.64
Comp.4070.982–8040.340.32–0.88–0.360.240.240.350.28–1.270.030.210.102.030.350.28–1.19–0.030.210.121.48
Average0.540.54–0.34–0.080.260.240.550.52–0.57–0.160.240.191.220.550.52–0.50–0.130.240.201.08

Note: Ent1 and Ent2 denote entry rates in the first and second half, respectively. Eff1 and Eff2 denote Efficiency scores in the first and second half. Similarly, Alt1 and Alt2 denote alternation rates in the first and second halves, respectively.

Table A3. The R-rate and A-rate in human data and the IBL-different and IBL-same models across 120 problems in estimation (Est.) and competition (Comp.) study in the TPT.
Table A3. The R-rate and A-rate in human data and the IBL-different and IBL-same models across 120 problems in estimation (Est.) and competition (Comp.) study in the TPT.
StudyProblemHigh (H)pHLow (L)Medium (M)R-rate (Human)Average R-rate (IBL-different)1R-rate (IBL-same)A-rate (Human)Average A-rate (IBL-different)2A-rate (IBL-same)
Est.1–8.70.06–22.8–21.40.250.230.230.170.170.15
Est.2–2.20.09–9.6–8.70.270.220.250.170.160.15
Est.3–20.1–11.2–9.50.250.230.210.160.160.15
Est.4–1.40.02–9.1–90.330.300.390.140.160.16
Est.5–0.90.07–4.8–4.70.370.370.380.130.180.16
Est.6–4.70.91–18.1–6.80.630.390.440.210.170.14
Est.7–9.70.06–24.8–24.20.300.400.300.170.170.16
Est.8–5.70.96–20.6–6.40.660.360.470.170.140.14
Est.9–5.60.1–19.4–18.10.310.290.240.090.190.16
Est.10–2.50.6–5.5–3.60.340.230.280.120.170.15
Est.11–5.80.97–16.4–6.60.610.640.540.120.120.14
Est.12–7.20.05–16.1–15.60.250.270.300.090.180.16
Est.13–1.80.93–6.7–20.440.440.390.110.150.15
Est.14–6.40.2–22.4–180.200.170.180.150.160.14
Est.15–3.30.97–10.5–3.20.160.300.320.100.140.15
Est.16–9.50.1–24.5–23.50.390.250.270.120.170.16
Est.17–2.20.92–11.5–3.40.470.440.440.130.130.14
Est.18–1.40.93–4.7–1.70.410.440.440.090.170.15
Est.19–8.60.1–26.5–26.30.490.310.380.180.160.17
Est.20–6.90.06–20.5–20.30.250.360.370.140.190.17
Est.211.80.6–4.11.70.080.170.160.100.120.13
Est.2290.97–6.79.10.140.310.280.110.170.13
Est.235.50.06–3.4–2.60.280.190.250.150.140.15
Est.2410.93–7.10.60.460.360.390.160.150.14
Est.2530.2–1.3–0.10.210.230.230.130.150.15
Est.268.90.1–1.4–0.90.230.200.290.120.120.16
Est.279.40.95–6.38.50.670.390.440.140.150.13
Est.283.30.91–3.52.70.580.240.390.170.120.14
Est.2950.4–6.9–3.80.390.300.240.170.170.15
Est.302.10.06–9.4–8.40.330.290.240.120.150.15
Est.310.90.2–5–5.30.880.390.550.090.180.17
Est.329.90.05–8.7–7.60.210.290.240.060.160.15
Est.337.70.02–3.1–30.280.340.380.100.170.15
Est.342.50.96–22.30.520.450.460.130.150.15
Est.359.20.91–0.78.20.560.420.380.090.150.13
Est.362.90.98–9.42.90.340.450.370.230.160.14
Est.372.90.05–6.5–5.70.300.250.250.170.170.15
Est.387.80.99–9.37.60.620.490.510.090.160.14
Est.396.50.8–4.86.20.320.170.200.080.140.12
Est.4050.9–3.84.10.460.340.380.080.140.13
Est.4120.10.956.519.60.500.360.380.200.120.11
Est.425.20.51.45.10.080.230.160.100.150.13
Est.43120.52.490.170.140.170.160.150.12
Est.4420.70.99.119.80.440.280.320.190.140.11
Est.458.40.071.21.60.200.300.290.120.160.15
Est.4622.60.47.212.40.410.230.190.200.140.12
Est.4723.40.937.622.10.720.290.370.140.120.10
Est.4817.20.0955.90.240.240.230.120.140.14
Est.4918.90.96.717.70.570.290.340.080.120.11
Est.5012.80.044.74.90.260.370.320.060.180.15
Est.5119.10.034.85.20.220.280.280.070.150.14
Est.5212.30.911.312.10.410.300.280.120.160.12
Est.536.80.936.70.410.290.330.110.140.14
Est.5422.60.39.2110.600.280.240.150.120.14
Est.556.40.090.51.50.280.240.230.120.140.14
Est.5615.30.065.97.10.170.220.210.100.160.14
Est.575.30.91.54.70.660.440.430.120.140.14
Est.5821.90.58.112.60.470.240.230.100.130.12
Est.5927.50.79.221.90.420.180.200.110.080.09
Est.604.40.20.71.10.380.240.300.120.130.16
Comp.1–0.30.96–2.1–0.30.330.380.400.280.160.15
Comp.2–0.90.95–4.2–10.50.400.410.10.160.15
Comp.3–6.30.3–15.2–12.20.240.230.210.140.160.15
Comp.4–100.2–29.2–25.60.320.230.200.130.160.15
Comp.5–1.70.9–3.9–1.90.450.380.410.190.160.15
Comp.6–6.30.99–15.7–6.40.680.490.480.180.160.15
Comp.7–5.60.7–20.2–11.70.370.300.290.270.150.14
Comp.8–0.70.1–6.5–60.270.300.290.130.170.16
Comp.9–5.70.95–16.3–6.10.430.390.420.120.150.14
Comp.10–1.50.92–6.4–1.80.440.370.390.150.150.15
Comp.11–1.20.02–12.3–12.10.260.350.350.110.170.16
Comp.12–5.40.94–16.8–6.40.550.420.460.170.150.14
Comp.13–20.05–10.4–9.40.110.250.240.10.160.15
Comp.14–8.80.6–19.5–15.50.660.320.310.160.160.15
Comp.15–8.90.08–26.3–25.40.190.290.270.130.170.16
Comp.16–7.10.07–19.6–18.70.340.280.260.120.170.16
Comp.17–9.70.1–24.7–23.80.370.280.280.190.170.16
Comp.18–40.2–9.3–8.10.340.260.240.220.160.16
Comp.19–6.50.9–17.5–8.40.490.400.430.180.150.14
Comp.20–4.30.6–16.1–4.50.080.150.140.10.130.12
Comp.2120.1–5.7–4.60.110.250.230.080.160.15
Comp.229.60.91–6.48.70.410.320.330.150.130.12
Comp.237.30.8–3.65.60.390.270.270.120.140.13
Comp.249.20.05–9.5–7.50.080.220.200.070.150.14
Comp.257.40.02–6.6–6.40.190.340.340.110.160.16
Comp.266.40.05–5.3–4.90.20.310.300.10.160.16
Comp.271.60.93–8.31.20.50.350.370.140.150.14
Comp.285.90.8–0.84.60.580.300.300.260.140.14
Comp.297.90.92–2.370.510.380.400.140.140.13
Comp.3030.91–7.71.40.410.400.430.180.140.13
Comp.316.70.95–1.86.40.520.390.410.110.150.14
Comp.326.70.93–55.60.490.390.420.110.140.13
Comp.337.30.96–8.56.80.650.410.420.080.140.13
Comp.341.30.05–4.3–4.10.30.340.340.10.170.16
Comp.3530.93–7.22.20.440.390.420.110.140.14
Comp.3650.08–9.1–7.90.090.250.230.070.160.15
Comp.372.10.8–8.41.30.280.240.240.230.140.13
Comp.386.70.07–6.2–5.10.290.250.230.140.160.15
Comp.397.40.3–8.2–6.90.580.320.300.20.160.16
Comp.4060.98–1.35.90.610.440.450.150.150.14
Comp.4118.80.87.615.50.520.310.310.110.130.12
Comp.4217.90.927.217.10.480.350.370.070.130.12
Comp.4322.90.069.69.20.880.570.580.070.160.15
Comp.44100.961.79.90.560.370.380.110.150.13
Comp.452.80.812.20.480.380.410.20.160.15
Comp.4617.10.16.980.320.240.220.120.150.14
Comp.4724.30.049.710.60.250.240.220.130.150.13
Comp.4818.20.986.918.10.590.410.410.140.140.12
Comp.4913.40.53.89.90.130.200.180.130.130.12
Comp.505.80.042.72.80.350.360.360.160.160.15
Comp.5113.10.943.812.80.520.360.370.090.140.13
Comp.523.50.090.10.50.260.300.290.120.160.15
Comp.5325.70.18.111.50.110.180.150.090.130.12
Comp.5416.50.016.970.180.360.360.140.160.15
Comp.5511.40.971.9110.660.460.490.10.140.13
Comp.5626.50.948.325.20.530.370.380.10.100.09
Comp.5711.50.63.77.90.450.260.240.270.140.13
Comp.5820.80.998.920.70.630.460.460.170.130.12
Comp.5910.10.34.260.320.240.220.190.150.14
Comp.6080.920.87.70.440.340.350.140.140.13
Average0.3880.3170.3240.1350.1510.141

Note: 1, 2 The reported R-rate and A-rate in the IBL-different model is obtained by averaging the R-rate and A-rate for each of the 4 players.

References and Notes

  1. Erev, I.; Ert, E.; Roth, A.E. A choice prediction competition for market entry games: An introduction. Games 2010, 1, 117–136. [Google Scholar]
  2. Gonzalez, C.; Lerch, J.F.; Lebiere, C. Instance-based learning in dynamic decision making. Cogn. Sci. 2003, 27, 591–635. [Google Scholar]
  3. Anderson, J.R.; Lebiere, C. The Atomic Components of Thought; Lawrence Erlbaum Associates: Hillsdale, NJ, USA, 1998. [Google Scholar]
  4. Gonzalez, C.; Lebiere, C. Instance-based cognitive models of decision making. In Transfer of Knowledge in Economic Decision-making; Zizzo, D., Courakis, A., Eds.; Macmillan (Palgrave Macmillan): New York, NY, USA, 2005; pp. 148–165. [Google Scholar]
  5. Martin, M.K.; Gonzalez, C.; Lebiere, C. Learning to make decisions in dynamic environments: Act-r plays the beer game. Proceedings of the Sixth International Conference, Cognitive Modeling, Pittsburgh, PA, USA, 2004; Lovett, M.C., Schunn, C.D., Lebiere, C., Munro, P., Eds.; Lawrence Erlbaum Associates Publishers: Pittsburgh, PA, USA, 2004; pp. 178–183. [Google Scholar]
  6. Gonzalez, C.; Dutt, V. Instance-based learning models of training. Proceedings of the Human Factors and Ergonomics Society 54th Annual Meeting, San Francisco, CA, USA, 2010; Human Factors and Ergonomics Society: San Francisco, CA, USA, 2010; pp. 2319–2323. [Google Scholar]
  7. Gonzalez, C.; Best, B.J.; Healy, A.F.; Bourne, L.E., Jr; Kole, J.A. A cognitive modeling account of simultaneous learning and fatigue effects. J. Cogn. Syst. Res. 2010, 12, 19–32. [Google Scholar]
  8. Dutt, V.; Yamaguchi, M.; Gonzalez, C.; Proctor, R.W. An instance-based learning model of stimulus-response compatibility effects in mixed location-relevant and location-irrelevant tasks. Proceedings of the 9th International Conference on Cognitive Modeling—ICCM2009, Manchester, UK, 2009; Howes, A., Peebles, D., Cooper, R., Eds.; Carnegie Mellon University: Pittsburgh, PA, USA, 2009. [Google Scholar]
  9. Lebiere, C.; Gonzalez, C.; Martin, M. Instance-based decision making model of repeated binary choice. Proceedings of the 8th International Conference on Cognitive Modeling, Ann Arbor, MI, USA, 2007; Lewis, R.L., Polk, T.A., Laird, J.E., Eds.; Taylor & Francis: London, UK, 2007; pp. 67–72. [Google Scholar]
  10. Lejarraga, T.; Dutt, V.; Gonzalez, C. Instance-based learning: A general model of repeated binary choice. J. Behav. Decis. Mak. 2010. (in press). [Google Scholar]
  11. Biele, G.; Erev, I.; Eyal, E. Learning, risk attitude and hot stoves in restless bandit problems. J. Math. Psychol. 2009, 53, 155–167. [Google Scholar]
  12. Nevo, I.; Erev, I. On surprise, change, and the effect of recent outcomes. 2010. (in preparation). [Google Scholar]
  13. Erev, I.; Barron, G. On adaptation, maximization, and reinforcement learning among cognitive strategies. Psychol. Rev. 2005, 112(4), 912–931. [Google Scholar]
  14. Rapoport, A.; Erev, I.; Abraham, E.V.; Olson, D.E. Randomization and adaptive learning in a simplified poker game. Organ.Behav.Hum.Decis.Process. 1997, 69, 31–49. [Google Scholar]
  15. Gonzalez, C.; Dutt, V. Integrating sampling and repeated choice paradigms of decision from experience. 2010. (submitted). [Google Scholar]
  16. Lebiere, C. Blending: An ACT-R mechanism for aggregate retrievals. Proceedings of the Sixth Annual ACT-R Workshop, George Mason University, Fairfax, VA, USA, 1999.
  17. Wong, T.J.; Cokely, E.T.; Schooler, L.J. An Online database of ACT-R parameters: Toward a transparent community-based approach to model development. Proceedings of the Tenth International Conference on Cognitive Modeling, Philadelphia, PA, USA; 2010; pp. 282–286. [Google Scholar]
  18. Busemeyer, J.R.; Diederich, A. Cognitive Modeling; Sage: New York, NY, USA, 2010. [Google Scholar]
  19. Busemeyer, J.R.; Wang, Y.M. Model comparison and model selections based on generalization criterion methodology. J. Math. Psychol. 2000, 44, 171–189. [Google Scholar]
  20. Erev, I.; Ert, E.; Roth, A.E.; Haruvy, E.; Herzog, S.; Hau, R.; Hertwig, R.; Stewart, T.; West, R.; Lebiere, C. A choice prediction competition for choices from experience and from description. J. Behav. Decis. Mak. 2010, 23, 15–47. [Google Scholar]
  21. Pitt, M.A.; Kim, W.; Myung, I.J. Flexibility versus generalizability in model selection. Psychon. Bull. Rev. 2003, 10, 29–44. [Google Scholar]
  22. Pitt, M.A.; Myung, I.J. When a good fit can be bad. Trends. Cogn. Sci. 2002, 6, 421–425. [Google Scholar]
  23. Navarro, D.J.; Pitt, M.A.; Myung, I.J. Assessing the distinguishability of models and the informativeness of data. Cogn. Psychol. 2004, 49, 47–84. [Google Scholar]
  24. Browne, M.W. Cross-validation methods. J. Math. Psychol. 2000, 44, 108–132. [Google Scholar]
  25. Barron, G.; Erev, I. Small feedback-based decisions and their limited correspondence to description-based decisions. J. Behav. Decis. Mak. 2003, 16, 215–233. [Google Scholar]
  26. Newell, A. Unified Theories of Cognition; Harvard University Press: Cambridge, MA, USA, 1990. [Google Scholar]

Share and Cite

MDPI and ACS Style

Gonzalez, C.; Dutt, V.; Lejarraga, T. A Loser Can Be a Winner: Comparison of Two Instance-based Learning Models in a Market Entry Competition. Games 2011, 2, 136-162. https://doi.org/10.3390/g2010136

AMA Style

Gonzalez C, Dutt V, Lejarraga T. A Loser Can Be a Winner: Comparison of Two Instance-based Learning Models in a Market Entry Competition. Games. 2011; 2(1):136-162. https://doi.org/10.3390/g2010136

Chicago/Turabian Style

Gonzalez, Cleotilde, Varun Dutt, and Tomás Lejarraga. 2011. "A Loser Can Be a Winner: Comparison of Two Instance-based Learning Models in a Market Entry Competition" Games 2, no. 1: 136-162. https://doi.org/10.3390/g2010136

APA Style

Gonzalez, C., Dutt, V., & Lejarraga, T. (2011). A Loser Can Be a Winner: Comparison of Two Instance-based Learning Models in a Market Entry Competition. Games, 2(1), 136-162. https://doi.org/10.3390/g2010136

Article Metrics

Back to TopTop