Choice modelling

Choice modeling attempts to model the decision process of an individual or segment in a particular context. Choice modeling may be used to estimate non-market environmental benefits and costs.[1]

Many alternative models exist in econometrics, marketing, sociometrics and other fields, including utility maximization, optimization applied to consumer theory, and a plethora of other identification strategies which may be more or less accurate depending on the data, sample, hypothesis and the particular decision being modelled. In addition Choice modeling is regarded as the most suitable method for estimating consumers’ willingness to pay for quality improvements in multiple dimensions.[2] The Nobel Prize for economics was awarded to a principal proponent of the Choice modeling theory, Daniel McFadden.[3]

Related terms for choice modeling

There are a number of terms which either are subsets or overlap with other areas of econometrics that may broadly be termed Choice modeling.

These include the following:

  1. Stated preference discrete choice modeling
  2. Discrete choice
  3. Choice experiment
  4. Choice set
  5. Conjoint analysis
  6. Controlled experiments

Theoretical background

Choice modeling was developed in parallel by economists and cognitive psychologists. The origins of choice modeling can be traced to Thurstone's research into food preferences in the 1920s and to random utility theory.

Choice modeling posits that with human choice there is an underlying rational decision process and that this process has a functional form. Depending on the behavioural context, a specific functional form may be selected as a candidate to model that behaviour. The multinomial logit or MNL model form is commonly used as it is a good approximation to the economic principle of utility maximisation. That is, human beings strive to maximise their total utility. The multinomial logit form describes total utility as a linear addition (or subtraction) of the component utilities in a context. Once the functional form of the decision process has been established, the parameters of a specific model may be estimated from available data using multiple regression, in the case of MNL. Other functional forms may be used or combined, such as binary logit, probit or EBA with appropriate statistical tests to determine the goodness of fit of the model to a hold out data set.

Methods used in choice modeling

Choice modeling comprises a number of specific techniques that contribute to its power. Some or all of these may be used in the construction of a Choice Model.

Orthogonality

For model convergence, and therefore parameter estimation, it is often necessary that the data have little or no collinearity. The reasons for this have more to do with information theory than anything else. To understand why this is, take the following example:

Imagine a car dealership that sells both luxury cars and used low-end vehicles. Using the utility maximisation principle and an MNL model form, we hypothesise that the decision to buy a car from this dealership is the sum of the individual contribution of each of the following to the total utility.

Using multinomial regression on the sales data however will not tell us what we want to know. The reason is that much of the data is collinear since cars at this dealership are either:

There is not enough information, nor will there ever be enough, to tell us whether people are buying cars because they are European, because they are a BMW or because they are high performance. The reason is that these three attributes always co-occur and in this case are perfectly correlated . That is: all BMW's are made in Germany and are of high performance. These three attributes: origin, marque and performance are said to be collinear or non-orthogonal.

These types of data, the sales figures, are known as revealed preference data, or RP data, because the data 'reveals' the underlying preference for cars. We can infer someone's preference through their actions, i.e. the car they actually bought. All data mining uses RP data. RP data is vulnerable to collinearity since the data is effectively from the wild world of reality. The presence of collinearity implies that there is missing information, as one or more of the collinear factors is redundant and adds no new information. This weakness of data mining is that the critical missing data that may explain choices, is simply never observed.

We can ensure that attributes of interest are orthogonal by filtering the RP data to remove correlations. This may not always be possible, however using stated preference methods, orthogonality can be ensured through appropriate construction of an experimental design. Decision making can be affected by whether choices are presented together or separately through the distinction bias.

Experimental design

In order to maximize the information collected in Stated Preference Experiments, an experimental design (below) is employed. An experimental design in a Choice Experiment is a strict scheme for controlling and presenting hypothetical scenarios, or choice sets to respondents. For the same experiment, different designs could be used, each with different properties. The best design depends on the objectives of the exercise.

It is the experimental design that drives the experiment and the ultimate capabilities of the model. Many very efficient designs exist in the public domain that allow near optimal experiments to be performed.

For example the Latin square 1617 design allows the estimation of all main effects of a product that could have up to 1617 (approximately 295 followed by eighteen zeros) configurations. Furthermore this could be achieved within a sample frame of only around 256 respondents.

Below is an example of a much smaller design. This is 34 main effects design.

0 0 0 0
0 1 1 2
0 2 2 1
1 0 1 1
1 1 2 0
1 2 0 2
2 0 2 2
2 1 0 1
2 2 1 0

This design would allow the estimation of main effects utilities from 81 (34) possible product configurations. A sample of around 20 respondents could model the main effects of all 81 possible product configurations with statistically significant results.

Some examples of other experimental designs commonly used:

More information on experimental designs may be found here.

Stated preference

A major advance in choice modeling has been the use of Stated Preference data. With RP data we are at the whim of the interrelated nature of the real world. With SP data, since we are directly asking humans about their preferences for products and services, we are also at liberty to construct the very products as we wish them to evaluate.

This allows great freedom in the creative construction of many improbable but plausible hypothetical products. It also allows complete militation against collinearity through experimental design.

If instead of using the RP sales data as in the previous example, we were to show respondents various cars and ask "Would you buy this car?"", we could model the same data. However, instead of simply using the cars we actually sold, we allowed ourselves the freedom to create hypothetical cars, we could escape the problems of collinearity and discover the true utilities for the attributes of marque, origin and performance. This is known as a Choice Experiment.

For example one could create the following unlikely, however plausible scenarios.

Information theory tells us that a data set generated from this exercise would at least allow the discrimination between 'origin' as a factor in choice.

A more formal derivation of an appropriate experimental design would consequently ensure that no attributes were collinear and would therefore guarantee that there was enough information in the collected data for all attribute effects to be identified.

Because individuals do not have to back up their choices with real commitments when they answer the survey, to some extent, they would behave inconsistently when the situation really happens, a common problem with all SP methods.

However, because Choice Models are Scale Invariant this effect is equivalent for all estimates and no individual estimate is biased with respect to another.

SP models may therefore be accurately scaled with the introduction of Scale Parameters from real world observations, yielding fairly accurate predictive models.

Preferences as choice trade-offs

It has long been known that simply asking human beings to rate or choose their preferred item from a scalar list will generally yield no more information than the fact that human beings want all the benefits and none of the costs. The above exercise if executed as a quantitative survey would tell us that people would prefer high performance cars at no cost. Again information theory tells us that there is no context-specific information here.

Instead, a choice experiment requires that individuals be forced to make a trade-off between two or more options, sometimes also allowing 'None or Neither' as a valid response. This presentation of alternatives requires that the at least some respondents compare: the cheaper, lower performance car against the more expensive, higher performance car. This datum provides the key missing information necessary to separate and independently measure the utility of performance and price.

Sampling and block allocation

Stated Preference data must be collected in highly specific fashion to avoid temporal, learning and segment biases. Techniques include:

Model generation

The typical outputs from a choice model are:

Choice modeling in practice

Superficially, a Choice Experiment resembles a market research survey; Respondents are recruited to fill out a survey, data is collected and the data is analysed. However two critical steps differentiate a Choice Experiment from a Questionnaire:

  1. An experimental design must be constructed. This is a non-trivial task.
  2. Data must be analysed with a model form, MNL, Mixed Logit, EBA, Probit etc...

The Choice Experiment itself may be performed via hard copy with pen and paper, however increasingly the on-line medium is being used as it has many advantages over the manual process, including cost, speed, accuracy and ability to perform more complex studies such as those involving multimedia or dynamic feedback.

Despite the power and general applicability of Choice modeling, the practical execution is far more complex than running a general survey. The model itself is a delicate tool and potential sources of bias that are ignored in general market research surveys need to be controlled for in choice models.

Strengths of choice modeling [4]

Choice modeling versus traditional quantitative market research

Choice Experiments may be used in nearly every case where a hard estimate of current and future human preferences needs to be determined.

Many other market research techniques attempt to use ratings and ranking scales to elicit preference information.

Ratings

Major problems with ratings questions that do not occur with Choice Models are:

Ranking

Rankings do introduce an element of trade-off in the response as no two items may occupy the same ranking position. Order preference is captured; however, relative importance is not.

Choice Models however do not suffer from these problems and furthermore are able to provide direct numerical predictions about the probability an individual will make a particular choice.

Maximum difference scaling

Maximum Difference Preference Scaling (or MaxDiff as it is commonly known) is a well-regarded alternative to ratings and ranking. It asks people to choose their most and least preferred options from a range of alternatives. By integrating across the choice probabilities, utility scores for each alternative can be estimated on an interval scale.

Uses of choice modeling

Choice modeling is particularly useful for:

Choice modeling is a standard technique in travel demand modeling. Classical references include Ben Akiva and Lerman (1985)[5] and Cascetta (2009),[6] while more recent methodological developments are described in Train (2003).[7]

Early applications of discrete choice theory to marketing are described in Anderson et al. (1992).[8]

Recent developments include a Bayesian approach to discrete choice modeling as set out in Rossi, Allenby, and McCulloch (2009).[9]

See also

References

  1. Jeff Bennet University of Queensland https://www.epa.qld.gov.au/publications?id=1585
  2. 2001 - Centre for International Economics - Review of willingness-to-pay methodologies
  3. Economics 2000
  4. 2001 The Centre for International Economics review of willingness-to-pay methodologies
  5. [Ben Akiva, M., Lerman,S. (1985) Discrete Choice Analysis: Theory and Application to Travel Demand. MIT Press]
  6. [Cascetta, E (2009) Transportation systems analysis. Springer]
  7. [Train, K. (2003) Discrete Choice Methods with Simulation. Cambridge University Press]
  8. [Anderson, P, De Palma, A., Thisse, J. (1992) Discrete choice theory of product differentiation. MIT Press]
  9. [Rossi, P., Allenby, G., McCulloch, R. (2009) Bayesian statistic and marketing. Wiley]

External links