Learning, fatigue and preference formation in discrete choice experiments

https://doi.org/10.1016/j.jebo.2015.08.018Get rights and content

Highlights

  • This paper explores learning, fatigue and preference formation in discrete choice experiments.

  • The models account for inconsistent preferences and variances at different stages.

  • Most respondents in our sample show stability, consistent with rational pre-existent preferences.

  • Allowing for learning and fatigue has implications on marginal WTP estimates.

Abstract

While the repeated nature of discrete choice experiments is advantageous from a sampling efficiency perspective, patterns of choice may differ across the tasks, due, in part, to learning and fatigue. Using probabilistic decision process models, we find in a field study that learning and fatigue behavior may only be exhibited by a small subset of respondents. Most respondents in our sample show preference and variance stability consistent with rational pre-existent and well formed preferences. Nearly all of the remainder exhibit both learning and fatigue effects. An important aspect of our approach is that it enables learning and fatigue effects to be explored, even though they were not envisaged during survey design or data collection.

Introduction

Discrete choice experiments (DCEs) are a stated preference elicitation method, whereby respondents choose their preferred alternative among several hypothetical alternatives in a choice task (e.g., see Louviere et al., 2003, Hensher et al., 2005 for introductions to the method). The method is widely used for valuing environmental goods and services. In this study we explore preferences for preservation of several rare and endangered fish species in the Lough Melvin Catchment in Ireland using a DCE. As is common practice in DCEs, respondents were asked to consider a number of multidimensional alternatives and to identify their preferred alternative in a choice scenario (or task) where, in our case, different fish species were or were not protected. As in any DCE, in addition to the number of attributes and alternatives per choice task, we had the opportunity to assign the number of choice tasks. In an attempt to increase sampling efficiency we included a large number of choice tasks giving rise to a panel of repeated choice tasks to be completed by each respondent.

A key advantage of using repeated valuation tasks is that they enable researchers to identify the extent to which respondents have clearly defined and established pre-existent preferences for the goods under consideration and the extent to which preferences are modified or even formed through the course of the elicitation process. Despite this, the issues and concerns relating to learning and fatigue are not routinely explored by researchers engaged in stated discrete choice analysis. In this paper we contribute to the literature by proposing a more flexible means for dealing with learning and fatigue in stated preference and more specifically in DCEs conducted in the field. In particular, we explore the extent to which respondents possess or form consistent preferences at different phases in the experiment and whether there is different variability of choice through identification of different scale parameters for each phase.

Our modeling approach builds on the standard multinomial logit (MNL) and random parameters logit (RPL) models, but, unlike previous studies, which have deterministically assumed that the same patterns of learning, fatigue or preference heterogeneity applies to the whole sample, we accommodate the fact that the patterns may be different across respondents. To achieve this we use a probabilistic decision process (PDP) model (e.g. Campbell et al., 2012, McNair et al., 2012, Hensher et al., 2013). This is similar in form to a latent class (LC) model, but the classes here are meant to describe a specific learning and fatigue behavior. The LC model is hence a tool to facilitate differences in learning and fatigue behavior across respondents. As a further departure from the standard LC specification, similar to Greene and Hensher (2013), we facilitate within class random taste variation to capture another layer of preference heterogeneity. We first use this approach to probabilistically determine the proportion of respondents who have consistent preferences as well as preferences that change due to learning or fatigue (or a mixture of the two). We then include scale-adjusted classes, as implemented in Magidson and Vermunt (2008) and Campbell et al. (2011), to ascertain probabilistically the share of respondents with a consistent error variance as well as those who's error variance is different (relative to the middle phase) in the early and/or late phases of the experiment. While both of these PDP models represent an improvement over the existing approaches, they both look at preference and variance consistency in isolation. To overcome this potential weakness, we propose an even more elaborate scale-adjusted PDP model that is aimed at uncovering both types of inconsistency simultaneously. The beneficial feature of this is that we can better disentangle the influence of learning and fatigue upon both the preference parameters as well as the scale parameter. Moreover, it offers a practical approach for DCE practitioners to investigate learning and fatigue, even though they were not considered during survey design or data collection.

Our results show that both learning and fatigue effects are present in this dataset. Our modeling results suggest that, while only a minority of respondents exhibit learning and/or fatigue behavior, expressions of utility (in terms of both preferences and variance) are different in the early and late phases of the experiment (relative to the middle phase) for those respondents identified as exhibiting signs of these patterns. Moreover, our final scale-adjusted PDP specification highlights the potential confound between the two types of inconsistency and, thus, the necessity for specifications that can accommodate both inconsistent preferences and error variance. Results from this model suggest that around two-thirds of respondents have consistent preferences and error variance across the sequence of choice tasks. The remaining respondents are shown to either adjust their preferences or choice variability in approximately equal proportions. Our results also show that model fit as well as marginal willingness to pay (WTP) are impacted by explicitly accommodating learning and fatigue effects on preferences and variability into our models. Our modeling approach also allows us to identify empirically the patterns of responses that may be exhibited in a repeat response DCE as outlined in Day et al. (2012, Table 1, p. 75). This application of the PDP model can be applied to field datasets to test for patterns associated with “standard” and “non-standard” preference formation. We find that in a large field dataset only a minority of respondents exhibit preference and variance instability but that patterns similar to those identified in Day et al. (2012) can be found. We find that two-thirds of respondents in our study appear to have a-priori well formed preferences in terms of demonstrating both preference and variance stability throughout the valuation sequence. One-third show instability of preference and scale throughout the sequence and appear to exhibit preference discovery between the early and middle phase or fatigue between the middle and late phase of the sequence or both. Empirical evidence from our findings suggests that the dominant form of preference and scale instability among this subset of respondents was the combination of preference learning in the early phase of the sequence combined with evidence of fatigue in the late phase.

The remainder of the paper is organized as follows. In the following section, we outline some background to learning and fatigue from a stated preference perspective. In Section 3 we detail our econometric approach and introduce our PDP model with random parameters specification to segment respondents based on their patterns of learning and/or fatigue. In Section 4 we briefly discuss the empirical case-study used to provide data for our analysis. Section 5 reports estimation and post-estimation results while, Section 6 discusses the implications of these findings and concludes.

Section snippets

Background

There is a well known theoretical and empirical literature suggesting that individuals may exhibit at least two forms of heterogeneity within the sequence of their choices. One type of heterogeneity has been attributed to engaging in some form of learning or discovery process when asked to identify preferences for a sequence of economic goods (see Bradley and Daly, 1994). One of the leading proponents of this learning effect within behavioral economics is Plott (1996), who coined the term “the

Background notation

DCEs are an application of the theory of value (Lancaster, 1966), combined with Random Utility Maximization theory (Thurstone, 1927, Manski, 1977). A central assumption of DCEs is that the choice is driven by the maximization of respondents’ utility. The utility, denoted by U, that each alternative brings to the respondents can be represented by the function:Unit=βxnit+ɛnit,where n indicates the respondent, i the chosen alternative in choice occasion t, β is a vector of parameters to be

Empirical case-study

To illustrate the proposed methodology on an empirical case-study we use stated preference data collected to estimate the existence value of rare and endangered fish species in the Lough Melvin Catchment in Ireland. Lough Melvin is a freshwater lake in the North West of Ireland which straddles the border between Northern Ireland and the Republic of Ireland. With a unique population of native fish species, the Lough Melvin Catchment has an internationally important conservation status. Lough

Results

As part of our analysis, rather than use an arbitrary definition of early, middle and late choices we ran a series of models with different number of choice tasks in each of the three phases. Conducting this type of investigation also gives an insight into the most likely patterns of learning and fatigue (i.e., when learning is most likely to end and fatigue is most likely to begin) for a given dataset. Based on this investigation we observe that the most significant gains in model fit (for our

Conclusions

In the present paper, we explored the issue of learning and fatigue in the context of inconsistent preferences and variance as respondents progress through the experiment.

Our paper started with the conventional approach of dealing with learning and fatigue, whereby respondents are all treated as either exhibiting inconsistent preferences or inconsistent error variance across early, middle, and late choice tasks. We then moved away from this deterministic method for uncovering learning and

References (52)

  • W.H. Greene et al.

    A latent class model for discrete choice analysis: contrasts with mixed logit

    Transp. Res. B

    (2003)
  • S. Hess

    Conditional parameter estimates from mixed logit models: distributional assumptions and a free software tool

    J. Choice Model.

    (2010)
  • S. Hess et al.

    Estimation of value of travel-time savings using mixed logit models

    Transp. Res. A

    (2005)
  • S. Hess et al.

    Not bored yet-revisiting respondent fatigue in stated choice experiments

    Transp. Res. A: Policy Pract.

    (2012)
  • J. Swait et al.

    Choice environment, market complexity, and consumer behavior: a theoretical and empirical approach for incorporating decision complexity into models of consumer choice

    Organ. Behav. Hum. Decis. Process.

    (2001)
  • A. Alberini

    Repeated questioning in choice experiments: are we improving statistical efficiency or getting respondents confused?

    J. Environ. Econ. Policy

    (2012)
  • D. Ariely et al.

    “Coherent arbitrariness”: stable demand curves without stable preferences

    Q. J. Econ.

    (2003)
  • M. Bradley et al.

    Use of the logit scaling approach to test for rank-order and fatigue effects in stated preference data

    Transportation

    (1994)
  • J. Braga et al.

    Preference anomalies, preference elicitation and the discovered preference hypothesis

    Environ. Resour. Econ.

    (2005)
  • R. Brouwer

    Constructed preference stability: a test-retest

    J. Environ. Econ. Policy

    (2012)
  • R. Brouwer et al.

    Choice certainty and consistency in repeated choice experiments

    Environ. Resour. Econ.

    (2010)
  • D. Campbell

    Willingness to pay for rural landscape improvements: combining mixed logit and random-effects models

    J. Agric. Econ.

    (2007)
  • D. Campbell et al.

    Position bias in best-worst scaling surveys: a case study on trust in institutions

    Am. J. Agric. Econ.

    (2015)
  • D. Campbell et al.

    Non-attendance to attributes in environmental choice analysis: a latent class specification

    J. Environ. Plan. Manage.

    (2011)
  • D. Campbell et al.

    Accommodating coefficient outliers in discrete choice modelling: a comparison of discrete and continuous mixing approaches

  • M. Czajkowski et al.

    Learning and fatigue effects revisited: investigating the effects of accounting for unobservable preference and scale heterogeneity

    Land Econ.

    (2014)
  • Cited by (42)

    • Estimating willingness-to-pay for a livestock vaccine among the marginalized: The role of reflective thought in discrete choice experiments

      2022, Preventive Veterinary Medicine
      Citation Excerpt :

      The proportions, presented in Table 5, were: 0.52 vs 0.46 for one attribute, and 0.07 vs 0.05 for three attributes. This result is consistent with learning that occurs during DCEs (Czajkowski et al., 2014; Campbell et al., 2015). At least 20% of respondents’ choices ignored no attributes.

    • Assessing the value of cooperative membership: A case of dairy marketing in the United States

      2021, Journal of Co-operative Organization and Management
      Citation Excerpt :

      The data collected in a DCE are limited in quality based on the ability for respondents to place themselves in a setting where they are behaving in a manner consistent to what would occur in a true willingness-to-accept scenario; i.e., hypothetical bias is present. Experimental designs that result in surveys taking over 15 min jeopardize the establishment of this setting and lead to increased rates of fatigue (Campbell, Boeri, Doherty, & Hutchinson, 2015). Based on our number of offers per question (2), six questions per respondent are recommended; i.e., each respondent sees six sets of two offers.

    View all citing articles on Scopus

    The authors wish to acknowledge the funding provided by the EU INTERREG IIIA Programme for Ireland/Northern Ireland for data collection. The usual disclaimer applies.

    View full text