-
Notifications
You must be signed in to change notification settings - Fork 0
/
draft.tex
2223 lines (1570 loc) · 298 KB
/
draft.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
\documentclass[a4paper,12pt]{article}
\usepackage{graphicx}
\usepackage{setspace}
\usepackage[lmargin=2.5cm,tmargin=3cm,rmargin=2.5cm,vscale=0.8,nohead]{geometry}
\usepackage{multicol}
\usepackage{wrapfig}
\usepackage{tikz}
\usepackage{tikz-network}
\usepackage{subfig}
\usepackage{amsmath, amsthm, amsfonts, amssymb}
\usepackage{bibentry}
\usepackage[round]{natbib}
%\usepackage[numbers]{natbib}
\usepackage{xpatch}
\usepackage{pgfplots}
\newcommand{\citeyearonly}[1]{\citeyearpar{#1}}
%\usepackage{fancyhdr}
\usepackage{mathrsfs}
\usepackage{authblk}
\usepackage[alwaysadjust]{paralist}
\usepackage{alltt}
\usepackage{caption}
\usepackage{array}
\usepackage[fit]{truncate}
%\pagestyle{fancy}
\usepackage{calc}
%\usepackage{fancyvrb}
\usepackage{float}
\usepackage[utf8]{inputenc}
\usepackage{lscape}
\usepackage{longtable}
\usepackage{tabularx}
\usepackage[figuresright]{rotating}
\usepackage{color}
\usepackage{hyperref}
\usepackage{enumerate}
\usepackage{url}
\usepackage[normalem]{ulem}
\usepackage{multirow}
\usepackage{rotating}
\usepackage{amsmath}
\usepackage{graphicx}
\usepackage{caption}
\usetikzlibrary{positioning, arrows}
\usepackage{multirow}
\usepackage{tabularx}
\usepackage{array}
\usepackage{booktabs}
\usepackage{algorithm}
\usepackage{algpseudocode}
\usepackage{array}
\usepackage{graphicx}
\usepackage{longtable}
\usepackage{adjustbox}
\usepackage[stable]{footmisc}
\usepackage[french, english]{babel}
\usepackage[T1]{fontenc}
\usepackage{lmodern}
\usepackage{epigraph}
\usepackage[page,toc,titletoc,title]{appendix}
\pgfplotsset{compat=1.18}
\newcommand{\doubleitem}{%
\begingroup
\stepcounter{enumi}%
\edef\tmp{\theenumi, }%
\stepcounter{enumi}
\edef\tmp{\endgroup\noexpand\item[\tmp\labelenumi]}%
\tmp}
\newenvironment{dedication}
{%\clearpage % we want a new page %% I commented this
\thispagestyle{empty}% no header and footer
\vspace*{\stretch{1}}% some space at the top
\itshape % the text is in italics
\raggedleft % flush to the right margin
}
{\par % end the paragraph
\vspace{\stretch{3}} % space at bottom is three times that at the top
\clearpage % finish off the page
}
\renewcommand{\footnotesize}{\normalsize}
\title{{\bf \Large The Power of Context in Decision-Making and Recommendations}
}
\author{Emil Mirzayev}
%March 28, 2023
\begin{document}
\begin{dedication}
To my father, my first mentor, who seeded my enduring passion for learning.
\end{dedication}
\spacing{1.5}
\maketitle
\clearpage
\begin{otherlanguage}{french}
\begin{abstract}
Cette dissertation comprend quatre études qui examinent les effets du contexte et appliquent ces connaissances pour améliorer les systèmes de recommandation dans les marchés en ligne. Je différencie deux types de contexte : interne et externe. Les conclusions de la dissertation indiquent que les effets du contexte interne dans les choix, qui ont été principalement étudiés et observés auparavant dans des situations expérimentales, sont également présents et détectables dans des situations concrètes. De plus, l'étude examinant le contexte externe trouve une relation positive entre les outils qui permettent aux utilisateurs de signaler leurs préférences aux systèmes de recommandation et leur adoption. Les applications empiriques des chapitres composant cette thèse reposent sur quatre ensembles de données distincts, le plus grand étant des données observationnelles provenant d'une situation concrète, et les trois autres provenant d'une situation expérimentale.
La première étude applique un modèle de décision computationnel à un ensemble de données conséquent de choix dans le monde réel, en faisant la première application de cette envergure. Les résultats indiquent que le contexte influence les choix des individus. L'étude suggère que les marchés en ligne pourraient utiliser de tels modèles pour approfondir la compréhension de la composition de l'ensemble de choix et de l'interaction entre différentes options sur les décisions des consommateurs.
La deuxième étude va au-delà des études de contexte traditionnelles en développant une méthodologie visant à séparer ses trois composantes principales, connues sous les noms d'attraction, de compromis et de similarité, les unes des autres. Cette étude contribue à comprendre l'interaction entre différents effets de contexte au sein d'un ensemble de choix et préconise le développement de conceptions de systèmes de recommandation et une compréhension plus profonde de la nature hétérogène de la dynamique du choix du consommateur. Les résultats de cette étude peuvent être utilisés pour naviguer dans le problème dit de démarrage à froid rencontré par les marchés numériques.
La troisième étude présente une approche novatrice pour aborder le problème de démarrage à froid du côté de l'utilisateur dans la conception de systèmes de recommandation. Elle s'appuie sur les résultats de l'étude précédente et applique les conclusions de la littérature sur le choix à deux étapes observé chez les individus pour générer des ensembles de considération. Ses conclusions ouvrent la voie à l'examen des effets de contexte provenant de l'extérieur des ensembles de choix, à savoir les préférences des individus et les outils qui leur permettent de signaler leurs préférences aux systèmes de recommandation. Cela est présenté comme l'un des principaux mécanismes pour créer des systèmes de recommandation plus efficaces.
La dernière étude a examiné l'effet du contrôle de l'utilisateur sur l'acceptation du système de recommandation en utilisant le Modèle d'Acceptation de la Technologie comme cadre théorique. Cette étude a trouvé que les systèmes de recommandation faciles à utiliser étaient perçus par les utilisateurs comme plus utiles et entraînaient une plus grande intention de les utiliser. Cependant, différents mécanismes de contrôle ont eu des impacts variés sur l'expérience utilisateur.
Cette thèse démontre l'existence d'effets de contexte dans des configurations multiattributs, multidimensionnelles et développe des méthodologies pour améliorer la conception des systèmes de recommandation avec ces effets de contexte. De plus, cette étude examine le contexte externe, à savoir les outils qui permettent aux utilisateurs d'exprimer leurs préférences et comment rendre les systèmes de recommandation meilleurs grâce à eux.
\end{abstract}
\end{otherlanguage}
\clearpage
\begin{abstract}
This dissertation comprises four studies investigating context effects and applying the knowledge to enhance recommender systems in online marketplaces. I differentiate between two types of context: internal and external. The findings of the dissertation indicate that the effects of the internal context in choice settings, which were primarily studied and observed before in experimental settings, are also present and detectable in field settings. Furthermore, the study investigating the external context finds a positive relationship between tools that enable users to signal their preferences to recommender systems and their adoption. The empirical applications of the chapters comprising this thesis rely on four distinct datasets, the largest being observational data from a field setting and the remaining three coming from an experimental setting.
The first study applies a computational decision-making model to a substantial dataset of real-world choice, making it the first application of this magnitude. The findings indicate that the context influences the choices of individuals. The study suggests that online marketplaces could use such models to gain further insight into how the composition of the choice-set and the interaction among different options affect consumer decisions.
The second study extends beyond traditional context studies by developing a methodology that aims to disentangle its three main components, known as attraction, compromise, and similarity, from each other. This study contributes to understanding the interaction between different context effects within a choice set and advocates the development of recommender system designs and a deeper understanding of the heterogeneous nature of the consumer choice dynamics. The results of this study can be used to navigate the so-called cold-start problem faced by digital marketplaces.
The third study presents a novel approach to addressing the user-side cold start problem in recommender system design. It builds on the results of the previous study and applies the findings of the two-stage choice literature observed in individuals to generate consideration sets. Its findings pave the way for investigating context effects arising from outside the choice sets, namely, the preferences of individuals and the tools that allow them to signal their preferences to recommender systems. This is argued to be one of the main mechanisms to create more effective recommender systems.
The final study investigated the effect of user control on recommender system acceptance using the Technology Acceptance Model as a theoretical framework. This study found that easy-to-use recommender systems were perceived by users as more useful and resulted in greater intention to use them. However, different control mechanisms had varying impacts on user experience.
This thesis demonstrates the existence of context effects in multiattribute, multidimensional settings and develops methodologies of enhancing recommender systems design with context effects. Additionally, this study investigates the external context, namely, the tools that allow users to express their preferences and how to make recommender systems better through them.
\end{abstract}
\clearpage
\section*{Acknowledgements}
Embarking on this scholarly journey has been both humbling and transformative, marked by countless days of meticulous study, intense deliberation, and profound discoveries. Indeed, the journey was punctuated with hurdles and moments of self-doubt. However, every trough was succeeded by a crest, each stumble by an ascent, and every challenge served as an opportunity. The path was steep, but these uphill struggles honed my perspective, expanded my horizons, and ignited my passion for knowledge. The lessons learned, the triumphs celebrated and the sense of accomplishment far outweighed the trials faced. Reflecting on these uplifting moments, I joyfully acknowledge those who accompanied me on this journey and made it not only possible, but also immensely rewarding.
Firstly, I extend my deepest gratitude to Zakaria Babutsidze, my Ph.D. supervisor. He has been a great mentor and a source of inspiration for me. His honesty and straightforwardness challenged me to think critically and creatively, shaping me into someone better day by day. I am indebted to Francesco Castellanetta, our Ph.D. program director, for his guiding hand and unwavering faith in my capabilities. I appreciate our long and intriguing discussions with Ludovic Diabaggio that transcended academia. I am grateful to Bill Rand for his invaluable insight, feedback, and co-authorship. Daniela Iubatti, your informal talks and unwavering support have been a source of inspiration. I appreciate Diego Zunino for his honest and invaluable guidance. Bruno Cirillo deserves special mention; our albeit slightly formal discussions have significantly shaped my work. My heartfelt appreciation goes to Renata Kaminska. Her support and intuitive understanding of my thoughts were essential in my pursuits. I am grateful to Benjamin Montmartin; our separate exchanges about wine (we were in France at the end) and statistics enriched my perspectives immeasurably. To Valerio Incerti, our creative discourses, even about less creative ideas, were a constant source of inspiration. Lapo Mola's valuable feedback motivated me to pursue excellence, and I value that wholeheartedly.
Aytan, my beloved wife and best friend, your unwavering support and guidance have been my Polaris, guiding me through the darkest nights and teaching me to only settle for infinity and beyond. Ecem Delicik, my dear friend, I am grateful for the shared stories, moments, laughter, and silence that have strengthened our friendship. The moment you realized that I was speaking to you in Turkish is unforgettable (I am sure you still remember it, too). Johanna Deperi, our shared moments and long conversations were invaluable for self-reflection, joy, and sometimes ``justified sadness'' throughout this journey. Artyom Yepremyan, our discussions helped us conquer the PhD mountain. I will forever cherish the times when I needed a hand, and you extended yours. Manon Desjardins, your constant support despite the towering stack of papers on your desk will always be appreciated. Mehdi Ibrahim, I am grateful for our coffee talks and the irresistible desserts from Morocco. Teymur Mardaliyev, your friendship and coding prowess have been invaluable; only you could convince me to settle for JavaScript there.
Finally, no words can fully express my gratitude to my parents. Father, your inquisitive nature has been my guide, instilling in me the curiosity that led me on this scholarly path. Your enthusiasm for the unexplored has been infectious, leading me to question, explore, and discover. Mother, you have been my anchor, reminding me of my roots and my identity in the academic whirlwind. Your unwavering encouragement and steadfast belief have reinforced my sense of self, allowing me to face challenges with courage and resilience. You taught me to stay grounded and true to myself, no matter how far I travel in the pursuit of knowledge. Your lessons continue to shape my journey, helping me navigate both personal and academic hurdles with grace and fortitude.
\textit{Çox sağ olun ki, varsınız!}
\clearpage
\tableofcontents
\newpage
\listoffigures
\clearpage
\listoftables
\newpage
\newpage
\section{Introduction}
\epigraph{We all make choices, but in the end, our choices make us.}{}
The main objective of this thesis is to make a contribution towards understanding the choice context and applying it to improve recommender systems. The thesis makes an effort to identify context effects arising from the choice set and to develop a methodology to implement this information in the design of recommender systems. Furthermore, it analyzes the effectiveness of user control mechanisms, which allow individuals to inform the system about their preferences and also amend them.
Recommender systems have become a crucial ally in the vast landscape of the digital world. They help individuals to navigate through the plethora of alternatives and find what they are looking for. They achieve this by using sophisticated algorithms that analyze the wide space of items, users, and their interactions with each other. System designers and businesses thrive to maximize the accuracy of recommendations, which means they want to maximize the consumption of recommendations. However, recent research streams show that accuracy is not the only way to make recommender systems effective. Other factors such as diversity and serendipity are also important \citep{kaminskas2016diversity}. Moreover, the black-box nature of the algorithms does not allow us to understand the reasons behind a particular user's choice behavior \citep{kotkovSurveySerendipityRecommender2016, samih2021exmrec2vec}. Therefore, more and more interest is directed towards understanding and utilizing the context around the particular choice \citep{adomavicius2005toward}.
The choice context has long received the attention of scholars in many fields, including marketing, psychology, and economics. Because it was studied in many domains, scholars refer to it differently. This thesis concentrates on and uses two of them. The first one posits that the choice context \footnote{For the remainder of the dissertation, I will follow the previous literature \citep{truebloodEtAl13} and refer to this definition of the choice context as context effects.} is ``the availability and nature of the choice alternatives'' \citep{tversky1972elimination, huberEtAl82, simonson89}. Previous research has demonstrated the existence of context effects in various settings \citep{herne1997decoy, soltani2012range, evangelidisEtAl18, wuConsguner20}. The second definition of context arises much later with the proliferation of recommender systems. Then, context is also referred to as ``the time and content of the choice, the location or sociodemographic characteristics of the decision maker...'' \citep{adomavicius2011context}\footnote{To distinguish between these two definitions I will refer to this definition as external context throughout this dissertation.}
However, modeling for context effects mathematically was a challenge because the existing models suffered from the independence of irrelevant alternative criterion \citep{luce59}, which meant that these models treated each alternative in isolation. This has shifted the scholar's interest towards computational models \citep{usher2001time, roe2001multialternative, trueblood2014multiattribute, noguchi2018multialternative}. Yet, none of these models has been applied to real-world choice data, and as a result, their applicability to field data remained an open question. I address this gap in the first study where I apply a computational model to a field data of high heterogeneity among dimensions. Building on my results in my second study, I have delved deeper into choice modeling and developed a methodology to account for the three main context effects studied in the literature \citep{truebloodEtAl13}.
When new users or items are introduced to recommender systems, they fail to function as intended because of the lack of information their algorithms need. In such cases, even when they do provide recommendations, the recommendations are far from personalized \citep{lika2014facing}. It is considered a key challenge in recommender system design \citep{park2009pairwise}. The previous literature has applied various methods to overcome this problem, including asking users to rate some items, share their preferences among others \citep{guy2009personalized, aharon2013off, bykau2013coping, saveski2014item}. However, the limitations of these approaches are that they ignore the context effects and concentrate on scenarios where information scarcity is temporary. Using the findings from my second study and combining it with decision-making and choice modeling literature, I propose an innovative solution to address continuous information scarcity issue in recommender system design by utilizing contextual information of the menu to generate relevant choice sets through a two-step choice modeling method.
Research agrees on the importance of metrics outside the boundaries of accuracy for recommender systems \citep{kaminskas2016diversity}. It is in the best interest of online marketplaces to provide users with recommendations that are not only in the category ``exactly what I want'', but also ``I never thought I would have liked this'' \citep{kotkovSurveySerendipityRecommender2016}. This can not only boost sales \citep{songWhenHowDiversify2019}, but also increase user satisfaction \citep{knijnenburgExplainingUserExperience2012}. However, it was observed that individual preferences are not stable and tend to change and system designers have proposed various tools to allow users to signal their preference shifts \citep{bostandjiev2012tasteweights,hijikata2012relation}. However, to better understand, it is necessary to study those tools in isolation from the context arising from the choice set and concentrate purely on external context that arises from the user side, e.g., their preferences \citep{adomavicius2011context}. In my last study, I address this gap by conducting an online experiment and applying Technology Acceptance Model \citep{davis1985technology} to measure users' acceptance of such tools.
All in all, the four studies aim to contribute to our understanding of the context and provide applications of the design of recommender systems using this information.
\newpage
\section{Exploring Context Effects in Multi-Attribute, Multi-Alternative Choice Environments}\label{chapter:simulationStudy}
\begin{abstract}
Previous computational decision making models that were developed to account for context effects have only been studied with an experimental data where only one effect was produced at a time. Using data coming from strictly controlled experimental environments hinders the understanding of context effects that occur in real-world choice scenarios where items have multiple dimensions and choice sets have dozens of alternatives. In this chapter, I apply a computational model that accounts for context effects to an observational data which was not done before. The data comes from an air travel industry and is ideal to study context effects in multiattribute, multialternative choice environments. I first find optimal parameters for the computational model using the differential evolution algorithm. Then, I complement a traditional choice model with its outputs and assess the significance of its contribution. This chapter contributes to context effect and decision-making literature by providing further insights on behavior of computational decision-making models in real-world choice data.
\end{abstract}
\newpage
\subsection{Introduction}
Context effects have been extensively studied and demonstrated in various domains, from psychology to marketing \citep{herne1997decoy, soltani2012range, truebloodEtAl13, frederickEtAl14, evangelidisEtAl18, wuConsguner20}. Some recent studies also concluded that multiple context effects may occur at the same time \citep{berkowitsch2014rigorously, noguchi2014attraction}. However, recent studies have also discovered boundary conditions for these effects \citep{liew2016appropriacy, spektor2018good, spektor2019similarity}. Familiarity with the choice domain was found to reduce the context effects experienced by individuals \citep{kim2005attraction, sheng2005understanding}. It was also found that some conditions may force these effects to completely reverse \citep{cataldo2019comparison}. The findings described above make it necessary to call for a model which could explain these effects.
Logit and Probit models have been traditionally used in choice settings \citep{gensch1979multinomial, kim2017probit}. However, those models cannot account for context effects because they only account for the attributes of the focal option, not taking into account the attributes of the other options in the choice set. Tversky has proposed a model of elimination by aspects that could explain the similarity effect \citeyearonly{tversky1972elimination}. The foundation of the model is attention switching of individuals between alternatives and attributes and their comparisons. Once attention is received, the attribute value of a given alternative is compared to a predetermined threshold value by the individual, and if the threshold does not meet, that alternative is eliminated from the decision. The step is then continued with another attribute until the final decision is made. Another model proposed by Tversky and Simonson could explain the effect of compromise \citeyearonly{tverskySimonson93}. The theoretical foundations of this model posited that alternatives are compared based on a weighted sum of attribute values and a local context comprising binary comparisons among alternatives. However, these two models were unable to successfully account for all three effects. Despite their drawbacks, the sequential decision making and attention-based mechanisms in these models laid the foundations of many upcoming computational choice models \citep{bhatia2013associations}.
In the last three decades, researchers have developed many computational models which account for context effects. Multialternative Decision Field Theory (MDFT) \citep{roe2001multialternative}, Leaky competing
accumulator \citep{usher2001time}, Multiattribute linear ballistic accumulator \citep{trueblood2014multiattribute}, Multialternative Decision by Sampling \citep{noguchi2018multialternative} are part of them. Some of these models have been extensively tested and studied, while others are relatively new and therefore have not received much attention from scholars \citep{truebloodEtAl13}. However, these studies have been performed with experimental data \citep{trueblood2014multiattribute, berkowitsch2014rigorously, evans2019response, busemeyer2019cognitive}. Research has proven that the behavior of individuals in the laboratory choice environment is different from real-world choice environment \citep{hogarth1989risk}. Hence, the applicability of such models to field data is an uninvestigated avenue because no previous study has been done where a computational model was applied to real-world observational data. The plan is to address this gap by applying the Multialternative Decision by Sampling (MDbS) model proposed by Noguchi and Stewart \citeyearonly{noguchi2018multialternative} to a field data from the airfare booking domain. This would allow me to assess the applicability level of this model complex field-data.
Also, applying MDbS to observational data would allow one to statistically assess the significance of the contribution of this model's ability to account for the context effect. Instead of testing this model against established choice models, I will attempt to complement them with MDbS. To do this, the random effect Probit model will be used as a variation of the Probit family and will be augmented with MDbS output. The Probit model is chosen as it does not explicitly assume IIA unlike the family of Logit models. To validate my results further, I will apply the same methodology and analysis to an experimental data.
The reasons for choosing MDbS are twofold. First, it is relatively new when compared to other models, hence it has not been further investigated before. Second, it is more robust and can account for a wider range of context effects (than other models) known to the literature \citep{noguchi2018multialternative}.
\subsection{Big three context effects}\label{chapter:bigThreeContextEffectsDescription}
We make choices all the time. Imagine the time you went to see a movie you had been waiting for a while and decided to grab some popcorn before entering. You may have seen something like that in the figure \ref{fig:decoyPopcornExample}, although prices may be higher these days. You are puzzled at first, but reminded that the movie is about to start, so you better hurry up and choose one.
The small one feels not enough for a 90 minute marathon. Then, there is a middle one which seems like an okay option at first. When you notice that big box, you immediately forget about the small one you saw moments ago. You start looking at sizes and prices of middle and large boxes and think: Well, that is easy. Big box seems like the way to go here, considering their prices are almost equal.
\begin{figure}[H]
\centering
\includegraphics[width=0.5\textwidth]{staticFiles/popcornDecoy.png}
\caption{Classic illustration of context effects.}
\label{fig:decoyPopcornExample}
\end{figure}
If this situation is familiar to you, then you have experienced the so-called context effect which can be understood as ``the composition and the nature of the choice set, availability of various options in it'' \citep{tversky1972elimination, huberPuto83}. For a long time, our understanding of the choice did not extend over the borders of two related principles. The first one is independence of irrelevant alternatives (IIA), which states that when having a choice between two two options $A$ and $B$ if a person prefers $A$ for example, regardless of adding a third option $C$ to this choice set, that person's preference must be unaltered \citep{luce59}. The second is the regularity principle, which states that the probability of choice of option $A$ cannot increase by the introduction of option $C$ \citep{luce59}.
However, research has concluded that the context of the choice set and the options in it have a substantial effect on how people make decisions. This effect has been extensively studied over the past five decades by many economists, marketing scholars, and psychologists \footnote{See Dowling et al. \citeyearonly{dowlingEtAl20} and Lichtenstein \citeyearonly{lichtenstein2006construction} for a more comprehensive review.} \citep{ kahnemanTversky79, simonson89, tverskySimonson93, lichtenstein2006construction, dowlingEtAl20}. Most of the research has focused on three context effects, also referred to as ``big three'': attraction, compromise and similarity \citep{howes2016contextual}.
To better understand these three context effects, let us think of a hypothetical choice set where options differ along two dimensions: Dimensions 1 and 2. We first start with a set of options consisting of two options: $A$ that has the values 20 and 80; $B$ that has the values 80 and 20 that figure \ref{fig:binaryChoiseSet} represents.
\begin{figure}[h]
\centering
\includegraphics[width=0.7\textwidth]{staticFiles/noEffect.png}
\caption{Binary choice set with two options.} % Add your description here
\label{fig:binaryChoiseSet} % This labels your figure for reference
\end{figure}
People who prefer Dimension 1 would choose $B$ whereas $A$ will be chosen by people who prefer Dimension 2. This can be described as equation \ref{eq:onlyTwoOptions} below.
\begin{equation}\label{eq:onlyTwoOptions}
P(A|A,B) = P(B|A,B)
\end{equation}
Where $P(A|A,B)$ corresponds to the probability of choosing $A$ given the choice set $A,B$. Same goes for $P(B|A,B)$.
\textit{Attraction effect}
Now, let us add a third option to this choice set, the option $D_A$ to create one variation, and $D_B$ to create the second variation of the ternary choice set. Both added options have lower values in both dimensions compared to $A$ and $B$, respectively. Huber created this type of scenario and has found what he has called the ``attraction effect'' \citeyearonly{huberEtAl82}. The attraction effect, which is also known in the literature as the asymmetric dominance effect, is a consistent violation of the regularity principle mentioned earlier. He suggested that when having a choice set consisting of options $A$ and $B$ the relative probability of choosing option $A$ can be increased if a third option with characteristics of $D_A$ is added to the same set of choices \citeyearonly{huberEtAl82}.
\begin{figure}[h]
\centering
\includegraphics[width=0.7\textwidth]{staticFiles/attractionEffect.png}
\caption{Attraction effect in ternary choice.} % Add your description here
\label{fig:attractionEffect} % This labels your figure for reference
\end{figure}
Figure \ref{fig:attractionEffect} shows these options and their respective values in each dimension. It can be seen that option $A$ and $B$ are located at two different ends of the choice space. Option $D_A$ is inferior to option $A$ in both dimensions and $D_B$ is inferior to $B$. With the attraction effect in place, equation \ref{eq:onlyTwoOptions} will change into \ref{eq:attractionProbability}.
\begin{equation}\label{eq:attractionProbability}
\frac{P(A|A,B,D_A)}{P(B|A,B,D_A)} > \frac{P(A|A,B)}{P(B|A,B)} \And \frac{P(B|A,B,D_B)}{P(A|A,B,D_B)} > \frac{P(B|A,B)}{P(A|A,B)}
\end{equation}
Huber noted that although other explanations are still possible, the addition of $D_A$ to the choice set would shift the preferences of people towards dimension 2 because this is where option $A$ appears advantageous \cite{huberEtAl82, bhatia2013associations}. However, this claim has not received unanimous support in future studies in which preference changes have been observed \citep{wedell1991distinguishing}.
\textit{Compromise effect}
When the third option we add is option $C$ instead of $D$, the preference shift happens differently. $C$ is virtually a middle option between $A$ and $B$ and therefore has a value of 50 in each dimension. In this case, the probabilities of choosing $A$ and $B$ will both decrease in favor of $C$, resulting in \ref{eq:compromiseProbability}:
\begin{align}\label{eq:compromiseProbability}
P(A|A,B,C) < P(A|A,B) \And P(B|A,B,C) < P(B|A,B)
\end{align}
Simonson was the first to describe such an effect \citeyearonly{simonson89}. He associated this with a difficulty to select: When people are not sure which attribute is important, they will find a justification to favor a compromise \citep{simonson89}. This argument can explain the reason why an individual may drift towards a middle choice in a three-option choice set. Such a compromise emerges as an important factor, acting as a tie-breaker when the decision maker is unsure between the initial two options.
It is worth noting that it is also possible to ``target'' a particular option from the binary choice set when adding a third option to create a compromise effect. One can add a target $C$ which makes $A$ a compromise option. In this case, the probability of choosing $A$ among this triple will increase, as it is considered a compromise between the remaining two options.
\begin{figure}[h]
\centering
\includegraphics[width=0.7\textwidth]{staticFiles/compromiseEffect.png}
\caption{Compromise effect in ternary choice.} % Add your description here
\label{fig:compromiseEffect} % This labels your figure for reference
\end{figure}
\textit{Similarity effect}
Although Becker et al. \citeyearonly{becker1964measuring} have mentioned it before, the first study of the similarity effect is known to be that of Tversky \citep{tversky1972elimination}. He noted that when faced with a binary choice set consisting of $A$ and $B$ individuals will gravitate towards $A$ more than when faced with a ternary choice set consisting of ${A, B, S_A}$ depicted in figure \ref{fig:similarityEffect}. He explained it by proposing the elimination by aspects theory, which states that one attribute will be chosen as elimination criteria, and all options that do not meet that criteria will be eliminated \citep{tversky1972elimination}. Therefore, in the set of choices ${A, B, S_A}$ if an individual selects dimension 2 as the elimination criteria, both $A$ and $S_A$ will be eliminated, leaving $B$ as a choice. In contrast, if the decision maker prefers dimension 1 more, then option
$B$ will be eliminated, leaving both $A$ and $S_A$ to share the ``victory'', hence resulting in equation \ref{eq:similarityProbability}.
\begin{equation}\label{eq:similarityProbability}
P(A|A,B,S_A) < P(A|A,B) \And P(B|A,B,S_B) < P(B|A,B)
\end{equation}
The similarity effect in the choice set ${A, B, S_A}$ will follow a similar route. Figure \ref{fig:similarityEffect} depicts both sets of choices, where an option similar to $A$ and $B$ was introduced separately.
\begin{figure}[h]
\centering
\includegraphics[width=0.7\textwidth]{staticFiles/SimilarityEffect.png}
\caption{Similarity effect in ternary choice in two scenarios.} % Add your description here
\label{fig:similarityEffect} % This labels your figure for reference
\end{figure}
\subsection{Computational decision making models}
MDbS belongs to the attention-based choice models. These models are type of decision-making models which take into account the attention allocation mechanisms when making decisions. They generally assume that people allocate attention to different attributes and those who receive more attention have more impact on decision making \citep{gabaix2000boundedly}. Before commencing with its underlying mechanisms and assumptions, it is beneficial to discuss the other two models that MDbS influenced by and have been studied extensively. After briefly discussing those models, I will continue with MDbS, its main assumptions, mechanisms, and how it accounts for context effects.
\textit{Multialternative Decision Field Theory}
The very first computational model which could account for all three context effects was Multialternative Decision Field Theory (MDFT) developed by Roe et al. \citeyearonly{roe2001multialternative} as an extension of Decision Field Theory \citep{busemeyer1993decision}. It is a dynamic model of decision making that accommodates multialternative preferential choice situations, which was not possible with Decision Field Theory \citep{hotaling2019quantitative}. MDFT assumes that decision making can be explained in three general mechanisms. First, attention allocation posits that attention switches over time between attributes stochastically. Second, the evaluation mechanism posits that the attribute value of the given option is compared with the average attribute values of other options, which makes sure that each option in the choice set participates in comparison. Third, the evidence accumulation mechanism, which is based on the results of the evaluations, gathers evidence in favor of the alternatives compared. As soon as the evidence gathered hits the externally set relative threshold, the choice is concluded \citep{busemeyer2002survey}. This means that as soon as the difference between the highest and the second highest evidence values is larger than the relative threshold, a choice is made. If this threshold is not met, the decision continues until the pre-set time limit is reached.
MDFT has been confirmed to account for similarity, compromise, and attraction effects in multialternative choice scenarios \citep{roe2001multialternative}. It has previously been tested against such random utility models of choice as Logit and Probit and has been concluded to be a better fit to empirical data \citep{berkowitsch2014rigorously}.
MDFT has also been adapted to account for preference changes \citep{mohr2017attraction} and decision making under time restrictions \citep{diederich2003mdft}.
\textit{Multiattribute linear ballistic accumulator}
The multiattribute linear ballistic accumulator (MLBA) is another attention-based decision-making model first proposed by Trueblood et al. \citeyearonly{trueblood2014multiattribute}. Similar to MDFT, it is also a dynamic model, which can be explained in three general mechanisms: attention allocation, evaluation of alternatives, and evidence accumulation. However, the two models have key differences. Firstly, MDFT emulates the search process of elimination by aspects proposed by Tversky \citeyearonly{tversky1972elimination} assuming that decision makers compare alternatives with one another over time. In contrast, MLBA assumes that individuals make comparisons and accumulate evidence of all alternatives independently from one another at the same time and then accumulate evidence \citep{trueblood15fragile}. Furthermore, MDFT assumes that individuals have limited cognitive capacity to process information when comparing items together, in contrast to MLBA, which considers individuals with unlimited cognitive capacity. Moreover, unlike MDFT, where decision is made based on relative threshold, in MLBA decision is based on absolute threshold, i.e. as soon as one alternative's evidence reaches the threshold, a decision is made in favor of that alternative. Another difference between these two models is the context effects for which they account. Although MDFT does account for attraction, compromise, and similarity effects \citep{hotaling2019quantitative}, MLBA additionally accounts for preference reversals arising from context \citep{trueblood15fragile}.
\subsection{Multialternative decision by sampling}
MDbS has its origins in the theory of decision-by-sampling, which assumes that individual preferences arise from binary, ordinal comparisons of alternatives on given attribute values with reference values from the memory \citep{stewart2006decision}. Unlike it, MDbS assumes that the information required for comparison also comes from the choice environment itself \citep{noguchi2018multialternative}. As in the other two previous models discussed above, its mechanisms can be explained using three stages. The next section will discuss this in detail.
\subsubsection{Mechanisms behind MDbS} \label{subsec:mechanismMDBS}
\textsc{Attention allocation}
According to MDbS, when comparing two tickets between Paris and New York, for example, the price of a ticket would be compared to the prices of other tickets in the choice set and also to the ones and which an individual has previously seen but are not in the current choice set. Comparisons are ordinal, meaning that evidence accumulated toward the ``winner'' at a rate of one irregardless of how large the difference was.
Previously, it was concluded that people tend to compare alternatives that are similar to each other more than dissimilar ones \citep{noguchi2014attraction}. Similarity-based attention is one of the main assumptions of MDbS. To better understand this, let $m_{ij}$ and $m_{kj}$ be two attribute values with $i \neq k \in \{1, \ldots, n_a\}$ and $j \in \{1, \ldots, n_d\}$. MDbS defines the similarity of $m_{ij}$ to $m_{kj}$ as
\begin{align}\label{similarityMDBS}
s_{ij,kj} = \exp \left( - \alpha \left| \frac{m_{ij} - m_{kj}}{m_{kj}} \right| \right) ,
\end{align}
with similarity parameter $\alpha$. Also, generally $s_{ij,kj} \neq s_{kj,ij}.$ Consider also
\begin{align}\label{sumOfSimilaritiesMDbS}
s_{ij} = \sum_{\substack{k \neq l \in \{1, \ldots, n_a\}}} s_{ij,kj} ,
\end{align}
which is the sum of all similarities for attribute $m_{ij}$ to other attributes on the same dimension. Consequently, by dividing this value by the sum of similarities in all other attributes across all dimensions, one can calculate the probability that $m_{ij}$ will be selected for comparison, which will be
\begin{align}\label{probabilityOfComparison}
p_{ij} = \frac{s_{ij}}{\sum_{l \in {1, \ldots, n_a}} \sum_{m \in {1, \ldots, n_d}} s_{lm}} .
\end{align}
\textsc{Evaluation of alternatives}
When evaluating alternatives with each other based on pairwise comparisons, MDbS defines the probability of winning a comparison as
\begin{align}\label{probabilityOneIsFavored}
P(m_{ij} \text{ is favored over } m_{kj}) =
\begin{cases}
F(\beta_0 (| \frac{m_{ij} - m_{kj}}{m_{kj}} |- \beta_1)) & \text{if } A_i > X_i \\
0 & \text{otherwise}
\end{cases} ,
\end{align}
where $F$ is a logistic sigmoid function and $\beta_0$ and $\beta_1$ correspond to the advantage value and the probability that this particular advantage value will be enough to be preferred. For example, consider the case where $\beta_0 = 0.1$ and $\beta_1 = 50$. This would mean that the advantage of 10\% would be preferred with the probability of 50\%. Consequently, if the difference is 20\%, then it will be preferred with the probability of 99\%. The logistic function brings the notion of ``soft'' comparison instead of ``hard'' comparison, in which case small differences would be ignored, while large differences would be extremely preferred \citep{noguchi2018multialternative}.
\textsc{Evidence accumulation}
As mentioned above, in MDbS the accumulation of evidence occurs at a rate of one. For each alternative and for each comparison, in case of winning that comparison, one evidence point is counted towards that alternative. Hence, the probability that evidence will increase by one point will be defined as
\begin{align}\label{probabilityOfEvidenceIncreasing}
p_i = \sum_{j \in {1, \ldots, n_d}} p_{ij} \cdot P(m_{ij} \text{ wins a comparison}).
\end{align}
In order to make a choice, MDbS sets a relative stopping rule $\theta$ following the study of Teodorescu \citeyearonly{teodorescu2013disentangling} which states that when deciding between more than two alternatives, the decision is made when the difference between the highest and the second best evidence is larger than the threshold, or the difference between the maximum and mean evidence becomes larger than the threshold. For computational feasibility, MDbS assumes $\theta = 0.1$, which means that a decision is made when the difference between the maximum and mean average evidence reaches $0.1$. Other parameters given externally are $\alpha, \beta_0, \beta_1$. As a last step, evidence for each alternative is divided by the sum of evidence for the entire choice set to convert them to choice probabilities.
After discussing the main mechanisms behind MDbS, the discussion about how MDbS accounts for attraction, compromise, and similarity effects becomes necessary. The next subsection will shed some light on this matter.
\subsubsection{Accounting for big three context effects}
To effectively illustrate the functioning of MDbS, employing an example choice set that encapsulates the context effects discussed can be beneficial. Therefore, the example dataset depicted in figure \ref{fig:MDBsContextExample} will be used. Although there are five alternatives in the figure \ref{fig:MDBsContextExample}, only three of them will be discussed at a time.
\begin{figure}[h]
\centering
\includegraphics[width=0.7\textwidth]{staticFiles/contextEffectExampleScatterplot.png}
\caption[MDbS’ account for big three context effects]{Example choice set to explain MDbS' account for big three context effects. $A$ and $B$ are considered original two alternatives (binary choice set). $D$ is dominated by $A$ on both dimensions. $C$ acts as compromise between the original two and $S$ is a similar option to $B$. Although these effects would be present in different variations of the choice set (for example one can make $A$ as compromise option), for the simplicity I will concentrate on this example.} % Add your description here
\label{fig:MDBsContextExample} % This labels your figure for reference
\end{figure}
\textit{Accounting for attraction effect}
When adding the option $D$ which is dominated by $A$ in both dimensions to the binary choice set $A$ $B$ (hereafter the binary choice set will be used instead of $A$ and $B$) one creates an attraction effect \citep{huberEtAl82, huberPuto83}. Huber et al. \citeyearonly{huberEtAl82} explained the attraction effect through weight shifts for individuals. The addition of $D$ would cause people to weigh dimension 2 more. Therefore, $A$ and $D$ will have higher ``interest'' among the people, and this is where $A$ wins over $D$.
However, MDbS takes a different approach. Adding $D$ will increase the probability of comparison of option $A$, due to its similarity to $D$ (recall equation \ref{similarityMDBS}). As a consequence, the probability of $A$ winning a comparison will be greater than the probability of $B$ winning a comparison, because $A$ dominates $D$ in both dimensions, while $B$ is only better in one and $A$ and $D$ will be selected more often to be compared with each other. As a result, $A$ will have the highest expected evidence accumulated, $B$ will be the runner-up, and $D$ will have the lowest evidence accumulated towards it. The value of $\alpha$ will determine how much attribute similarity is translated into being selected for comparison.
\textit{Accounting for compromise effect}
In the scenario where option $C$ is added to the binary choice set, the compromise effect arises \citep{simonson89}. This makes $C$ more likely to be chosen because people become uncertain about the importance of attributes and, therefore, experience a choice difficulty. This results in the choice of $C$ as it is easier to justify \citep{simonson89}.
MDbS approach differs here as well. Recall that the probability of accumulating evidence towards an option is a product of its probability to be selected for comparison and its probability to win that comparison, as described in equation \ref{probabilityOfEvidenceIncreasing}. $C$ is more similar to $A$ and $B$ than $A$ is similar to $B$ (and vice versa). This will increase the probability that $C$ will be chosen as a comparison. Although $C$ will not win every comparison, the fact that it will be chosen more as a comparison will increase its probability of accumulating evidence.
\textit{Accounting for similarity effect}
When Tversky \citeyearonly{tversky1972elimination} explained the similarity effect in a choice set consisting of $A$, $B$ and $S$, he explained it via his famous elimination by aspects theory. When $B$ and $S$ are similar to each other, they will be eliminated together or stay together. Hence, having $S$ in the set of options will ``steal'' the probability of choice from $B$.
In MDbS this is explained by $\beta_0$ and $\beta_1$ from equation \ref{probabilityOneIsFavored}. Recall that the reason of having the sigmoid function with arguments $\beta_0$ and $\beta_1$ is to make sure that small differences would be relatively ignored. This is in line with previous literature that states that people tend to ignore small differences between alternatives when making a decision \citep{kalwani1992consumer}. Hence, the small differences between $B$ and $S$ would be ignored, which would translate into a decreased probability that any of them would win over the other when compared. This will indirectly increase the accumulation of evidence for $A$, resulting in ``shared'' evidence between $B$ and $S$.
Not only does MDbS account for the three main context effects, it also successfully accounts for other known effects to decision-making literature, such as the attribute spacing effect \citep{cooke1998multiattribute}; centrality effect \citep{brown2011decision}; background contrast effect \citep{tverskySimonson93}; endowment effect \citep{knetsch1989endowment} and others\footnote{For the full list of the effects MDbS can account for please refer to Noguchi \citeyearonly{noguchi2018multialternative}.}. Overall, the authors claim that MDbS can theoretically account for up to 25 variations in the context effect.
Considering its ability to account for wider range of context effects, MDbS offers a novel and more insightful path for studying multidimensional and multiattribute choice data. It also offers a fine trade-off between the complexity of the model, namely its dynamic attributes, which makes it more practical analytically.
After delving into the theoretical mechanisms of MDbS and its handling of the three main context effects, the focus now shifts to the practical component of the research: the methodology. The following section provides a deeper exploration into the specifics of the research process. It offers a brief overview of the dataset, followed by an in-depth explanation of the steps involved in the analysis.
\subsection{Empirical application}
\textbf{Observational Data}\label{section:observationalDataDescription}
The observational dataset is created by the merger of two sources. The first dataset constitutes a list of all air travel reservations made in Europe on European routes between December 2013 and June 2014, extracted from the MIDT database (Marketing Information Data Tapes). Besides all of the booking details (e.g., number of passengers, price), it also contains the timestamp of booking and the identity of the booking office (all offline and online outlets have unique identifiers). The second source of data contains information on all air travel searches performed on one of the most comprehensive air travel booking services operated by Amadeus S.A.S. This dataset also contains trip specifics, as well as the identifier of the office where the search was performed. Most importantly, the latter dataset contains information on all possible alternatives that could have been presented to the traveler at the time of search, but does not contain information on which of the options (if any) has the traveler chosen. The matching of these two datasets across office identifier, search / booking time, origin and destination of the trip, travel dates, and number of passengers results in a merged dataset that allows us to identify the itineraries chosen within the options offered at the search \footnote{Office ID, trip origin and destination, trip dates and number of passengers are matched exactly. The time interval between the reservation and the preceding search is minimized. If, given exactly matched attributes, the booking was not performed within 24 hours after a given search, the search is declared unmatched. If, given exactly matched attributes, no search is found during the 24 hours preceding a given booking, the booking is declared unmatched. Unmatched searches and bookings are removed from the analysis}. An important limitation of the data is that there is no way of ensuring that the consumer has actually seen the exhaustive list of alternatives available to him/her at the time of booking. However, dataset consists of all the options that they could have seen. Although this is a drawback for a researcher, this is a standard experience for the practitioner (e.g., recommender system designer). Practitioners designing recommender systems need to create algorithms based on the set of existing alternatives without much visibility on the subset of options a particular user will be interested in, or will eventually see.
The matched dataset (previously used by Mottini and Acuna-Agost \citeyearonly{mottiniAcunaAgost17}) consists of 13000 choice sessions with around 1 million choice alternatives in total. Every alternative is a round-trip flight and has a number of attributes including ticket price, date and times of all inbound and outbound flights, number of flights in the itinerary, number of airlines, days before booking, and a few more, less important attributes.
Menus (i.e., choice sets) with one single available alternative do not allow the consumer to make choices and are therefore discarded. Data on choices contain at most 100 alternatives for each choice session, even if more choices potentially existed. As a result, our data are truncated from the right. This creates a large number of menus, including exactly 100 alternatives, some of which may be incomplete. To deal with this oddity, we simply confine our research to menus having between 2 and 99 alternatives (excluding those with one option, since there is no choice involved). In the end, we are left with a dataset with 6,297 choice sessions with 368,723 alternatives in total.
\begin{table}
\centering
\begin{tabular}{l|cccccc}
\hline
Variable & Count & Mean & St.Dev. & Min & Max \\
\hline
Price (in EUR) & 368,723 & 647.12 & 1,105.120 & 59.55 & 16,997 \\
Trip duration (in minutes) & 368,723 & 518.98 & 555.04 & 70 & 2,715 \\
Number of flights & 368,723 & 2.94 & 0.95 & 2 & 6 \\
Number of airlines & 368,723 & 1.25 & 0.45 & 1 & 5 \\
Menu size & 368,723 & 58.077 & 30.267 & 2 & 99 \\
Days before departure & 368,723 & 32.36 & 38.03 & 0 & 340\\
Domestic travel & 368,723 & 0.49 & 0.49 & 0 & 1\\
Intercontinental travel & 368,723 & 0.06 & 0.23 & 0 & 1\\
\hline
\end{tabular}
\caption{Descriptive statistics of variables in observational data.}
\label{tab:descriptiveStats}
\end{table}
These are the attributes that are designated as vertical in the choice process. For the purposes of this dissertation, it is assumed that consumers prefer lower values for each of them (e.g., all consumers prefer lower prices, shorter trips, fewer layovers, and not having to change airlines too frequently). Apart from vertical attributes, there are also three attributes that do not vary across alternatives within each menu. These are the number of days between when the choice was made and the start of the trip, whether the trip is
domestic or international and whether it is intercontinental. In addition to vertical attributes, the data also contains two sets of horizontal attributes, the departure times and the dates of outbound and inbound flights. These attributes are treated as horizontal, as there is no clear way of defining consumer preferences over them. To eliminate potential scale effects, z-score normalization on vertical attributes was performed as $Z = \frac{{x - \overline{x}}}{{\sigma}}
$ where $\overline{x}$ is the mean and $\sigma$ is the standard deviation of the variable $x$.
Due to the MDbS nature of comparing dimensions with one another, for the purposes of this study, I cannot use variables that do not differ between menus. Additionally, I cannot utilize horizontal attributes in my analysis because they do not follow the standard ``the greater the better mathematical approach''. However, these variables are used later in this dissertation in chapters \ref{chapter:jmrPaper} and \ref{chapter:hicssPaper}. As a result, I am bound to utilize only four vertical attributes in my analysis. Table \ref{tab:descriptiveStats} provides descriptive information about the variables. I have also multiplied the four vertical variables by \-1 to convert them to a negative scale due to the MDbS nature of comparing absolute values and the assumption that consumers prefer lower values along vertical dimensions.
\textbf{Experimental data}
In the course of this study, I also introduce an additional dataset sourced from a controlled experiment conducted by Noguchi \citeyearonly{noguchi2018multialternative}, distinct from the primary observational data. Although not as diverse in terms of alternatives, dimensions, and choice sets, these experimental data carry significant value, not as a principal analytic focus, but rather as a means to corroborate my main findings. I will apply the same analytical techniques employed in the observational analysis and utilize experimental dataset as a robustness check to verify the validity of my results. Henceforth, the role of these data is primarily confirmatory.
This dataset comes from an experiment conducted by Noguchi \citeyearonly{noguchi2018multialternative} in which 503 participants, aged 18 to 75 years, participated in Amazon Mechanical Turk, resulting in a total of 5295 observations. Participants faced eight randomly sampled decision scenarios with descriptions consisting of two and three alternative sets, each ternary choice set containing only one of the attraction, compromise, similarity effects. For ternary choice sets, to create a context effect, one alternative was randomly selected as focus and third alternative was generated following three scenarios: a) for attraction effect scenario, third options' both dimensions were reduced by 25\% of the difference between the remaining two options' dimensions; b) for compromise scenario, the third option was generated in a way that it would make the randomly chosen target a compromise; c) in similarity scenario 2\% of the difference was added to one dimension while 2\% was subtracted from another dimension for the third option. The dimensions of the alternatives are described in the table \ref{tab:noguchiDescriptions}.
\begin{table}
\centering
\begin{tabular}{l|lll}
\hline
Product & Dimension & Alternative A & Alternative B \\
\hline
\multirow{2}{*}{Mouthwash} & Breath & 4.5 hours & 7.2 hours \\
& Germs killed & 77\% & 56\% \\[2ex]
\multirow{2}{*}{Exercise class} & Fee & \$9.49 & \$6.49 \\
& Calories & 356 kcal & 259 kcal \\[2ex]
\multirow{2}{*}{Box of chocolate} & Amount & 26 oz & 33 oz \\
& Variety & 9 & 5 \\[2ex]
\multirow{2}{*}{GPS} & Update & 3.04 Hz & 5.62 Hz \\
& Accuracy & 4.97 m & 7.83 m \\[2ex]
\multirow{2}{*}{Mobile battery} & Price & \$19.93 & \$13.49 \\
& Talk time & 14.55 hours & 9.25 hours \\[2ex]
\multirow{2}{*}{Light bulb} & Life & 1309 hours & 1923 hours \\
& Price & \$1.35 & \$2.50 \\[2ex]
\multirow{2}{*}{Air purifier} & Noise & 64.7 dB & 39.3 dB \\
& Efficiency & 325 cfm & 203 cfm \\[2ex]
\multirow{2}{*}{Strawberry} & Quantity & 407 g & 452 g \\
& Price & \$2.58 & \$2.85 \\
\hline
\end{tabular}
\caption[Attribute values used in experiment]{Attribute values used in the experiment. Sourced from Noguchi \citeyearonly{noguchi2018multialternative}.}
\label{tab:noguchiDescriptions}
\end{table}
\textbf{Parameter optimization}
Recall the three dynamic parameters for MDbS, $\alpha, \beta_0, \beta_1$, which were discussed in section \ref{subsec:mechanismMDBS}. They allow MDbS to take into account various context effects. Noguchi \citeyearonly{noguchi2018multialternative} demonstrates MDbS performance using a fixed set of parameters throughout the article using: $\alpha = 3, \beta_0 = 0.1, \beta_1 = 50$.
Those parameters are essential controls of the behavior of the model and they create the underpinnings of the choice set, impacting the generated choice probabilities. Therefore, it is fundamental to identify the optimal parameters that will fit the observed data. While identifying the optimal parameters could ideally be purely theory-driven, in reality the theoretical guidance will often fall short. This will leave a plethora of potential parameter values. Hence, this requires a systematic search method to explore the parameter space and identify the optimal parameters that would fit the data the best.
\textit{Parameter space definition}
Before continuing further with the method, one must first define the parameter space over which the search process will commence. Recall that there are three parameters to be optimized, which were $\alpha, \beta_0, \beta_1$. Two of them, $\beta_0$ and $\beta_1$, have theoretical boundaries.
The effect of the attribute range was first investigated by Mellers \citeyearonly{mellers1994trade}, who described the tendency of people to scale the perceived attractiveness of an alternative in a given attribute using the entire range of that attribute. Therefore, $\beta_0$ in MDbS represents the fraction of the difference between attributes compared to the entire attribute range, which is bounded between 0 and 1. On the other hand, $\beta_1$ represents the percentage of preference of that difference for an individual; hence, it is also bounded with values between 0 and 100. I have created 99 $\beta_0$ values evenly spaced between 0 and 1 and 99 inclusive $\beta_1$ values evenly spaced between 0 and 100.
The parameter $\alpha$ on the other hand, does not have a theoretical upper bound. However, since it is used to determine which alternatives to compare with each other, it must be greater than 0 because otherwise no alternative will be selected for comparison. I have randomly generated 4,000 samples where the alpha ranged between 0.1 and 10. My observations have demonstrated that the performance of MDbS significantly deteriorates when $\alpha \ge 5$. Hence, to balance the need for a flexible model with the requirement of stable performance, an upper bound of 5 has been set for $\alpha$. I have created 49 $\alpha$ values between 0.1 and 5.
In total, the full parameter space has been created with combinations of the three parameter values reaching 480,249 triples.
\textit{Optimization method}
Parameter optimization is a task of high importance in many scientific and engineering applications, where the goal is to find the optimal values of a set of parameters that best fit a given model or system. There are various methods available for parameter optimization, ranging from differential equation-based methods to brute force and other optimization algorithms. I have chosen the differential evolution algorithm proposed by Storn \citeyearonly{storn1997differential} for this purpose. It has several advantages over other algorithms. Firstly, it can be easily implemented. Secondly, it is ideal when the parameter space is large \citep{lin2019applying}. Thirdly, it is especially suitable for complex and non-linear functions \citep{omran2009bare}.
The way differential evolution works resembles other genetic algorithms. First, it creates an initial population $P$ with a size of $n$ within a given parameter space $S$ and assesses its fitness using the evaluation metric $F$. Then, it randomly selects three members of $P$ and creates a new member. If it is better than a randomly selected one within this triple, it replaces it. This process continues until termination criterion is met, which is either: a) $F$ has reached its global minimum, b) the number of iterations has reached the threshold, or c) $F$ has not improved considerably within the predefined number of iterations. The pseudocode below describes its workflow:
\begin{algorithm}
\caption{Simplified Differential Evolution.}
\begin{algorithmic}[1]
\State Initialize population of $P$ from the parameter space $S$
\While{not met termination criterion}
\For{each individual in $P$}
\State \textbf{Mutation:} Select three distinct individuals from population. Compute the donor by adding the weighted difference of two individuals to the third.
\State \textbf{Crossover:} Create trial individual by mixing parameters of current individual and the donor, decided by random draw and crossover rate.
\State \textbf{Selection:} Compare trial and current individuals on using $F$. If the trial performs better, replace the current individual with the trial in the population.
\EndFor
\EndWhile
\State \Return Best individual from the final population as optimal parameters.
\end{algorithmic}
\end{algorithm}
Differential evolution itself has parameters that must be defined in advance. The population size parameter in this algorithm defines the number of candidate solutions it considers during each iteration. Those candidates are selected following uniform distribution in the parameter space which achieves evenly distributed candidates. There is a trade-off between high population size leading to finer exploration of parameter space and low population size leading to faster conversion, albeit not optimal. I set it to 15 to achieve both good exploration and conversion speed. The second parameter, the crossover probability, controls the extent to which the algorithm combines information from different solutions. The higher value will further diversify the population, encouraging exploration of new regions in the parameter space. On the other hand, lower values will lead to more exploitation of the current space. I have set this to 0.5. For other parameters, I will use the values suggested in the literature \citep{omidi2020differential}.
\textit{Evaluation metric}
After discussing the importance of optimal parameter search and defining the optimization algorithm, the remaining question is the evaluation metric of the MDbS. Previous studies which have applied various dynamic choice models to experimental data have used the mean absolute error of aggregate choice shares for the entire dataset as the main metric. Albeit an interesting approach itself, this will not be a feasible approach for me because the experimental data these models have been applied to entailed ternary choice sets, whereas the observational data are not ternary. It comprises choice sets with minimum of 2 and maximum of 99 alternatives.
Designers and engineers of choice sets have long used ``Top n'' accuracy metrics when designing choice sets or testing the performance of statistical models \citep{ricci2015recommender}. ``Top $n$'' accuracy metric measures whether or not the true class of the option matches the top $n$ predictions of the model. I will follow and adopt this metric because it is well established in the literature, mirrors the real-world decision making, and it fits the contribution of the thesis the best. I will use ``Top 1'' accuracy, which ranges from 0 to 1 as my optimization metric for an individual choice set. Because choice sets in the data vary significantly in size, I will use the average Top 1 accuracy metric weighted by menu size. This will ensure that smaller menus contribute proportionally to their sizes. Also, to comply with the aim of minimization of the differential evolution algorithm, I will multiply this metric by -1. As an additional measure to explore the parameter space thoroughly, I employ the multiple-run approach for the differential evolution algorithm. Specifically, I will execute it ten times across the entire dataset. This reposition will allow me to further explore the parameter space, mitigating the risk of missing any region that can potentially contain an optimal solution.
\subsection{Results}
This section presents the initial results of parameter optimization, followed by the outcomes derived from choice modeling. First, an examination of the results obtained from observational data takes precedence. Afterwards, a concise discussion of the results derived from experimental data will accompany this analysis.
\textbf{Observational data results}
When looking at optimization results on observational data one can immediately see that the values of $\beta_0$ tend to fluctuate around 0.82 and 0.96 while $\beta_1$ is generally below 10\%. which indicates that MDbS tends to be more strict in terms of defining the winners when comparing, on average preferring 90\% of the ``advantage'' in a given dimension only a little shy of 15\% of the time. Also, it appears that the $\alpha$ values tend to be preferred in the lower half of the parameter space, so only very similar alternatives were chosen by the model for comparison. This behavior is understandable considering that the average choice set had 55 alternatives. Table \ref{tab:optimizationAmadeusResults} contains the results of the parameter optimization using the differential evolution algorithm. It is worth noting that, its top 1 accuracy performances, albeit higher than random chance, still would fall far behind the pure statistical models, such as MNL based ones.
At first sight, such model behavior might seem surprising. Recall that the nature of MDbS is to compare alternatives with each other and collect evidence based on the won comparisons. In the choice experiments, the usual size of the menu is three, and only one of the context effects is generated at a given time. However, in observational data, the number of alternatives in the menu is much higher. The presence of a large number of alternatives potentially introduces also other context effects. Also, MDbS is bound to only dimensions which are mathematically comparable with each other. In the observational dataset there also present horizontal attributes for which only the decision maker can decide in a given scenario whether or not given the same price, flight duration, the flight which is at 5:00 in the morning is better than the one that is at 14:00 in the afternoon.
\begin{table}
\centering
\begin{tabular}{ccccc}
\hline
Iteration & $\alpha$ & $\beta_0$ & $\beta_1$ & Average top 1 accuracy \\
\hline
1 & 2.58 & 0.912 & 8.832 & 0.125 \\
2 & 1.622 & 0.948 & 6.001 & 0.124 \\
3 & 1.88 & 0.843 & 6.876 & 0.124 \\
4 & 1.883 & 0.832 & 6.83 & 0.124 \\
5 & 1.856 & 0.91 & 8.297 & 0.124 \\
6 & 2.154 & 0.954 & 8.021 & 0.124 \\
7 & 2.204 & 0.963 & 8.52 & 0.124 \\
8 & 0.234 & 0.859 & 7.03 & 0.123 \\
9 & 0.559 & 0.829 & 55.076 & 0.122 \\
10 & 0.235 & 0.844 & 58.893 & 0.122 \\
\hline
\end{tabular}
\caption{Optimization results for observational data.}
\label{tab:optimizationAmadeusResults}
\end{table}
I have estimated two models by using the random effect probit model with standard errors at cluster levels. The first model only included vertical attributes, whereas the second model extended the first one through the addition of the output from the MDbS model. In both cases, it seems that individuals have strong preferences for faster alternatives with lower prices and fewer layovers. This supports our initial assumption that individuals prefer lower values of vertical attributes.
\begin{table}
\centering
\begin{tabular}{lcc}
\hline
& Model 1 & Model 2 \\
\hline
Price & -0.309*** & -0.282*** \\
& (0.006) & (0.006) \\[1ex]
Trip duration & -0.185*** & -0.158*** \\
& (0.007) & (0.006) \\[1ex]
Number of flights & -0.195*** & -0.178*** \\
& (0.007) & (0.007) \\[1ex]
Number of airlines & -0.262*** & -0.245*** \\
& (0.008) & (0.008) \\[1ex]
MDbS output & & 2.085*** \\
& & (0.097) \\[1ex]
Constant included & Yes & Yes \\[1ex]
Menu size as control & Yes & Yes \\[1ex]
Number of observations & 368,723 & 368,723 \\[1ex]
Akaike information criteria & 48,532.341 & 47,968.61 \\[1ex]
Log-likelihood & -24,260.171 & -23,977.305 \\[1ex]
\hline
\end{tabular}
\caption[Outputs of Probit model for observational data]{Outputs of Probit model with random effects for observational data. Standard errors in parentheses. Statistical significance levels: *** $p<0.01$, ** $p<0.05$, * $p<0.1.$.}
\label{tab:amadeusProbitResults}
\end{table}
Recall that variable ``MDbS output'' refers to the probabilities produced by MDbS. Model 2 results show a positive and statistically significant effect for the information provided. It shows that MDbS is able to capture additional information about the choice by accounting for context effects. To better understand the significance of this result, figure \ref{fig:marginsAmadeusGraph} shows the average marginal effects of the information provided by the computational model. One can immediately observe the downward trend. It is not surprising. As the number of alternatives increases, each additional alternative adds less to the likelihood of choice than the previous. In the context of menus, it could imply that when there are fewer options (smaller menus), the likelihood that any particular choice is selected is more significantly influenced by MDbS output. On average, for every 0.5 increase in MDbS output, the probability of choice has increased by 0.037 percentage points \footnote{Marginal effect of MDbS output across the whole dataset was 0.074.}. This effect was as high as 0.12 percentage points for menus containing as few as 5 alternatives.
\begin{figure}[h]
\centering
\includegraphics[width=0.7\textwidth]{staticFiles/marginsAmadeusGraph.png}
\caption[Marginal effects of MDbS output]{Average marginal effects of MDbS output with respect to different menu sizes. Horizontal lines represent 95\% confidence interval boundaries.} % Add your description here
\label{fig:marginsAmadeusGraph} % This labels your figure for reference
\end{figure}
\textbf{Experimental data results}
At first sight, the results of optimization indicate that the optimal parameters differ between datasets. Considering the nature of these two datasets, such a result is expected. Although the optimal $\alpha$ tends to be higher than the one in the observational data, both the optimal $\beta_0$ and $\beta_1$ values are lower than their counterparts. A higher $\alpha$ can indicate that for smaller choice sets, MDbS tends to be less strict in comparison criteria. While the performance metrics seem higher than for the observational data, the menus are considerably smaller. Table \ref{tab:optimizationNoghuchiResults} gives further information.
\begin{table}
\centering
\begin{tabular}{ccccc}
\hline
Iteration & $\alpha$ & $\beta_0$ & $\beta_1$ & Average top 1 accuracy\\
\hline
1 & 2.888 & 0.572 & 4.567 & 0.52 \\
2 & 2.918 & 0.577 & 4.598 & 0.52 \\
3 & 2.939 & 0.569 & 4.655 & 0.52 \\
4 & 2.747 & 0.579 & 4.349 & 0.519 \\
5 & 2.936 & 0.743 & 3.911 & 0.518 \\
6 & 2.925 & 0.715 & 3.967 & 0.518 \\
7 & 2.997 & 0.706 & 4.098 & 0.518 \\
8 & 3.342 & 0.61 & 4.93 & 0.517 \\
9 & 2.01 & 0.494 & 3.788 & 0.517 \\
10 & 2.01 & 0.494 & 3.788 & 0.517 \\
\hline
\end{tabular}
\caption{Optimization results for experimental data.}
\label{tab:optimizationNoghuchiResults}
\end{table}
Overall, considering the differing natures of these two datasets, comparing two optimal parameter combinations would not give any useful knowledge. However, this is not the case for the results from the choice modeling. These results follow the ones from observational data and confirm them. As with field data, here, the MDbS output is proven to provide statistically significant information for a choice model with coefficient in the positive direction. A 0.5 increase in MDbS' ``assessment'' about the alternative resulted in a 0.46 percentage points increase in the actual choice probability among the participants. This effect did not differ between the choice sets having two or three alternatives.
\begin{table}
\centering
\renewcommand{\arraystretch}{1.5}
\begin{tabular}{lcc}
\hline
& Model 1 & Model 2 \\
\hline
X & 0.000*** & 0.000*** \\
& ($0.000$) & ($0.000$) \\[1ex]
Y & 0.000* & 0.000 \\
& ($0.000$) & ($0.000$) \\[1ex]
MDbS output & & 2.518*** \\
& & (0.317) \\[1ex]
Constant included & Yes & Yes \\[1ex]
Menu size as control & Yes & Yes \\[1ex]
Number of observations & 5,295 & 5,295 \\[1ex]
Akaike information criteria & 6,987.151 & 6,893.77 \\[1ex]
Log-likelihood & -3,489.576 & -3,441.885 \\
\hline
\end{tabular}
\caption[Outputs of Probit model for experimental data]{Outputs of Probit model with random effects for experimental data. Standard errors in parentheses. Statistical significance levels: *** $p<0.01$, ** $p<0.05$, * $p<0.1.$.}
\label{tab:noguchiProbitResults}
\end{table}
\subsection{Conclusion}
In this study I have applied MDbS to observational data and showed the consistency of my findings using experimental data. This is the first account of an application of a computational model to real-world choice data of this magnitude. The results indicated that computational models can account for context effects that affect choice behavior not only in experimental settings but also in field settings.
The results of this study create implications for online marketplaces. In today's world, these platforms aggregate immense amounts of products and services, providing consumers with dozens of choices. To help consumers in their choice, these platforms employ sophisticated algorithms which aim to curate product lists and create recommendations with a side goal of influencing the buying decisions of individuals. By applying mathematical decision-making models to choice datasets, these platforms can gain crucial information about the context within the choice sets, which might influence choice decisions towards particular alternatives. This information may also be used to create product bundles with a heterogeneous context to satisfy the needs of consumers.
This study has limitations. I have utilized only one computational model, namely MDbS. This limits the generalizability of my results. Different models ``behave'' differently, and although they are trying to capture the same effects, applying other decision models and investigating their differences can be an interesting avenue to pursue. Another limitation is that the proposed approach was applied only to data originating from one type of choice setting, namely, air travel. Application of this approach to other types of multi-dimensional, multiattribute choice data may help to better generalize the results.
The use of MDbS in the current study has shown it has the ability to potentially capture a wide range of context effects, including attraction, compromise, and similarity. While these results have yielded valuable insights, the general nature of MDbS and other computational models is their inability to successfully isolate these effects from one another. The main reason for that is that they have only been tested in experiments with one effect present at a time. When the number of alternatives in the dataset increases the potential interplay between options and the existence of other context effects come into play.
My findings provide a strong foundation that leads to a crucial but also challenging future direction: the development of a methodology which would allow disentangle this ``general'' context effect. I will computationally differentiate among three main components, attraction, similarity, and compromise in a multi-dimensional, multialternative choice setting. This goal provides great motivation for the next chapter of my thesis.
\newpage
\section{Enhancing Choice Modeling in Multi-Attribute, Multi-Alternative Settings\footnote{This chapter is based on a joint work with my supervisor Zakaria Babutsidze, William Rand, Nobuyuki Hanaki, Ismael Rafai, Rodrigo Acuna Agost and Thierry Delahaye.}}\label{chapter:jmrPaper}
\begin{abstract}
Previous approaches to modeling the effect of context on choices consider neat, compact environments, often in laboratory settings. Such an approach severely limits the study of context effects and, as a consequence, the applicability of findings. In this paper, the authors generalize the existing approach in modeling choice with context effects and apply it on large-scale observational data. The authors consider three main context effects: the attraction, compromise, and similarity effects. The proposed methodology relies on an ex ante calculation of each context effect measure for every alternative in the choice set. This approach minimizes the computational complications of estimating the resulting choice model. The proposed approach is applied to two empirical settings: the choice of airfare using observational data and the choice of daily commute mode using data from a stated choice experiment. The presence of attraction and similarity effects in both empirical settings is demonstrated. The authors also document the existence of the reverse compromise effect in airfare choice, highlighting the fact that travelers possess rigid rankings among flight attributes and are essentially maximizing their utility in terms of one (or few) attribute(s).
\end{abstract}
\subsection{Introduction}
The fact that behavioral biases exist in individual decision making is well established (see Dowling et al. \citeyearonly{dowlingEtAl20} for a recent review of evidence). One type of systematic departure from the classic utility maximization approach that seems particularly important is a set of context effects \citep{truebloodEtAl13, kocherEtAl19}. The theory behind these effects posits that the context in which choices are made influences the decision. While the choice context could have a very wide meaning, in this literature it is the availability and nature of choice alternatives which is referred to as ``context'' \citep{tversky1972elimination, huberEtAl82, simonson89}.
Context effects have been systematically studied in marketing and psychology \citep{kivetz04, roodrkerkEtAl11, frederickEtAl14, dotsonEtAl18}. However, virtually all such studies have used controlled experiments in neat, compact settings. Namely, the settings where decision makers are presented with few options and (very) few attributes across which these options differ. In contrast, most actual choices take place in much messier environments. Especially today, when much of our search and shopping activity has shifted online. Proliferation of search engines allows each option to be easily compared with many alternatives across many different characteristics. In chapter \ref{chapter:simulationStudy} I focused on quantifying context as an aggregate. Although my previous study could successfully identify context in complex, multiattribute setting, the knowledge about the prevalence of context effects in these environments is still scarce. Precise measurements of context effects in multi-option and multiattribute setting is one way to contribute. Defining how to measure these effects is a minimum requirement for proceeding to evaluate the existence of context effects using observational data.
Recent attempts in computer science have been made to define some of these effects. The machine learning community has incorporated context effects in discrete choice models applied to observational data \citep{pfannschmidt2019learning, bowerBalzano20}. However, the objective was to increase the prediction accuracy of the choice models \citep{tomlinsonBenson21}. As a result, incorporation of context effects takes the form of generalizing choice models to allow for departure from the strict rationality assumptions\footnote{Recent examples of this approach are Contextual Multinomial Logit by Yusefi Maragheh et al. \citeyearonly{yousefi2020choice} and the linear context logit of Tomlinson and Benson \citeyearonly{tomlinsonBenson21}.}. These proposed generalizations of estimated functional forms generally do not distinguish across various different types of context effects. Additionally, these approaches often run into computational difficulties, i.e. the estimation process is NP-hard \citep{yousefi2020choice}.
In this chapter, I propose measures of different context effects in multi-option and multidimensional settings. Following Rooderkerk, Van Heerde, and Bijmolt \citeyearonly{roodrkerkEtAl11}, I consider three context effects - attraction, compromise, and similarity effects. Although I have discussed each of these three effects in more detail in the chapter \ref{chapter:bigThreeContextEffectsDescription}, a very brief recap of this ``trinity'' seems appropriate. The attraction effect refers to the increase in attractiveness of a set of options as a result of adding an alternative to the choice set, the compromise effect refers to the inclination of consumers to prefer options that represent a compromise across extreme sets of alternatives, while the similarity effect refers to the drop in choice likelihood for an alternative once another similar alternative has been added to the choice set. Each of the measures corresponding to the three aforementioned effects requires a specific approach to make the measurements applicable to the observational data. Each of these measures is calculated prior to choice estimation, which avoids computational problems. After presenting the generalized measures of the three effects, we perform an empirical analysis of the choices based on the new measures using observational data. We use an extensive dataset of airfare choices for this exercise. We identify that attraction and similarity effects influence the choices in air-travel booking data. We also detect a reverse compromise effect that seems to indicate that air travelers consistently prefer extreme alternatives (i.e., the cheapest or the shortest flight) to alternatives that constitute a compromise among extreme options.
\subsection{Context effect and choice modeling}
Over the years, multiple empirical models have been developed to model context effects. Empirical approaches usually model context effects in either the structural part of utility or in the error covariance part \citep{kamakuraSrivastava84, dotsonEtAl18}. Some of these models have the capacity to take into account multiple effects at the same time \citep{tverskySimonson93, orhun09}. These models extend a classical random utility model \citep{mcfadden01} in multiple directions using discrete choice modeling \citep{benAkivaLerman85}. However, Rooderkerk, Van Heerde, and Bijmolt \citeyearonly{roodrkerkEtAl11} present a unifying model that takes into account all three context effects. Instead of using advanced statistical techniques to address violations of utility maximization assumptions associated with the existence of context effects \citep{luce59}, their approach focuses on additive specification and ex ante calculation of individual measures for each of the three context effects for each item in the menu. Namely, the authors assume that the choice estimator is additive in three context effects (along with a generic preference-driven part) and develop the methodology of quantifying three effects for each alternative prior to calculating the estimator. This is a particularly flexible approach, which also ensures that the researcher does not run into computational difficulties (i.e., $NP$ hard calculations). I follow the suite and formulate the utility that a consumer c attaches to an option $i$, under a given menu $m$, as being additive in two parts:
$$U_{c,i}^m = u_{c,i} + v_{c,i}^m$$
The first summand in this equation $u_{c,i}$ denotes an inherent utility that the consumer $c$ can derive from the option $i$. This part depends only on the tastes of the consumer $c$ towards the characteristics of the option $i$. It is independent of the other options contained in the menu. The second summand $v_{c,i}^m$, denotes the context-dependent utility. I additionally assume that the context-dependent part of the utility can be represented as a linear combination of three contextual effects,
$$v_{c,i}^m = a_1 \text{Attraction} + a_2 \text{Compromise} + a_3 \text{Similarity}.
$$
Thus, the measures of three context effects that are necessary to estimate empirical discrete choice models based on the utility formulation above need to be computed ex ante. Measures developed by Rooderkerk, Van Heerde, and Bijmolt \citeyearonly{roodrkerkEtAl11}, are adapted to experimental data with a small number of alternatives in the choice set and a small number of attributes characterizing alternatives. This significantly limits the application of the unifying model of context effects. In the next section, I present a generalization of three context effect measures to multi-option, multi-attribute environment which will further allow for the application of the unifying model to observational data.
\subsection{Generalizing context effect measures}
\textbf{Approach to generalization}
Naturally, generalizing across many alternatives and many attributes presents challenges in both dimensions. The fact that theoretical underpinnings of the three effects are diverse does not simplify the task. In the following sections, I will discuss specificities involved in the generalization of each measure. First, however, I focus on common challenges.
Conceptualizations of contextual effects commonly hinge on the choice frequency comparisons between two alternatives. For example, in case of attraction effect, if adding a third alternative to a two-item menu induces some of the consumers to switch their choices to the other incumbent alternative - one could conclude that attraction effect is present. This is suitable for experimental setups where the researcher has control over menus and can observe choices in both cases (i.e., in case of an original two-item menu, as well as after adding the third item). However, given that the aim is to generalize context effect measures for application to a wider range of situations, and most importantly to observational data, it is necessary to take a more fine-grained view and quantify the context in which each of the alternatives is embedded. Quantifying the choice context for each alternative would create an opportunity to study the effect of the context on choice probabilities through inference across (very) different choice sets. Such an approach would be general enough to consider not only the addition of a new alternative to the menu, but also any alteration of attributes for any of the items in the menu. For example, increasing the price of an alternative could decrease the probability of its choice. This would have a direct effect on the choice probabilities of other alternatives. However, the same price increase could also change the choice context and have additional knock-on effect on choice probabilities of (at least some) alternatives.
\begin{figure}[t]
\centering
\includegraphics[width=0.7\textwidth]{staticFiles/contextEffectZaksScatterPlot.png}
\caption[Accounting for attraction effect]{Visualization of accounting for attraction effect. \\ Note: The figure represents three alternative choice sets, each comprising of six options, described along two characteristics. Options A, B, C, E and F and common across three menus. Menus differ in the identity of the 6th option (D, D' or D").}
\label{fig:attractionZakVisualization}
\end{figure}
Rooderkerk, Van Heerde, and Bijmolt \citeyearonly{roodrkerkEtAl11} take this approach for simple two-attribute products. Using the attraction effect as an example once more, the idea is to quantify how much attraction power does a given menu provide to a given alternative. If option $A$ dominates option $B$ (i.e., it is superior in at least some attributes and not inferior in any of the attributes), while option $B$ is not dominated by option $C$, the attraction power of $A$ compared to $C$ could be measured by the degree to which option $A$ is better than option $B$. The more pronounced the dominance, the more pronounced the attraction effect. However, once we leave a neat context of tree-item menus, we wonder into a possibility that option A dominates not one, but multiple alternatives at the same time. Consider the situation depicted in figure \ref{fig:attractionZakVisualization}. Here we have the menu with six alternatives ${A, B, C, D, E, F}$ each of which is characterized by two attributes $V1$ and $V2$. In this example, option $A$ dominates three alternatives ${B, C, D}$. To simply extend the approach by Rooderkerk, Van Heerde, and Bijmolt \citeyearonly{roodrkerkEtAl11} and calculate the attraction power of the alternative $A$, we could find the center among the three dominated variables and then measure the distance. Such a measure would capture the difference between two choice sets ${A, B, C, D, E, F}$ and ${A, B, C, D', E, F}$. In the latter case, the attraction power of $A$' is lower because the option $D'$ is closer to $A$ than $D$. However, such a measure would not accurately capture the difference between scenarios ${A, B, C, D', E, F}$ and ${A, B, C, D", E, F}$. In the latter case $A$ dominates only two alternatives ${B, C}$. Therefore, the setting changes qualitatively. Such qualitative differences are avoided in experimental settings by design. However, they are prevalent in observational data. While it is acknowledged that the move from $D$ to $D'$ changes the choice context, I argue that the context change is more pronounced in the case of the move from $D'$ to $D"$. Although the ideal measure would combine the characteristics of the number of dominated alternatives and the (some measure of average) distance between the focal alternative and the group of dominated options, in this article I take the approach of focusing on the former, as this is likely to have a more pronounced impact \footnote{Combining frequency and distance measures in one metric requires arbitrage across the two drivers of context effects. It is not clear how to solve such a problem (that is, it is not clear if dominating one option that is at a certain distance from a focal alternative generates more or less attraction than dominating two alternatives that are at a half that distance).}.
As a result, our approach would capture the context change between ${A, B, C, D, E, F}$ and ${A, B, C, D", E, F}$ or ${A, B, C, D', E, F}$ and ${A, B, C, D", E, F}$, but it will not evaluate the context difference between ${A, B, C, D, E, F}$ and ${A, B, C, D', E, F}$. In what follows, the same approach is applied to similarity and compromise measures.
Once one moves towards choices which have multiple attributes, it is quick to realize that there are two distinct types of choice characteristics that our measures should potentially handle. One type of attributes constitute product characteristics over which preferences are fairly similar for all customers, and their effects can be readily anticipated from basic economic theory. These attributes can easily be ordered from most preferred to least preferred. The most obvious of these characteristics is price. One can assume that every customer would prefer to obtain a given product for a lower price. We call such product characteristics vertical attributes. These are usually attributes that can be represented using numeric values. Previous work measuring context effects only considers such (vertical) attributes \citep{trueblood2014multiattribute, noguchi2018multialternative, noguchi2014attraction}. This is a requirement for defining preferential relationships that are necessary to identify attraction and compromise effects. The same approach is adopted, where I consider only vertical attributes when defining attraction and compromise effects.
On the other hand, there exists another set of attributes where there is no obvious, homogeneous ordering. For example, consider the color. There is no theoretical ground to assume that all consumers would prefer a car that is blue over a car that is green (all other attributes remain constant). The same is true about attributes that at first sight are not strictly labeled as categorical, for example, time. When buying a cinema or plane ticket, there is no theoretical reason for explaining how a ticket for 15:00 is better or worse than one at 17:00. I refer to these as horizontal attributes. The potential heterogeneity between decision makers in ordering such attributes makes inclusion of such features in the calculation of attraction and compromise effects impossible. In experimental settings, these attributes are often constant between treatments to avoid confounding effects. However, in the field, this usually cannot be done. Therefore, the study of context effects with observational data requires them to be statistically controlled.
However, unlike the measurement of attraction and compromise effects, measuring the similarity across the alternatives does not require the existence of a single universal ranking. In fact, many clustering methods can identify options that are more or less similar to each other based on a wide range (numeric and categorical) of variables. Therefore, in what follows, I will incorporate all (vertical, as well as horizontal) attributes in the measurement of similarity between a pair of alternatives.
\textbf{Attraction effect}
Previous studies of the attraction effect concentrate on carefully designed small choice sets in experimental settings \citep{huberEtAl82, huberPuto83}. In such settings, an alternative is added to the choice set in a position that is unequivocally inferior to (only) one of two items already present in the menu. Notice again that identification of inferiority requires the attribute under consideration to be vertical, and this cannot be achieved with horizontal attributes. This manipulation introduces an asymmetry between the two incumbent alternatives; one alternative now dominates the decoy, while the other does not. The attraction effect implies that such manipulation increases the attractiveness of the dominant incumbent option with respect to the other incumbent alternative.
A standard measure of the attraction effect considers a trade-off between two (vertical) characteristics. Let us consider $i \in \mathbb{N}$ vertical attributes $V_i$ for a set of two options A and B. In two dimensions $n = 2$, we start with $V_1(A) > V_1(B)$ and $V_2(A) < V_2(B)$, and then introduce an alternative $C$ such that $V_1(A) > V_1(C) > V_1(B)$ and $V_2(C) < V_2(A) < V_2(B)$. Under such circumstances, $C$ is dominated by $A$, but not by $B$. This introduces asymmetry in consumer considerations and increases the probability that the consumer will choose option $A$. Generalizing this concept to multiple (vertical) attributes is straightforward. For $N > 2$, we again start with $A$ being preferred over $B$ in some $j > 0$ dimensions, while $B$ is preferred to $A$ in some others $k > 0$, such that $j+k \le N$. Then we need an alternative $C$ that will be strictly worse than $A$ in at least one dimension, while not being better in any other dimensions and being better than $B$ in some dimensions while being worse in some others. As long as these two conditions are satisfied, the attraction effect states that $C$ will result in $A$ being favored.
Generalizing this approach to multiple alternatives is somewhat more challenging. The reason for this is that, instead of one comparison ($A$ vs. $B$ in the case above), for a set of choices with $M$ alternatives, there are $\frac{M(M-1)}{2}$ potential comparisons to consider. Under real-life circumstances, it is easy to identify situations where more than one of the $\frac{M(M-1)}{2}$ relationships has the potential for an attraction effect. Besides, for any given pair of choices, we could have multiple decoy options generating attraction effect. The final complication is that option $A$ may have one set of decoy alternatives and option $B$ another set of decoy alternatives. In these contexts, it is not clear which option the attraction effect favors.
To quantify the attraction effect generated by the menu for a given alternative, I propose to calculate the number of options present in the menu that the focal alternative dominates. This is done across all vertical dimensions. Then, two alternatives present in the same menu can be compared by examining how many choices they dominate. Under such circumstances we can consider different positions option $C$ can take with respect to options $A$ and $B$. If option $C$ is neither superior (dominant) nor inferior (dominated) by any of the options ${A and B}$, or if it dominates both focal options, then it does not generate an attraction effect for $A$ or $B$. If option $C$ is dominated by both options in the focal pair, it generates an attraction effect for both of them (compared to other alternatives). In all of these cases, the location of option $C$ contributes similarly to the choice probability of both options ${A, B}$. Finally, if option $C$ is dominated by only one of the two focal alternatives (say by $A$, but not by $B$) - it generates a discriminatory attraction effect favoring option $A$ and increasing its probability of being chosen. As a result, the number of options that the current alternative dominates in a menu (appropriately normalized by the size of menu for a comparison across different choice settings) measures the (relative) extent of the attraction effect generated by the menu. For example, compare the probability of choosing option $A$ versus $E$ in figure \ref{fig:attractionZakVisualization} between two sets of menus ${A, B, C, D, E, F}$ and ${A, B, C, D", E, F}$. This probability is higher in the former situation (where $A$ dominates three alternatives, while $E$ dominates one) than in the latter case (where $A$ only dominates two alternatives, while $E$ still dominates one). Although in these cases both alternatives do have some attraction effect, the relative attraction effect of option $A$ compared to option $E$ is stronger in the former scenario. Therefore, I measure the attraction effect that favors the focal option $F$ as
$$Attraction(F)=O(Dominated),$$
where $O(Dominated)$ measures the number of alternatives in the menu that the focal option $F$ dominates. Given the measure, we expect that the higher the attraction effect in favor of the focal option, the higher the probability of choice of the focal option (\textit{ceteris paribus}).
\textbf{Compromise effect}
The compromise effect is traditionally understood and operationalized in a three-option, two-attribute (experimental) setting \citep{simonson89, dharEtAl00}. It is worth mentioning again here that these two attributes need to be vertical so that we can define universal preference relationships. Let us consider the similar starting situation of options $A$ and $B$ as in the previous subsection: $V_1(A) > V_1(B)$ and $V_2(A) < V_2(B)$. The addition of option $C$ to this menu such that $V_1(C) > V_1(A) > V_1(B)$ and $V_2(C) < V_2(A) < V_2(B)$, makes option $A$ a compromise between two extreme options $B, C$. The compromise effect
maintains that such an alteration of the menu would disproportionately benefit alternative $A$ compared to alternative $B$.
\begin{figure}[t]
\centering
\includegraphics[width=0.7\textwidth]{staticFiles/compromiseEffectZaksScatterPlot.png}
\caption[Compromise effect generalization]{Visualization of the compromise effect generalization.\\ Note: The figure represents a generalization of the compromise effect across multiple alternatives. $F$ represents a focal option. $Gr_1$ collects alternatives dominated by $F$, $Gr_2$ collects alternatives dominating $F$. Focal option represents a compromise between alternatives in $Gr_3$ and $Gr_4$.}
\label{fig:compromiseZakVisualization}
\end{figure}
To formulate the general measure of the compromise effect, let's first consider the case of multiple options $M$ in two dimensions (attributes, $N = 2$).The compromise effect calculation over multiple options is visualized on figure \ref{fig:compromiseZakVisualization} with $M = 7$ case. To quantify the extent of the compromise introduced by the focal option $F$ in the menu, I propose to split all other alternatives $M - 1$ into four groups. Let group 1 contain all alternatives for which $V_1(G_1 ) \le V_1(F)$ and $V_2(G_1 ) \le V_2(F)$. These are the alternatives dominated by the focal option. In the case of figure \ref{fig:compromiseZakVisualization}, this set contains only the option $A$. Let group 2 contain all alternatives for which $V_1(G_2 ) \ge V_1(F)$ and $V_2(G_2 ) \ge V_2(F)$. All these options dominate the focal option. This set contains option $E$ in figure \ref{fig:compromiseZakVisualization}. Clearly, the focal option cannot constitute a compromise between any pair of alternatives which is included in any of these first two groups of alternatives. Next, let group 3 contain all alternatives for which $V_1(G_3 ) > V_1(F)$ and $V_2(G_3 ) < V_2(F)$, and group 4 contain all alternatives for which $V_1(G_4 ) < V_1(F)$ and $V_2(G_4 ) > V_2(F)$. In the case of figure \ref{fig:compromiseZakVisualization}, group 3 contains options $B$, $C$ and $D$, while group 4 contains option $G$. The focal alternative can be viewed as a compromise between groups 3 and 4. For the quantification of the extent of such a compromise, I define
\begin{align}\label{}
\text{Compromise}(F) = \frac{\min(O(G_3), O(G_4))}{\max(O(G_3), O(G_4))} * (O(G_3) + O(G_4)) ,
\end{align}
where $O(G_i)$ measures the number of alternatives in the group $i$. The first multiplier (the ratio) in the measure quantifies the asymmetry across the sizes (in terms of number of alternatives) of the two groups, while the second multiplier (the sum) quantifies the joint size of two groups across which the focal option is a compromise. For the option $F$ in figure \ref{fig:compromiseZakVisualization}, this value is $Compromise(F) = \frac{1}{3} * 4 = 1.33$. If any of the two concerned groups are empty, the value is zero, corresponding to the fact that the focal alternative is at the extreme edge of one of the dimensions and therefore is not a compromise. As a result, our compromise measure will be strictly zero for options $B$, $C$, $E$ and $G$. On the other hand, the better the balance between the size of the two groups, the more valuable compromise alternative $F$ provides. So, the same measure for option $D$ in figure \ref{fig:compromiseZakVisualization} is 4. Alternative $D$ also corresponds to the compromise between 4 alternatives (like option $F$), but the comparison groups are better (in this case, perfectly) balanced. Notice that the measure also increases the number of total options in two comparison groups. Notice that the same measure for option $A$ is 2, even though (similar to option $D$) it also exhibits the correct balance between the sizes of two comparative groups. This reflects the fact that option $D$ is a compromise between larger sets of extreme alternatives \footnote{An alternative way to quantify the compromise between two sets of extreme options is to count the number of all possible pairs for which a given focal option is a compromise. This would result in $Compromise'(F) = O(G_3) * O(G_4)$. This measure behaves very similarly to the one discussed in the paper. In fact, the correlation between the two compromise measures in the dataset used in this paper is 0.825. All the results reported in the paper are qualitatively unaltered by the replacement of the compromise measure with this alternative. However, I prefer to work with the compromise measure in the paper, as it takes a more ``collective'' view of the choice process.}.
Extending the compromise measure to multiple dimensions is somewhat more challenging. The challenge relates to the fact that increasing number of dimensions (i.e., vertical dimensions) presents exponentially increasing opportunities of different ways a given option can be a compromise. The $N = 2$ case has one pair of groups to compare. However, in the case of $N = 3$, a focal alternative can be a compromise between multiple pairs of option groups. For example, option $F$ can be a compromise between two groups $Z$ and $Y$ such that all options in group $Z$ are superior to option $F$ in dimensions 1 and 2, but inferior in dimension 3, while options in group $Y$ are inferior to option $F$ in dimensions 1 and 2, but superior in dimension 3. Permutation calculus guarantees that there are three such potential comparisons. However, this is not all. Option $F$ can also be a compromise between two groups $X$ and $W$ such that all options in group $X$ are superior to option $F$ in dimension 1, but inferior in dimension 2, while options in group $W$ are inferior to option $F$ in dimension 1, but superior in dimension 2, as long as dimension 3 is constant in all options in groups $X$ and $W$, as well as $F$. The permutation calculus guarantees three additional such comparisons.
As a result, moving from 2 to 3 dimensions increases the number of potential comparisons to calculate the value of the alternative as a compromise option in the range from 1 to 6. Appendix \ref{appendix:compromiseCalculation} derives the number of comparison alternatives necessary to cover all potential ways in which a focal alternative can be a compromise as a function of the number of dimensions. However, as all the groups defined above are mutually exclusive (i.e., each alternative can only belong to one and only one of such potential comparison groups), generalization of the compromise measure in $N$ dimensions would require summation of the specific comparison group's compromise measure over all comparison groups. And thus,
\begin{align}\label{eq:compromiseEffectGeneralFormula}
\text{Compromise}(F) = \sum_j \text{Compromise}(F)_j ,
\end{align}
where $j$ runs over all possible comparison groups. Summation, instead of averaging, is used in order to reward options that constitute a compromise across multiple (many) comparison groups. Given the measure of the compromise effect, I expect that the higher the compromise effect, the higher the probability of choice of the focal alternative.
It is worth mentioning here that as both attraction and compromise measures only generalize across vertical dimensions, it is important to control for all relevant horizontal dimensions in choice models employing these measures of the two context effects.
\textbf{Similarity effect}
Operationalizing the measure of the similarity effect in three options and two vertical dimensions is straightforward \citep{roodrkerkEtAl11}. Increasing the size of the menu introduces an important challenge of defining the border between options that are similar to the focal alternative and those that are not similar to it. At the same time, unlike the previous two context effects, the theory pertinent to the similarity effect does not require dimensions to necessarily be vertical \citep{tversky1972elimination}. The sufficient condition to quantify the similarity effect requires detecting the number of other alternatives that are similar to the focal option.
Clustering, using machine learning, gives a possibility to operationalize the measure of the similarity effect across all dimensions. Several clustering algorithms have been developed that can take multidimensional lists and partition them into groups of similar objects. Clustering algorithms are unsupervised machine learning techniques that do not require explicit guidance on the definition of similarity. They use different internally consistent evaluation criteria in order to partition the input group of objects into multiple subgroups. Items belonging to the same group are judged to be similar to each other, while items belonging to two different groups are regarded dissimilar. Some algorithms, like K-means clustering \citep{lloyd82}, require additional input on (or an optimization layer for calculating) how many subgroups the user would like to detect. Others, such as Affinity Propagation \citep{freyDueck07}, automatically calculate the optimal number of detected clusters. Appendix \ref{appendix:clusteringAlgorithms} provides a summary of two popular clustering algorithms that can be used for this purpose. We argue that being able to autodetect the number of clusters is significant in terms of minimizing necessary input, as well as minimizing computational power, and use Affinity Propagation in the empirical application bellow.
As a result, I propose using a clustering algorithm (in this case, Affinity Propagation) in order to detect clusters within the menu of proposed options. Once such clusters have been identified, the size of the cluster to which the focal option belongs can be used as a straightforward measure of similarity. Hence, I measure the similarity effect as
$$\text{Similarity}(F) = O(\text{Cluster}_F) ,
$$
where, $Cluster_F$ refers to the cluster to which the focal option belongs. Given this measure of the similarity effect, we expect that the higher the similarity, the lower the choice probability of the focal alternative.
\subsection{Empirical applications}
In this section, I present two empirical applications using the generalization of three contextual measures and an estimate unifying model of context effects. Both applications come from a travel context in Europe. The first application uses a large set of observational data on airfare booking. This is a very heterogeneous dataset, and choice sets vary in terms of the number of alternatives, as well as between city pairs of origin-destination. The second application uses experimental data of stated choice in urban commutes. These data are less exciting in terms of menu variability, but it allows one to address several potential concerns with the main observational dataset. As a result, this is used as a validation exercise.
\textbf{Observational data}\label{section:additionalPreprocessingObservationalData}
Observational data I use are the same as that used in the previous chapter. The dataset is described in detail in section \ref{section:observationalDataDescription}. These data have been subject to preprocessing rules, which are also discussed in the section \ref{section:observationalDataDescription}. The descriptive statistics of the context variables are shown in the table \ref{tab:descriptivesContextOnly}. The descriptive information for other covariates can be found in the table \ref{tab:descriptiveStats}.
\begin{table}[ht]
\centering
\begin{tabular}{lrrrrr}
\hline
Variable & Count & Mean & St.Dev & Min & Max \\
\hline
Attraction & 368,723 & 19.78 & 20.33 & 0 & 98 \\
Compromise & 368,723 & 1.73 & 3.96 & 0 & 63.01 \\
Similarity & 368,723 & 11.27 & 5.88 & 1 & 77 \\
\hline
\end{tabular}
\caption[Descriptive statistis of context variables]{Descriptive statistics of context variables. \\ Note. Statistics before normalization.}
\label{tab:descriptivesContextOnly}
\end{table}
\textbf{Measurement of context effects}
The measurement of the attraction and compromise effects is pretty straightforward. I follow the methodology outlined in the previous section. For the attraction effect, I count the number of alternatives dominated by a given option within the menu. This is implemented across all four vertical attributes. The compromise effect, given by equation \ref{eq:compromiseEffectGeneralFormula}, is also measured across all four dimensions. This results in 25 pairs of comparison groups for each alternative (see equation \ref{eq:compromiseEffectDetailedCalculation} in Appendix \ref{appendix:compromiseCalculation}).
However, before proceeding to the measurement of similarity effect, it is needed to normalize the flight departure variables for the clustering algorithms. In order to identify similar alternatives within the menu, the clustering method needs a variable that allows it to measure the distance between any two departure values. This is achieved by transforming these variables into the Coordinated Universal Time format preserving dates, hours, and minutes of departure time. This way, the algorithm is able to measure the distance between any pair of alternatives in minutes. For normalization purposes, I also subtract the timestamp of the earliest flight in a menu from the departure times of every flight in that menu; thus, all times are measured as times after the earliest time.
After this transformation, I use Affinity Propagation for obtaining sets of similar options within each menu. I feed the clustering algorithm with the data on all the vertical and horizontal variables for each alternative. The algorithm returns an identifier for the group of options that comprises each of the options. Affinity propagation detects on average 7.62 clusters within the choice sets. In order to develop the measure of the similarity effect, I calculate the number of alternatives in the cluster to which the focal alternative belongs.
\textbf{Choice modeling}
To examine the context effects on choices in the airline booking data, I estimate random effects Probit models augmented by the context effect measures. These models have the crucial advantage of interpretability. Another advantage (over, for example, Logit) is the feature that Probit does not explicitly require the assumption of the independence from irrelevant alternatives. If the augmented model perfectly accounts for all context effects (IAA), this would not be a concern. However, as one cannot guarantee that human choices are not affected by any other context features (that have not yet been hypothesized and examined), having this feature is an additional advantage. I, however, also present robustness checks by fitting alternative statistical models in the Appendix \ref{appendix:LogitMixedAndFixedEffectResults} (Logit, Mixed Logit, and Fixed Effects Probit) \footnote{An additional robustness check in terms of the usage of the clustering method is also presented in the same appendix. There I use K-mean clustering (augmented with the use of Silhouette score \citep{rousseeuw1987silhouettes} to calculate the optimal number of clusters) to calculate the similarity measure. The results are robust to this alteration as well.}.
An important point to note here is the fact that the context variables incorporate menu size effects. For example, the attraction variable cannot take any value higher than 5 in a menu of size 6. However, the same variable can take the value of 49 in the 50 size menu. One way to deal with this feature would be to normalize the context variables by the size of the menu. Another alternative is to account for this feature statistically by controlling for the size of the menu in the regression equation. I opt for the latter because it guarantees higher flexibility in the empirical model structure. It also allows to account for menu size effects that could go further than context effects (for example potential choice overload). An additional advantage is that it is much simpler to interpret marginal effects of unscaled context variables.
\clearpage
\begin{sidewaystable}[ht]
\centering
\begin{tabular}{p{5cm}|*{9}{p{1.7cm}}}
\hline
Variable & Model 1 & Model 2 & Model 3 & Model 4 & Model 5 & Model 6 & Model 7 & Model 8 & Model 9 \\ \hline
Price & -0.225*** & -0.232*** & -0.200*** & -0.227*** & -0.219*** & -0.184*** & -0.186*** & -0.184*** & -0.186*** \\
& (0.006) & (0.006) & (0.008) & (0.006) & (0.006) & (0.008) & (0.007) & (0.007) & (0.007) \\
Trip duration & -0.136*** & -0.116*** & -0.087*** & -0.112*** & -0.124*** & -0.093*** & -0.092*** & -0.094*** & -0.092*** \\
& (0.009) & (0.009) & (0.010) & (0.009) & (0.009) & (0.010) & (0.010) & (0.010) & (0.010) \\
Number of flights & -0.342*** & -0.322*** & -0.325*** & -0.309*** & -0.289*** & -0.279*** & -0.289*** & -0.281*** & -0.290*** \\
& (0.008) & (0.008) & (0.008) & (0.008) & (0.009) & (0.009) & (0.009) & (0.009) & (0.009) \\
Number of airlines & -0.208*** & -0.209*** & -0.197*** & -0.210*** & -0.205*** & -0.193*** & -0.199*** & -0.193*** & -0.199*** \\
& (0.010) & (0.010) & (0.011) & (0.010) & (0.010) & (0.010) & (0.010) & (0.010) & (0.010) \\
Attraction & & & 0.003*** & & & 0.003*** & & 0.003*** & \\
& & & (0.001) & & & (<0.001) & & (<0.001) & \\
Compromise & & & & -0.038*** & & -0.034*** & -0.031*** & & \\
& & & & (0.003) & & (0.003) & (0.003) & & \\
Similarity & & & & & -0.020*** & -0.020*** & -0.031*** & -0.020*** & -0.031*** \\
& & & & & (0.001) & (0.001) & (0.002) & (0.001) & (0.002) \\
Attraction within cluster & & & & & & & 0.025*** & & 0.0245*** \\
& & & & & & & (0.002) & & (0.002) \\
Attraction outside cluster & & & & & & & 0.002*** & & 0.002*** \\
& & & & & & & (<0.001) & & (<0.001) \\
Compromise within cluster & & & & & & & & -0.149*** & -0.136*** \\
& & & & & & & & (0.020) & (0.020) \\
Compromise outside cluster & & & & & & & & -0.028*** & -0.027*** \\
& & & & & & & & (0.004) & (0.004) \\
Constant included & YES & YES & YES & YES & YES & YES & YES & YES & YES \\
Horizontal variables as controls & NO & YES & YES & YES & YES & YES & YES & YES & YES \\
Number of observations & 368723 & 368723 & 368723 & 368723 & 368723 & 368723 & 368723 & 368723 & 368723 \\
Number of choices & 6297 & 6297 & 6297 & 6297 & 6297 & 6297 & 6297 & 6297 & 6297 \\
Consistent Akaike information criterion & 49592 & 48906 & 48878 & 48764 & 48716 & 48565 & 48477 & 48555 & 48469 \\
Log likelihood & -24761 & -24356 & -24335 & -24278 & -24254 & -24165 & -24114 & -24153 & -24103 \\ \hline
\end{tabular}
\caption[Choice model estimation results]{Choice model estimation results.\\ Notes: Outputs from random effects Probit regressions. Standard errors in parentheses. Statistical significance levels: $*** p<0.01$, $** p<0.05$, $* p<0.1$.}
\label{tab:mainResultsRandomProbitModel19AmadeusData}
\end{sidewaystable}
\clearpage
Before going to the estimation, one needs to transform the departure-time variables into outbound and inbound flight pairs. The transformation that was performed for the clustering exercise cannot be used directly because it estimated coefficients that are not interpretable. To make this information as tractable as possible, I generate a set of variables. First, a day-of-the-week variable is generated for the outbound flight. Second, a variable is generated that measures the duration of stay at the destination \footnote{For the regression analysis, similar to other numeric variables, in order to eliminate any scale effects, I perform a z-score transformation of duration of stay variable.}. These two variables together describe the inbound and outbound flight timing characteristics at the level of the day. However, consumer preferences can be defined on a smaller scale. Therefore, I also generate two variables that describe the exact time of the day of the outbound and inbound flights. These variables, $t_out,t_i \in [0;1)$, are measured as a fraction of a day such that $t_i = 0$ corresponds to midnight, while $t_i=0.5$ corresponds to midday. I further apply a cosine transformation to these variables, i.e., $\cos(t_i ) = 2 \pi t_i$. This confines the departure time variable to the interval [-1;1], and ensures the smooth transition in departure times at midnight. These transformations result in a total of four variables describing departure timestamps for outbound and inbound flight pair - the horizontal attribute of the alternative.
I estimate a sequence of 9 models and present the results in table \ref{tab:mainResultsRandomProbitModel19AmadeusData}. I estimate these models by using random-effects Probit regressions with robust standard errors. First, I start out by fitting two simple baseline models of consumer choice. Model 1 is the simplest estimation, which includes only the four vertical attributes as independent variables. Model 2 further extends this model by adding four horizontal attributes. In both cases, with or without horizontal attribute controls, all vertical variables generate meaningful results. Consumers clearly have preferences for shorter, cheaper flights with fewer layovers and airline changes. Travelers also seem to have preferences for the outbound flight during the day and for the inbound flight during the night (recall that the cosine transform variables reach a maximum at midnight and a minimum at midday).
\begin{table}[ht]
\centering
\renewcommand{\arraystretch}{1.3}
\setlength{\tabcolsep}{0.4em}
\begin{tabular}{p{3cm}*{7}{p{1.5cm}}}
\hline
Variable & Model 3 & Model 4 & Model 5 & Model 6 & Model 7 & Model 8 & Model 9 \\
\hline
Attraction & 0.0187 & & 0.0170 & & 0.0173 & & \\
Compromise & & -0.1900 & & -0.1646 & & -0.1509 & \\
Similarity & & & -0.0747 & -0.0756 & -0.1222 & -0.0740 & -0.1207 \\
Attraction within cluster & & & & & 0.1042 & & 0.1040 \\
Attraction outside cluster & & & & & 0.0133 & & 0.0135 \\
Compromise within cluster & & & & & & -0.6039 & -0.5444 \\
Compromise outside cluster & & & & & & -0.1377 & -0.1323 \\
\hline
\end{tabular}
\caption[Marginal effects of choice model on observational data]{Average marginal effects for relevant models.\\ Note. Average marginal effects implied by various models. All p-values were significant at $p<0.001$ level.}
\label{tab:marginalEffectsAmadeusModel39}
\end{table}
To further extend model 2, three models (3 through 5) that each incorporate one of the context effects, and one model that incorporates all three context effects at once (model 6) were estimated. Table \ref{tab:mainResultsRandomProbitModel19AmadeusData} indicates the consistency between the coefficient estimates of model 6 and those of models 3-5. This set of models also allows one to evaluate the effect of the three context effects on consumer choice. In line with the theory, the presence of attraction and similarity effects is observed. That is, if the attraction measure increases for a given option, this increases the likelihood that the option is chosen. On the other hand, when the similarity measure increases, it decreases the probability that the option is chosen. Both of these effects are statistically highly significant and are in the hypothesized direction. To better understand the economic significance of the estimated effects, table \ref{tab:marginalEffectsAmadeusModel39} presents (average) marginal effects of the relevant models. From table \ref{tab:marginalEffectsAmadeusModel39} one can read that if the attraction measure increases by one unit (that is, having one more option dominated by the focal alternative, \textit{ceteris paribus}) the likelihood of an option being chosen goes up by about 0.02 percentage points on average. On the other hand, if the similarity measure increases by one unit, the likelihood that a given option is chosen goes down by about 0.08 percentage points on average.
Tables \ref{tab:mainResultsRandomProbitModel19AmadeusData} and \ref{tab:marginalEffectsAmadeusModel39}, however, also indicate the existence of a reverse compromise effect. The compromise effect posits that if an option represents a compromise between extreme alternatives, it will have a higher likelihood of being chosen. On the contrary, our results indicate that increasing our compromise measure decreases the likelihood of an option being chosen. This effect is again statistically and economically significant. From this we can conclude that in the context of airfare choice, consumers prefer extreme options to those that represent compromise. This implies that the preferences of individual consumers are strongly anchored to one of the four vertical attributes. For example, if a traveler attaches particular importance to price, she will be reluctant to trade away an option that is cheap for increases in the attractiveness in any other (vertical) dimension. This, in fact, is rather understandable given the context of current empirical exercise: The two largest groups of air travelers are holidaymakers, who are price-sensitive and do not readily trade away price advantage for shorter travel time, and business travelers, who are time-sensitive and do not trade away flight duration for a decrease in price.
Next, I investigate the interaction between several context effects. Previous literature has hypothesized and demonstrated the interaction between attraction and similarity effects in laboratory environments \citep{huberEtAl82, huberPuto83, roodrkerkEtAl11}. The interaction between similarity and compromise effects has not been studied in literature; however, one can consider that if the similarity effect efficiently identifies comparable alternatives that could constitute a consideration set of the consumer, the compromise effect, which considers options outside the consideration set, will not constitute an adequate guide for consumer behavior. Given that the similarity measure hinges on identifying clusters of similar options, the interplay between the similarity effect and the other two context effects is rather straightforward to study.
For this, four additional measures are calculated for each option that decomposes attraction and compromise effects along the cluster lines identified by the similarity measure. More precisely, the attraction and compromise measure for a given option is calculated: 1) by taking only the alternatives that belong to the same cluster to which this particular option belongs, and 2) by considering only the alternatives that do not belong to the same cluster. This way, one can get a measure of attraction and compromise effects of an alternative within the cluster (i.e., among comparable alternatives, or within the consideration set) and outside the cluster (i.e., among relatively noncomparable alternatives, or outside the consideration set).
In models 7 through 9 I study the comparative effects of pairs of these effects. Model 7 decomposes the attraction effect of model 6 into two parts (inside and outside the cluster). Model 8 decomposes the compromise effect along the same lines, and model 9 estimates the model that includes both decompositions simultaneously. The results are again consistent and meaningful. Models 7 and 9 imply that attraction effect within the cluster, i.e., among comparable alternatives, has a much stronger impact on the purchase likelihood than that of the impact of the measure calculated based on nonsimilar alternatives. Table \ref{tab:marginalEffectsAmadeusModel39} indicates a difference in the size of the order of magnitude. Similarly, as indicated by models 8 and 9, being a compromise among comparable alternatives has a much higher detrimental effect on purchase likelihood than being a compromise among remote alternatives. The p values for all tests of coefficient pair equality (i.e., estimated coefficient for attraction within cluster being equal to that of the attraction outside cluster, and coefficient for compromise within the cluster being equal to the compromise coefficient outside the cluster) are below 0.001, indicating that cluster-based measure of similarity may be an efficient indicator of the consumer's consideration set.
\clearpage
\begin{sidewaystable}[ht]
\centering
\begin{tabular}{p{5cm}*{9}{p{1.7cm}}}
\hline
Variable & Model 1 & Model 2 & Model 3 & Model 4 & Model 5 & Model 6 & Model 7 & Model 8 & Model 9 \\ \hline
\addlinespace
Price & -0.249*** & -0.259*** & -0.253*** & -0.257*** & -0.249*** & -0.238*** & -0.245*** & -0.238*** & -0.245*** \\
& (0.008) & (0.009) & (0.013) & (0.009) & (0.009) & (0.013) & (0.013) & (0.013) & (0.013) \\
Trip duration & -0.167*** & -0.143*** & -0.137*** & -0.143*** & -0.147*** & -0.139*** & -0.144*** & -0.138*** & -0.143*** \\
& (0.013) & (0.013) & (0.016) & (0.013) & (0.013) & (0.015) & (0.015) & (0.015) & (0.015) \\
Number of flights & -0.414*** & -0.392*** & -0.393*** & -0.386*** & -0.373*** & -0.369*** & -0.373*** & -0.371*** & -0.375*** \\
& (0.014) & (0.014) & (0.014) & (0.014) & (0.014) & (0.015) & (0.015) & (0.015) & (0.015) \\
Number of airlines & -0.161*** & -0.163*** & -0.160*** & -0.164*** & -0.164*** & -0.162*** & -0.165*** & -0.161*** & -0.164*** \\
& (0.015) & (0.015) & (0.015) & (0.015) & (0.015) & (0.015) & (0.015) & (0.015) & (0.015) \\
Attraction & & & 0.002 & & & 0.002 & & 0.003 & \\
& & & (0.003) & & & (0.003) & & (0.003) & \\
Compromise & & & & -0.066*** & & -0.063*** & -0.064*** & & \\
& & & & (0.013) & & (0.013) & (0.013) & & \\
Similarity & & & & & -0.023*** & -0.023*** & -0.029*** & -0.022*** & -0.029*** \\
& & & & & (0.003) & (0.003) & (0.004) & (0.003) & (0.004) \\
Attraction within cluster & & & & & & & 0.016*** & & 0.016*** \\
& & & & & & & (0.005) & & (0.005) \\
Attraction outside cluster & & & & & & & -0.001 & & -0.001 \\
& & & & & & & (0.003) & & (0.003) \\
Compromise within cluster & & & & & & & & -0.182*** & -0.180*** \\
& & & & & & & & (0.048) & (0.048) \\
Comrpmise outside cluster & & & & & & & & -0.029 & -0.031* \\
& & & & & & & & (0.018) & (0.018) \\
Constant included & YES & YES & YES & YES & YES & YES & YES & YES & YES \\
Horizontal variables as controls & NO & YES & YES & YES & YES & YES & YES & YES & YES \\
Number of observations & 79080 & 79080 & 79080 & 79080 & 79080 & 79080 & 79080 & 79080 & 79080 \\
Number of choices & 3954 & 3954 & 3954 & 3954 & 3954 & 3954 & 3954 & 3954 & 3954 \\
Consistent Akaike information criterion & 25568 & 25124 & 25134 & 25106 & 25088 & 25085 & 25087 & 25101 & 25104 \\
Log likelihood & -12759 & -12482 & -12481 & -12467 & -12458 & -12444 & -12439 & -12446 & -12441 \\
\hline
\end{tabular}
\caption[Choice model results for the reduced dataset]{Choice model estimation results from the reduced dataset.\\ Notes: Outputs from random effects Probit regressions. Standard errors in parentheses. Statistical significance levels: $*** p<0.01$, $** p<0.05$, $* p<0.1$.}
\label{tab:reducedResultsRandomProbitModel19AmadeusData}
\end{sidewaystable}
\clearpage
An important drawback of the dataset is the feature that there is no way to know which available options on the market reached the eyeballs of the consumer. One way of thinking about this problem is to consider the most likely way such menus are delivered to decision-makers. Most online booking sites and flight aggregators use specific and proprietary algorithms to rank available menus at the search point. These rankings decide which options are shown to the customer. Although the information on specific ranking algorithms is not public, we know that the attribute that usually plays the most important role is the price. Given the robust findings that price negatively affects choice probability, the best guess for a simple ranking mechanism that would capture a wide variety of sorting mechanisms would be options sorted in decreasing order with respect to price. Assuming that each user was reached by the same number of options, we could construct a reduced dataset for a sensitivity check. For this exercise, I construct a dataset that only contains menus with more than 20 options, and only retain 20 cheapest alternatives per menu. There are also cases where the chosen option is not part of the set of 20 cheapest alternatives in the menu \footnote{20 alternatives are chosen so that there are enough entries to have variance in key variables. Results are robust to different menu sizes, with the characteristics that, as I reduce menu size, more effects seem to lose significance}. I also eliminate these choice cases from the reduced dataset. This leaves me with about four thousand choice cases.
\clearpage
\begin{table}[!ht]
\centering
\renewcommand{\arraystretch}{1.1}
\setlength{\tabcolsep}{0.3em}
\begin{tabular}{>{\fontsize{10pt}{11pt}\selectfont}p{3cm}>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l}
\hline
Panel A & Model 3 & Model 4 & Model 5 & Model 6 & Model 7 & Model 8 & Model 9 \\ \hline
Attraction & 0.0138 & & 0.0191 & & 0.0010 & & \\
& [0.512] & & [0.362] & & [0.317] & & \\
Compromise & & -0.5507 & & -0.5239 & & -0.5335 & \\
& & [$<0.001$] & & [$<0.001$] & & [$<0.001$] & \\
Similarity & & & -0.1916 & -0.1890 & -0.2425 & -0.1859 & -0.2375 \\
& & & [$<0.001$] & [$<0.001$] & [$<0.001$] & [$<0.001$] & [$<0.001$] \\
\renewcommand{\arraystretch}{1.}
Attraction within cluster & & & & & 0.1357 & & 0.1329 \\
& & & & & [0.001] & & [0.001] \\
Attraction outside cluster & & & & & -0.0111 & & -0.0080 \\
& & & & & [0.628] & & [0.728] \\
\renewcommand{\arraystretch}{1.}
Compromise within cluster & & & & & & -1.5127 & -1.4961 \\
& & & & & & [$<0.001$] & [$<0.001$] \\
Comrpmise outside cluster & & & & & & -0.2374 & -0.2566 \\
& & & & & & [0.107] & [0.083] \\
\end{tabular}
\vspace{10pt}
\begin{tabular}{>{\fontsize{10pt}{11pt}\selectfont}p{3cm}>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l>{\fontsize{11pt}{13pt}\selectfont}l}
\hline
Panel B & Model 3 & Model 4 & Model 5 & Model 6 & Model 7 & Model 8 & Model 9 \\ \hline
Attraction & 0.0359 & & 0.0291 & & 0.0096 & & \\
& [$<0.001$] & & [$<0.001$] & & [$<0.001$] & & \\
Compromise & & -0.4150 & & -0.3482 & & -0.3112 & \\
& & [$<0.001$] & & [$<0.001$] & & [$<0.001$] & \\
Similarity & & & -0.2037 & -0.2082 & -0.3084 & -0.2038 & -0.3034 \\
& & & [$<0.001$] & [$<0.001$] & [$<0.001$] & [$<0.001$] & [$<0.001$] \\
\renewcommand{\arraystretch}{1.}
Attraction within cluster & & & & & 0.2454 & & 0.2436 \\