Informational Asymmetries and Observational Learning in Search
Informational Asymmetries and Observational Learning in Search
30.90 30.90
All gures in US Dollars.
Optimal refers to the earnings obtained by
setting 81 at all periods.
Note that subjects were guaranteed $20
(so the few subjects who obtained less than
that received $20 in the end of the experiment).
subjects understand the setup, it seems from the answers that some of them had a very
weak idea of what they were supposed to do during the experiment. For comparing the
means and medians of reservation prices across the two experimental conditions (and only
for that), these subjects introduce biases and increase the variance, so I chose to drop them
from the sample for this specic analysis.
18
The reason is that a random draw from the
distribution has a mean of about 50, while the majority of the observed reservation prices
are in the range of 70 to 85. It turned out that subjects in the experimental treatment were
more likely to get confused and to think that they do not understand the instructions. This is
due to the additional observed draws, which had no practical effect on the outcome (beyond
their effect through learning). Therefore, comparing means with these random subjects
in the sample is likely to bias the results downward in this treatment, when compared to
the control group, and hence create problems in the interpretation of the simple comparison
of means across groups. In all other statistical analyses, such random subjects can be
treated as a white noise, so I keep them in the sample, and they have little effect on the
results.
Table 2 and Figure 1 show the average and median reservation price for each group,
as it evolved over the 100 periods of the experiment. As in previous experiments, which
are discussed in Section 1, it is shown that, on average, subjects search too little. For both
groups in all periods the average reservation price is below the optimal reservation price
of 81, making subjects stop searching too early. The results conrm the hypothesis that
subjects who may experience symmetric regret by observing higher prices after they have
stopped searching end up searching longer and setting higher reservation prices, closer to
the optimal one. Once the random subjects are dropped, there is much less noise in the
averages, so I am able to obtain statistically signicant differences between the two groups.
There are a fewadditional interesting results that should be mentioned. First, even subjects
in the experimental treatment, who search more, still search too little by setting average
reservation prices that are signicantly lower than the optimal one, 81. This can be explained
248 EINAV
Table 2. Averages of reservation prices.
Control Treatment
All subjects
First period 63.3 (2.76) 63.1 (3.18)
All periods 68.7 (2.33) 69.4 (2.26)
Last 50 periods 70.8 (2.41) 70.8 (2.32)
Last 25 periods 71.3 (2.37) 70.8 (2.43)
Last 10 periods 72.0 (2.46) 71.3 (2.43)
Last period 72.7 (3.02) 70.8 (2.93)
No. of subjects 35 40
All Non-Random subjects
First period 66.5 (2.17) 66.7 (3.38)
All periods 72.6 (1.83) 75.2 (1.45)
No. of subjects 30 30
0.146
0.188
0.158
0.172
0.092
0.026
(0.35) (0.10) (2.46) (1.59)
Constant 0.767
0.795
0.760
0.795
0.806
0.851
0.0031
0.0035
0.0031
0.0043
0.0036
,
implies signicance at the 1 and 5% condence level, respectively.
The regressions are run for each experimental treatment separately. The control group and the pseudo control
group use the same data. The only difference is that in the pseudo control an upward regret regressor is
computed, as if subjects could observe post-purchase information. The coefcient on Downward regret
suggests the existence of observational learning. The most important observation, however, is the difference
in the magnitude and signicance in the coefcient on Upward regret in the pseudo control group and the
treatment group regressions. Its signicance only in the treatment group regression implies that it is only the
actual experience of regret that affects behavior. The rest of the regressors are controls. They all obtain the
predicted signs. The regressions include all subjects.
To gain some intuition for the magnitude of the effect, a subject who would otherwise not
change her reservation price will decrease (increase) it by about 0.7 (0.15) after experiencing
downward (upward) regret.
22
This magnitude is quite large considering the fact that, on
average, each subject increases her reservation price by about 0.08 per period. Loosely
speaking, a downward regret offsets about nine periods without regret, while an upward
regret is equivalent to three periods without regret.
The control variables all have the predicted signs. Higher reservation prices make it more
likely to adjust downwards, and higher number of bids have similar effect. Earnings do
not seem to have a signicant effect on the strategy adjustment. These coefcients on the
control variables illustrate why it is important to use them. Downward regret is more likely
after a high reservation price and high number of bids. Thus, if we did not control for these
variables in the regression, we would obtain stronger estimated effect of downward regret,
which would only be driven by statistical properties of regret. After using these variables
252 EINAV
Table 4. Multinomial logit regressions.
Control Pseudo control Treatment
Downward Upward Downward Upward Downward Upward
Dependent variable change change change change change change
Downward regret
t 1
4.34
1.46 4.98
1.56 4.53
1.28
(3.98) (1.05) (4.27) (1.18) (4.71) (0.60)
Upward regret
t 1
1.10 1.15 1.02 1.42
1.12 2.38
1.20 1.85
1.58
(2.00) (0.36) (2.23) (0.56) (2.00) (1.20)
Upward regret
t 1
1.05 1.06 0.95 1.24
1.005 1.051
1.006 1.054
0.948
0.988
,
implies signicance at the 1% and 5% condence level, respectively.
In all regressions, the base category is no change in the reservation price.
The regressions reported below are similar to those reported in Table 3. Rather than using the magnitude of the
change in reservation price, these regressions replicate the same pattern of Table 3, when the dependent variable is
the discrete choice of whether to change the reservation price, and if so in which direction. The top panel runs the
regressions with no controls, while the bottom panel adds several controls. The results are qualitatively similar.
The specication is of a multinomial logit, where the reported gures are of odds ratios.
As in Table 3, the Downward regret works as predicted: it increases the probability of changing the reservation
price downwards, but has no signicant effect on a change upwards. As before, the most interesting contrast is
between the effect of Upward regret in the pseudo control group, where it is not signicant, vis-` a-vis its signicant
effect in the treatment group. Again, this implies that it is only the actual experience of regret that affects behavior.
INFORMATIONAL ASYMMETRIES AND OBSERVATIONAL LEARNING IN SEARCH 253
as controls, it seems much more likely that the remaining effect of regret is indeed due to
learning.
Table 4 repeats the exercise, using multinomial logit regression. Rather than using the
dependent variable as a continuous measure, Table 4 only analyzes the direction of strategy
adjustment, tofocus onthe directional learningidea. Thus, the dependent variable is discrete:
upward adjustment of the reservation price (r p
t
r p
t 1
> 0), no adjustment (r p
t
r p
t 1
= 0), or a downward adjustment (r p
t
r p
t 1
< 0). The results are similar to
those presented in Table 3. In all groups, observing a downward regret makes it more likely
that subjects adjust their reservation price downwards, while observing an upward regret
makes it more likely that subjects in the treatment group adjust their reservation prices
upwards. In sharp contrast, (hypothetical) upward regret has no signicant effect in the
pseudo control group, suggesting that it is only the actual observation of regret that affects
behavior.
Again, the magnitude of the effect is quite substantial. On average, each subject ad-
justs her reservation price downwards about 28 percent of the times. The odds ratio of
4.34 in the top panel of Table 4 implies that after experiencing a downward regret, this
probability is much higher, and is about 64 percent. Similarly, the odds ratio of 1.42 for
upward regret (the last column of Table 4) implies that the probability of adjusting the
reservation price upwards become about 40 percent, compared to an average probability of
32 percent.
3.3. Summary
The results replicate well previous results that found that subjects search too little. In this
setup it is shown that subjects in both groups set on average a reservation price which is
signicantly lower than the optimal reservation price of 81. In addition, I showthat subjects
who are able to obtain more symmetric information structure tend to search better, closer
to optimal, albeit still too little. Finally, the results document quite strongly the effect of
actual observation of regret, in both directions, upward as well as downward. The analysis
suggests that observing regret can be thought of as a feedback to the subjects, which enhance
their learning. Once the feedback is objective, symmetric in both directions, subjects react
accordingly and better converge towards the optimal reservation price.
4. Conclusions and further research
Previous experimental search papers have consistently found that subjects in experiments
search too little. In this paper I test for a possible explanation for this robust result, and nd
evidence in its support. Subjects, who can evaluate their strategies more symmetrically by
observing post-purchase information, search more intensively.
In addition, the results show the importance of observational learning in the form of
observational regret as a learning-enhancing device. Using the individual search problem it
is shown that actual observation of regret leads to a directed adjustment of played strategies,
while probabilistic regret has no statistically signicant effect. It is evident that subjects
254 EINAV
who are able to actually observe regret end up searching in a more efcient way by setting
reservation prices that are closer to the optimal one.
We can try to extend these results to strategic interactions. For example, one could look
at a sequential two-player game, using both the action method and the strategy method.
23
In the strategy method subjects are informed about the full strategy of their opponents, so
they can calculate their unrealized payoffs. In this sense, the strategy method is similar to
the experimental treatment and allows for observational regret, while the action method is
similar to the control group, in which subjects may only speculate what could have been
their earnings had they chosen a different action. The results from this paper suggest that
learning and convergence to equilibrium would be faster when the strategy method is used.
The observation of post-purchase information is not only a mechanical device to test the
explanation for too little search. It may also be a closer description of certain markets in the
real world. It is quite often that a consumer buys a certain good and later observes the same
good being sold at a lower price. This new information can be obtained from a friend who
bought the same good, fromadvertisement, or fromvisiting the same retail store for another
reason. Moreover, it seems plausible that consumers who bought a certain product recently
are more likely to internalize newinformation about it. Consequently, the results obtained in
this paper suggest an important empirical implication: search is going to be closer to optimal
in markets in which consumers are more likely to obtain post-purchase information. This
prediction may be testable. This can be done by comparing search behavior in markets
for similar products that differ in their information structure.
24
In markets where search is
repeated but post-purchase information is not frequently available, a more efcient search
can be achieved by designing search strategies that will make information ow in a more
symmetric manner, making post-purchase information more readily available.
Appendix A: Experiment instructions
You are participating in an experiment about individual decision making. Hence, nothing
that you do can affect the other participants, and nothing they do will affect you. In addition,
the setup is not completely identical for all of you (and a lot of it is random), so trying to
learn from others choices does not make much sense.
The experiment consists of 100 identical periods. At the start of each period you will get
an articial object for 50 cents. You will have to sell this object to the computer. Each period,
before the computer makes bids for the object, you will have to announce the minimum
price at which you are willing to sell the object. Then, the computer will make bids for the
object, until there is a bid greater than or equal to the minimum price you have announced.
The selling price will be this last computer bid.
You can think of this setup as a search for a new job. In this analogy, each computer bid is
a wage offer for the same job, and your search criterion denes a minimum wage at which
you are willing to take the job.
There are three things that are very important to know:
1. Every bid costs 2 tokens. Using the wage analogy, these costs can stand for the time you
spend looking for a job. Hence, your costs in a period are two tokens times the number
of bids in that period, plus the 50 tokens that you had to pay to get the object.
INFORMATIONAL ASYMMETRIES AND OBSERVATIONAL LEARNING IN SEARCH 255
2. Your earnings in a period are the price at which you make the sale (the rst bid that is
greater than or equal to the minimum price announced at the beginning of that period),
minus the costs in the same period (i.e. Earning = Sale Price 2 Number of Bids
50). Note that it is possible to have negative earnings in a period. If that happens, the
loss will be deducted from the earnings in other periods.
3. All bids are integers between 1 to 100 tokens (1 and 100 included). Each bid is indepen-
dent of previous bids and of the minimum price you announced. The bids have already
been generated randomly, prior to the experiment, so any decision that you make during
the experiment is not going to affect the bids you will get later on. In page 4 of the
instructions you can take a look at a histogram of 100,000 bids that were generated by
the same process that generated the bids you will get during the experiment. The height
of each bar is the probability of getting the corresponding bid.
At the beginning of each period you will start by having a sample of computer bids
(arranged in a circle shape) which were generated by the same process as the real ones.
These sample bids are independent of the real ones, and are provided only in order to
give you a better understanding of the distribution of bids. There are no tricksif you
feel that you understand the distribution from which the bids are generated, it is totally
sensible to ignore these sample bids. The sample bids are provided only in order to
give you a better feel of the distribution of bids. After you have looked at the sample
bids, you can choose your minimum selling price for that period and bidding will take
place.
At the end of each period, some of you will get the opportunity to observe the next three
computer bids (you get it the same way you received the previous bids). These additional
bids cannot be used for sale. They are provided in order to let you see how more bids would
have looked like in case the minimum selling price you had set would not have made you
stop where you stopped, so bidding would have continued.
The above will go on for 100 periods. Your total earnings will be the sum of the earnings
in each period.
Note that by choosing your minimum selling price you choose both the expected price
you will receive and the expected number of bids you will pay for. Therefore, by choosing
your minimum selling price thoughtfully, you can increase your earnings.
Appendix B: Computer program description
You begin the experiment by having one of the instructors activating the program. This
will make the welcome window disappear, and you will be able to start the experiment by
clicking on the Next Period button, which appears in the right bottomcorner of the screen
(please raise your hand if you do not see this button).
After clicking the Next Period button, the minimum price window (shown below) will
pop up. The circle that appears in the lower part of the window is the sample bids that are
discussed in the instructions. You can then choose the minimum price for that period from
the ten menus in the top of the window. After you made your choice, click on the OK
button, which appears in the top right corner of the window.
256 EINAV
Now bidding will take place, as can be seen in the screen below. You can get new bids by
clicking the Next Bid button (left bottom corner) for each new bid. When bids take place,
they appear on the left side of your screen. There are three columnsthe bid number, the bid,
and the amount you would have earned if the bid was the last one for the period (calculated
by: bid 2 bid number 50). On the right side of your screen, your performance in
past periods is recordedthe period, the minimum price you announced for that period,
the number of bids you got, the sale price and the total earnings you had for the period. At
the top of the screen, you can see also your subject number, what period it is, the minimum
price you announced for the period, and your total and average earnings so far. Once a
period ends, click the Next Period to go on.
INFORMATIONAL ASYMMETRIES AND OBSERVATIONAL LEARNING IN SEARCH 257
After you are done, the last screen will pop up. Fill in your name and social security
number and click the Print and Quit button. We will collect the printouts from the printer,
prepare your payment and call you to pick it up. The receipt is the only place in which your
name and social security number are recorded. Other than that, only the random number
assigned to you is recorded.
If there are no further questions, you are ready to start. The instructors will approach
your desk and activate the program.
Enjoy!
Acknowledgments
This paper is a revised version of chapter 3 of my 2002 Harvard University disserta-
tion. I am particularly grateful to Al Roth for continuous guidance, helpful discussions,
and nancial support of the pilot studies. I thank Drew Fudenberg, David Laibson, an
anonymous referee, and the editor for helpful comments and suggestions. Estelle Cantilon,
Guillermo Caruana, Yair Eilat, Muriel Niederle, Robert Slonim, Manuel Trajtenberg, Leeat
Yariv, Elena Zoido, and seminar participants at Harvard University and at the 1999 Econo-
metrics in Tel-Aviv provided suggestions and comments on earlier drafts. I thank Joep
Sonnemans for sharing with me his experimental data. Financial support by the Russell Sage
Foundation (RSF Project #98-00-07) is gratefully acknowledged. All remaining errors are
my own.
Notes
1. See also the introduction to Erev and Roth (1998) and the references therein.
2. See Stigler (1961) and Kohn and Shavell (1974).
3. See, for example, Schotter and Brounstein (1981), Hey (1982, 1987), Kogut (1990), Cox and Oaxaca (1996),
Moon and Martin (1996), Butler and Loomes (1997), and Sonnemans (1998).
4. The control group is identical to the design of Sonnemans (1998).
5. Kohn and Shavell (1974) and Telser (1973) discuss the implications of extensions to this setup.
6. See the references listed in endnote 3.
7. More formally, suppose a search strategy yields expected payoff of e with payoff variance of v. Now, instead
of running one such search problem, we run N i.i.d identical problems, such that the sum of the expected
payoffs remains the same (so the size of each problem is N
1
times the original problem). Thus, it is easy
to see that the variance of the same strategy in the repeated problem is N(vN
2
) = vN
1
.
8. Although not in the scope of this paper, a possible way to directly test the power of the risk-aversion explanation
is by using binary lottery techniques. These techniques use offers as probabilities of winning a binary lottery,
instead of direct monetary payments. See Roth and Malouf (1979), Berg et al. (1986), and Kagel and Roth
(1995) for more details. As mentioned before, Rabin (2000) suggests that this may not be necessary.
9. Sonnemans (1998) mentions this explanation, and documents that the behavior of subjects after experiencing
regret is different fromtheir behavior in other periods. He does not, however, formally analyzes this hypothesis,
as it is not the focus of his paper.
10. Such ideas are consistent with directional learning theory (Selten and Buchta, 1994), and, to a lesser extent,
with regret matching (Hart and Mas-Colell, 2000).
11. The xed cost per period is not a choice and has no effect on the optimal strategy. It is useful because it
increases the salience of net payoffs and directs subjects towards setting higher reservation prices, speeding
258 EINAV
up learning and convergence. Pilot studies without xed cost have indeed shown that the xed cost is quite
useful for this purpose.
12. Even with full information, one may wonder how well subjects understand the information they are given. In
the experiment described below I try to help subjects understand what the distribution is by showing them a
random sample from it.
13. See, for example, Roth and Erev (1995) and the references they cite.
14. Sonnemans (1998) had 19 subjects who played for 45 periods. They started (period 1) with an average
reservation price of 67.5 and ended (period 45) with an average reservation price of 77.9, which is still below
the optimal reservation price of 81.
15. As emphasized in the previous section, it is important to distinguish between the concept of observational
regret, which is dened above, and the more common use of (anticipated) regret, which enters into the utility
function. Throughout the paper, regret plays a role in enhancing learning, but not through any direct effect on
utility.
16. The complete instructions can be found in Appendix A. Appendix B provides a description of the computer
program, which was also given to the subjects before the beginning of the experiment. These, and a short
pre-experiment exercise, were read aloud to subjects prior to the experiment.
17. The identical draws eliminate some of the noise that can arise fromthe randomdifferences in the draws among
subjects. It is also true, however, that identical draws may cause some bias in the results (if, for example, the
very rst rounds are crucial for learning). With a relatively small sample, the trade off is best resolved by
having identical draws. With large enough sample, however, it would be denitely better to allow different
draws for each individual.
18. The criterion for dropping subjects was very conservative. I drop any subject who chose reservation price
lower than 50, for at least 5 times during the last 40 (out of 100) periods. In general, a reservation price lower
than 50 seems somewhat unreasonable given the xed cost of 50. I look only at the last 40 periods because it
makes sense that subjects may not realize immediately that less than 50 is a bad strategy, but only later on.
It goes without saying that the criterion was chosen without looking at its affect on the results. This criterion
led to dropping as many as 15 subjects (5 (out of 35) from the control group and 10 (out of 40) from the
treatment group). As could be predicted, more subjects got confused (and were dropped) in the somewhat
more confusing condition, namely the treatment group, in which they get to observe post-purchase bids,
which have no effect on the outcome.
19. At the very rst period subjects from either group do not know which condition they are in. Only in the end
of the rst period, once they observe or do not observe post-purchase prices, they realize to which one of the
groups they were randomly assigned.
20. These three additional prices are known to us, because the sequences of bids were previously generated. In
fact, for our purposes we could also generate new i.i.d draws from the same distribution, and think of them as
the three potential future bids.
21. Simpler descriptive statistics also lead to similar conclusions, and are reported in Einav (2002). The qualitative
results are not sensitive to the particular specication or to whether the random subjects are included or not.
Tables 3 and 4 provide results with all subjects included.
22. This is calculated by inverting the logistic transformation. For an estimated coefcient of a, its effect on the
dependent variable, which would otherwise be zero, is given by log(0.5 +a) log(0.5 a).
23. Take the ultimatum game for example, in the action method the responder would only accept or reject the
offer he actually received, while in the strategy method the responder would have to specify a fully contingent
strategy of whether he would have accepted or rejected any potential offer.
24. For example, we can think of two markets for the same product, a regular one and an online one. Alternatively,
one can think of two U.S. states that differ in the law regarding advertisement of tobacco products. Post-
purchase information is likely to be better in states that allow for such advertisements. It must be noted,
however, that a clean test of the prediction would require the two markets to differ in the post-purchase
information, but to be similar in the pre-purchase information, as well as in the mix of consumers who
participate in them. Such examples may be hard to nd in real world, which is one of the reasons that make
experiments so useful.
INFORMATIONAL ASYMMETRIES AND OBSERVATIONAL LEARNING IN SEARCH 259
References
Berg, J. E., L. A. Daley, J. W. Dickhaut, and J. R. OBrien. (1986). Controlling Preferences for Lotteries on Units
of Experimental Exchange, Quarterly Journal of Economics 101, 281306.
Butler, D. and G. Loomes. (1997). Quasi-Rational Search under Incomplete Information: Some Evidence from
Experiments, The Manchester School 65, 127144.
Cox, J. C. and R. L. Oaxaca. (1996). Testing Job Search Models: The Laboratory Approach, Research in Labor
Economics 15, 171207.
Einav, L. (2002). Essays in Industrial Organization, Ph.D. Dissertation, Harvard University, Chapter 3.
Erev, I. and A. E. Roth. (1998). Predicting How People Play Games: Reinforcement Learning in Experimental
Games with Unique, Mixed Strategy Equilibria, American Economic Review 88, 848881.
Goeree, J. K., C. A. Holt, and T. R. Palfrey. (2002). Quantal Response Equilibrium and Overbidding in Private-
Value Auctions, Journal of Economic Theory 104, 247272.
Hart, S. and A. Mas-Colell. (2000). A Simple Adaptive Procedure Leading to Correlated Equilibrium, Econo-
metrica 68, 11271150.
Hey, J. D. (1982). Search for Rules for Search, Journal of Economic Behavior and Organization 3, 6581.
Hey, J. D. (1987). Still Searching, Journal of Economic Behavior and Organization 8, 137144.
Kagel, J. H. and A. E. Roth. (1995). The Handbook of Experimental Economics. Princeton University Press.
Kogut, C. A. (1990). Consumer Search Behavior and Sunk Costs, Journal of Economic Behavior and Organi-
zation 14, 381392.
Kohn, M. G. and S. M. Shavell (1974). The Theory of Search, Journal of Economic Theory 9, 93123.
Moon, P. and A. Martin. (1996). The Search for Consistency in Economic Search, Journal of Economic Behavior
and Organization 29, 311321.
Rabin, M. (2000). Risk Aversion and Expected-Utility Theory: A Calibration Theorem, Econometrica 68,
12811292.
Roth, A. E. and I. Erev. (1995). Learning in Extensive-Form Games: Experimental Data and Simple Dynamic
Models in the Intermediate Term, Games and Economic Behavior 8, 164212.
Roth, A. E. and M. W. Malouf. (1979). Game-Theoretic Models and the Role of Information in Bargaining,
Psychological Review 86, 574594.
Schotter, A. and Y. D. Brounstein. (1981). Economic Search: An Experimental Study, Economic Inquiry 19,
125.
Selten, R. and J. Buchta. (1994). Experimental Sealed Bid First Price Auctions with Directly Observed Bid
Function, Discussion Paper No. B-270, University of Bonn.
Sonnemans, J. (1998). Strategies of Search, Journal of Economic Behavior and Organization 35, 309332.
Stigler, G. J. (1961). The Economics of Information, Journal of Political Economy 69, 213225.
Telser, L. G. (1973). Searching for the Lowest Price, American Economic Review 63, 4049.