Bengt Holmström Prize Lecture: Pay for Performance and Beyond



Yüklə 2,49 Mb.
Pdf görüntüsü
səhifə5/11
tarix09.08.2018
ölçüsü2,49 Mb.
#62134
1   2   3   4   5   6   7   8   9   10   11

Pay For Performance and Beyond 

423


So, assume as before that the agent can choose between low (L) or high (H) 

effort with a (utility) cost-differential equal to c > 0. Note that the likelihood 

ratio f

L

(x)/f



H

(x) with the normal distribution goes to infinity as x goes to negative 

infinity. Therefore, the right-hand side of (4), which is supposed to characterize 

the optimal solution, must be negative for sufficiently small values of x, because 

the values for λ and μ are strictly positive when high effort is implemented as 

discussed earlier. But that is inconsistent with the left-hand side being strictly 

positive for all x regardless of the function s

H

(x). How can that be? Mirrlees 



conjectured and proved that this must be because the first-best solution can be 

approximated arbitrarily closely, but never achieved.

10

In particular, consider a scheme s(x) which pays the agent a fixed amount 



above a cutoff z, say, and punishes the agent by a fixed amount if x falls below 

z. If the utility function is unbounded below, Mirrlees shows that one can con-

struct a sequence of increasingly lenient standards z, each paired with an increas-

ingly harsh punishment that maintains the agent’s incentive to work hard and 

the agent’s willingness to accept all contracts in the sequence. Furthermore, this 

sequence can be chosen so that the principal’s expected utility approaches the 

first-best outcome.

One intuition for this paradoxical result is that while one cannot set μ = 0 in 

(4), one can approach this characterization by letting μ go to zero in a sequence 

of schemes that assures that the right-hand side stays positive.

The inference logic suggests an intuitive explanation. Despite the way the 

normal distribution appears to the eye, tail events are extremely informative 

when one tries to judge whether the observed outcome is consistent with the 

agent choosing a high rather than a low level of effort. The likelihood ratio f

L

(x)/


f

H

(x) goes to infinity in the lower tail, implying that an outcome far out in the 



tail is overwhelming evidence against the H distribution. Statistically, the normal 

distribution is more akin to the moving box example discussed in connection 

with first-best outcomes. Punishments in the tail are effective, because one can 

be increasingly certain that L was the source.



III. TOWARD A MORE REALISTIC MODEL

A. Why Are Incentive Schemes Linear?

One could brush aside Mirrlees’ example as extremely unrealistic (which it is) 

and proceed to study linear contracts because they are simple and widely used. 

But this would defeat the purpose of studying general incentives. An additive 




424 

The Nobel Prizes

production function with a normal error term is a very natural example to study. 

The unrealistic solution prods us to look for a more fundamental reason for lin-

earity than simplicity and to revisit the assumptions underlying the basic princi-

pal-agent model. This led us to the analysis in Holmström and Milgrom (1987).

The optimum in the basic model tends to be complex, because there is an 

imbalance between the agent’s one-dimensional effort space and the infinite-

dimensional control space available to the principal.

11

 The principal is therefore 



able to fine-tune incentives by utilizing every minute piece of information about 

the agent’s performance resulting in complex schemes.

The bonus scheme that approaches first-best in Mirrlees’ example works for 

the same reason. It is finely tuned because the agent is assumed to control only 

the mean of the normal distribution. If we enrich the agent’s choice space by 

allowing the agent to observe some information in advance of choosing effort, 

the scheme will perform poorly. Think of a salesperson that is paid a discrete 

bonus at the end of the month if she exceeds a sales target. Suppose she can see 

her progress over time and adjust her sales effort accordingly. Then, if the month 

is about to end and she is still far from the sales target, she may try to move 

potential sales into the next month and get a head start towards next month’s 

bonus. Or, if she has already made the target, she may do the same. This sort 

of gaming is common (see Healy 1985 and Oyer 1998). To the extent that such 

gaming is dysfunctional one should avoid a bonus scheme.

Linear schemes are robust to gaming. Regardless of how much the salesper-

son has sold to date, the incentive to sell one more unit is the same. Following this 

logic, we built a discrete dynamic model where the agent sells one unit or none 

in each period and chooses effort in each period based on the full history up to 

that time. The principal pays the agent at the end of time, basing the payment on 

the full path of performance. Assuming that the agent has an exponential utility 

function and effort manifests itself as an opportunity cost (the cost is financial), it 

is optimal to pay the agent the same way in each period—a base pay and a bonus 

for a sale.

12

 The agent’s optimal choice of effort is also the same in each period, 



regardless of past history. Therefore, the optimal incentive pay at the end of the 

term is linear in total sales, independently of when sales occurred.

13

The discrete model has a continuous time analog where the agent’s instanta-



neous effort controls the drift of a Brownian motion with fixed variance over the 

time interval [0,1]. The Brownian model can be viewed as the limit of the discrete 

time model under suitable assumptions.

14

 It is optimal for the principal to pay 



the agent a linear function of the final position of the process (the aggregate 

sales) at time 1 and for the agent to choose a constant drift rate (effort) at each 




Yüklə 2,49 Mb.

Dostları ilə paylaş:
1   2   3   4   5   6   7   8   9   10   11




Verilənlər bazası müəlliflik hüququ ilə müdafiə olunur ©www.genderi.org 2024
rəhbərliyinə müraciət

    Ana səhifə