[EM] Why the concept of "sincere" votes in Range is flawed.

Abd ul-Rahman Lomax abd at lomaxdesign.com
Tue Nov 25 11:36:15 PST 2008


And now that rarity from me, an original post....

Approval Voting is a special case of Range, with 
rating values restricted to 0 and 1. When Brams 
proposed Approval, it was as a method free of 
vulnerability to "tactical" or "strategic" 
voting, i.e., voting with reversed preference in 
order to produce a better outcome. And, indeed, 
both Range and Approval are immune to that, i.e., 
there is no advantage to be gained by it, ever 
(at least not in terms of outcome).

The proponents of other methods attacked this by 
redefining -- without ever being explicit about 
it -- the meaning of strategic voting. Because 
the concept was developed to apply to methods 
using a preference list, whether explicit on the 
ballot or presumed to exist in the mind of the 
voter, a strategic vote was one which reversed 
preference, simple. But with Approval and Range, 
it is possible to vote equal preference. Is that 
insincere if the voter has a preference? The 
critics of Range and Approval have claimed so, 
and thus they can claim that Range and Approval 
are "vulnerable to strategic voting."

Arrow, in explaining why he did not study 
cardinal rating methods (like Range and 
Approval), methods that allow equal ranking, 
wrote that they offended him because there is no 
single sincere vote. I.e., a whole set of votes 
could be considered sincere. If the voter prefers 
A>B>C, the voter could vote for A or for A and B 
(and, for that matter, for A and B and C), and still be sincere.

(A side-note: unless a preferential ballot allows 
ranking all candidates, it does allow equal 
ranking *at the bottom,* indeed it requires it. 
But we've tended to focus on the winner only.)

The critics, I've seen, will consider a vote for 
A only, with Approval, when the voter supposedly 
"approves" both A and B, to be "strategic." It 
certainly is strategic in the sense of "smart," 
under some conditions. However, this is where 
preference strength comes in, and a strange twist 
of the definitions takes place. We must assume 
that if the voter votes only for A, the voter 
does, indeed, prefer A. So with a preferential 
method, as with Plurality, the vote for A alone 
is sincere, and a vote for B alone would be 
insincere. In other words, Approval voting is 
"vulnerable" to a voter voting what would be 
considered a sincere vote in a method that does 
not allow equal ranking. This is having the critical cake and eating it too.

So what are "sincere" Range and Approval votes? 
Should voters in Range vote "sincerely?" Or 
should they vote "strategically," which means 
that their vote is different depending on their 
perception of the election probabiities. The 
voter votes only for A in the example above, even 
though the voter supposedly "approves" of B as 
well, because the voter perceives the important 
choice as being between A and B, with C being 
unimportant. If the voter sees C as possibly 
winning, with significant probability, the voter 
is much more likely to vote for both A and B.

The root of the critical problem is that votes 
have been considered expressions of preference 
alone, and the goal has been to find a voting 
method that works, even in the presence of voter 
knowledge of the election probabilities, just 
like a zero-knowledge election. The problem is 
that this is a strange and artificial creation, 
when we look at it carefully. It doesn't exist in 
the real world, and there are many obstacles in 
the way of it, including Arrow's theorem and how 
people will always behave. When we ask people 
what they want, they will *always* modify the 
answers according to how they perceive the probabilities of each possibility.

Which would you prefer, $10 or $100? Seems 
simple, eh? And I've argued that any good ballot 
design will allow you to express that preference. 
However, suppose there are three alternatives, $0 
or $10 or $100. We can easily rank these, but 
suppose that these are personal utilities for the 
three alternatives, and they are not identical 
for all the voters, and some voters will prefer 
the outcomes in a different order. And if we vote 
100>10>0, the probabilities, in our judgement, 
are that we'll get 0. While the 100 outcome is 
obviously preferable to us, we consider it 
unlikely. So how do we vote in an Approval 
election? I've set it up to be obvious. We vote 
for 100 and for 10. Now, how do we vote in Range? 
The supposed sincere vote, based on true personal 
utilities, which we've made obvious, would be, in 
Range 100, to vote the dollar values. Yet that 
would be almost as foolish as to vote for 100 only.

I came across the following piece, at 
http://cepa.newschool.edu/het/essays/uncert/vnmaxioms.htm


>In the von Neumann-Morgenstern hypothesis, 
>probabilities are assumed to be "objective" or 
>exogenously given by "Nature" and thus cannot be 
>influenced by the agent. However, the problem of 
>an agent under uncertainty is to choose among 
>lotteries, and thus find the "best" lottery in D 
>(X). One 
>of 
><http://cepa.newschool.edu/het/profiles/neumann.htm>von 
>Neumann and 
><http://cepa.newschool.edu/het/profiles/morgenst.htm>Morgenstern's 
>major contributions to economics more generally 
>was to show that if an agent has preferences 
>defined over lotteries, then there is a utility 
>function U: D (X) ® R that assigns a utility to 
>every lottery p Î D (X) that represents these preferences.
>
>Of course, if lotteries are merely 
>distributions, it might not seem to make sense 
>that a person would "prefer" a particular 
>distribution to another on its own. If we follow 
><http://cepa.newschool.edu/het/essays/uncert/bernoulhyp.htm>Bernoulli's 
>construction, we get a sense that what people 
>really get utility from is the outcome or 
>consequence, x Î X. We do not eat 
>"probabilities", after all, we eat apples! Yet 
>what von Neumann and Morgenstern suggest is 
>precisely the opposite: people have utility from 
>lotteries and not apples! In other words, 
>people's preferences are formed over lotteries 
>and from these preferences over lotteries, 
>combined with objective probabilities, we can 
>deduce what the underlying preferences on 
>outcomes might be. Thus, in von 
>Neumann-Morgenstern's theory, unlike 
>Bernoulli's, preferences over lotteries 
>logically precede preferences over outcomes.
>
>How can this bizarre argument be justified? It 
>turns out to be rather simple actually, if we 
>think about it carefully. Consider a situation 
>with two outcomes, either $10 or $0. Obviously, 
>people prefer $10 to $0. Now, consider two 
>lotteries: in lottery A, you receive $10 with 
>90% probability and $0 with 10% probability; in 
>lottery B, you receive $10 with 40% probability 
>and $0 with 60% probability. Obviously, the 
>first lottery A is better than lottery B, thus 
>we say that over the set of outcomes X = ($10, 
>0), the distribution p = (90%, 10%) is preferred 
>to distribution q = (40%, 60%). What if the two 
>lotteries are not over exactly the same 
>outcomes? Well, we make them so by assigning 
>probability 0 to those outcomes which are not 
>listed in that lottery. For instance, in Figure 
>1, lotteries p and q have different outcomes. 
>However, letting the full set of outcomes be (0, 
>1, 2, 3), then the distribution implied by 
>lottery p is (0.5, 0.3, 0.2, 0) whereas the 
>distribution implied by lottery q is (0, 0, 0.6, 
>0.4). Thus our preference between lotteries with 
>different outcomes can be restated in terms of 
>preferences between probability distributions 
>over the same set of outcomes by adjusting the set of outcomes accordingly.
>
>But is this not arguing precisely what 
><http://cepa.newschool.edu/het/profiles/bernoulli.htm>Bernoulli 
>was saying, namely, that the "real" preferences 
>are over outcomes and not lotteries? Yes and no. 
>Yes, in the sense that the only reason we prefer 
>a lottery over another is due to the implied 
>underlying outcomes. No, in the sense that 
>preferences are not defined over these outcomes 
>but only defined over lotteries. In other words, 
>von Neumann and Morgenstern's great insight was 
>to avoid defining preferences over outcomes and 
>capturing everything in terms of preferences 
>over lotteries. The essence of von Neumann and 
>Morgenstern's expected utility hypothesis, then, 
>was to confine themselves to preferences over 
>distributions and then from that, deduce the 
>implied preferences over the underlying outcomes.

"Preferences" in Range Voting is preferences over 
lotteries, not preferences over outcomes, as 
such. I, of course, support voting methods which 
allow the expression of both, hybrid methods, and 
which resolve the occasional conflict between a 
sum-of-votes approach and a pairwise winner 
approach, using not the original ballot, but a 
new one, i.e., a runoff that turns the choice 
involved back to the voters. Some supporters of 
Range are disturbed by this, because, supposedly, 
the Range Votes, summed, elect the social utility 
winner, which, they argue is the best winner for 
society. However, they've neglected the overall 
process in favor of resolving it in a single 
ballot. If a single ballot *must* be used, no 
matter what the cost, the Range outcome is indeed 
the closest we can get to ideal, I suspect. But 
we are not limited to that, and we can go back to 
the voters -- a different set of voters, usually! 
-- and ask them. The exact details of that 
additional election I'll leave for another paper; 
parliamentary procedure would suggest that it be 
an entirely new election, informed by the results 
of the first one, plus additional campaigning, 
but practicality may suggest something different. 
Regardless, it's apparent to me that two ballots 
is better than one, when one doesn't come up with 
a clear majority choice or better.

Economics. It seems to be a field that is 
disreputable to political scientists. But this 
is, of course, a field where substantial 
theoretical expertise has been applied to the 
problem of making decisions. A voting system is, 
obviously, such a problem. Warren Smith found 
this paper and pointed it out to us:

http://ideas.repec.org/a/ecm/emetrp/v67y1999i3p471-498.html

This is a paper by Dhillon and Mertens. An abstract:

>In a framework of preferences over lotteries, 
>the authors show that an axiom system consisting 
>of weakened versions of Arrow's axioms has a 
>unique solution, 'relative utilitarianism.' This 
>consists of first normalizing individual von 
>Neumann-Morgenstern utilities between zero and 
>one and then summing them. The weakening 
>consists chiefly in removing from IIA the 
>requirement that social preferences be 
>insensitive to variations in the intensity of 
>preferences. The authors also show the resulting 
>axiom system to be in a strong sense independent.

Relative Utilitarianism is an analytical method 
which takes as input Range Votes; as Warren Smith 
has stated he prefers, the Votes are rational 
numbers (I think) in the range of 0-1, with no 
restriction on resolution. I.e., practical Range 
Voting uses some specified resolution; I define 
Range N as being Range with N+1 choices, so Range 
1 is Approval (with two choices, 0 and 1), and we 
can express Range votes as 0-N; i.e, Range 100 
may vote as 0-100, but is really 0-1 in steps of 1/100 vote.

In pure relative utilitarianism, then, unless the 
voter is indifferent to a choice, the vote in 
that choice will always show preference, but the 
magnitude of the preference will vary according 
to perceived probabilities. In practical Range 
Voting, if the von Neumann-Morgenstern utilities 
get rounded off, thus showing equal preference 
when the reality is that there is an underlying 
preference, but with a combination of absolute 
magnitude and relative probability that brings it 
within the resolution of the Range method.

Now, voters don't sit down with a calculator, but 
what was claimed in the first paper above is that 
this is, in fact, how we make decisions. It's 
much simpler than one might think, and in real 
elections, the normal procedure for determining 
these utilities is quite simple in *most* 
elections: Pick two frontrunners (which depends 
on probabilities only, not personal preferences). 
Then use preferences to rate one of them at 
maximum and one at minimum. If one has 
preferences of significance outside this set and 
this range (i.e., one has a candidate preferred 
over the best frontrunner and one over the worst 
frontrunner, then one might consider, if the 
method has sufficient resolution, pulling the 
frontrunner down a notch or the worst up a notch, 
to preserve preference expression. Alternatively, 
perhaps the method allows expression of preference independently of rating.

Only when there are three candidates considered 
possible winners does it get more complicated. 
But the point of all this is that voters will 
always consider election probabilities, and thus 
pure Independence of Irrelevant Alternatives is a 
real stumbling block if insisted upon. The voting 
power I assign to the pairwise preference of $10 
to $100 must depend on, not only my pure 
intensity of preference, but on my perception of 
the probabilities. Voters in Range are choosing 
lotteries with specified prizes, with values and 
probabilities as estimated by the voter, and setting their votes accordingly.

And they sincerely choose them, i.e., they 
attempt to maximize their personal expected 
return, and this is *exactly* what we want them 
to do. It's not "greedy" or "selfish," it's 
"intelligent." Now, the shift in votes due to the 
probability perceptions can be mistaken. A dark 
horse candidate may not receive the full vote 
strength that the candidate would receive in a 
zero-knowledge election, with all candidates 
being considered equally likely. Thus, we'd need 
runoffs to fix problems, which might be detected 
through preference analsyis. But the method is 
theoretically ideal, as Dhillon and Mertens show, 
it is a unique solution to Arrow's theorem (with 
a minimal tweak, one that is utterly necessary; 
the requirement of absolute Independence of 
Irrelevant Alternatives was, quite simply, a 
mistaken intuition. Relative Utilitarianism 
doesn't require -- actually does not allow, in 
its pure form, -- the suppression of preferences, 
but the *magnitude* of the expressed preference 
varies with the alternatives, and that 
technically violates IIA, as it was understood.

This brings us to a problem with Range Voting. If 
voters expect that their task, with Range, is to 
give "sincere ratings," regardless of the effect 
on results, Range will suffer badly from IIA and, 
indeed, as claimed by critics, voters who pay no 
attention to irrelevant candidates, in 
determining how they vote for two frontrunners, 
will have an advantage, through "bullet voting" 
or through "exaggerating." Range Voting is still 
*voting,* but merely with fractional votes 
allowed, not required, when N is greater than 1. 
Approval is still voting. It is not about 
"approving" the candidates, except in the sense 
that by voting for a candidate, one is approving 
the election of that candidate, *compared to the 
likely alternatives*. It is not a sentiment, it's 
an action, adding weight to an outcome, choosing 
to effectively participate in it. Add weight to 
an irrelevant alternative, it doesn't matter, by 
definition. In almost all elections, there are 
two frontrunners, and, this is why Plurality 
usually works, and only breaks down through the 
related spoiler or center squeeze effects, 
because of the restriction against voting for 
more than one. All advanced voting methods -- 
with one exception, not applied anywhere that I'm 
aware of, Asset Voting -- allow voting for more 
than one, but through various procedures. (Even 
Asset would normally allow voting for more than 
one, the original form was proposed for an STV 
ballot with optional preferential voting, to deal 
with the very common problem of exhausted ballots.)

Comments invited.




More information about the Election-Methods mailing list