Chapter 5: Aggregation
Chapters 1–4 focused on preferences of a single decision-maker. This chapter asks: how do we aggregate preferences across multiple individuals?
Motivating examples:
| Symbol | Meaning |
|---|---|
| \(N = \{1, \ldots, n\}\) | Set of \(n\) voters (agents) |
| \(A = \{a_1, \ldots, a_m\}\) | Set of \(m\) alternatives |
| \(\succ_i\) | Voter \(i\)’s strict preference ordering over \(A\) |
| \(\mathcal{L}(A)\) | Set of all strict linear orders over \(A\) |
| \(f: \mathcal{L}(A)^n \to A\) | Social choice function (SCF): profile \(\to\) winner |
| \(F: \mathcal{L}(A)^n \to \mathcal{L}(A)\) | Social welfare function (SWF): profile \(\to\) ranking |
| \(\text{SP}(Y)\) | Single-peaked preferences on totally ordered set \(Y\) |
| \(p(\succ_i)\) | Peak (ideal point) of voter \(i\)’s preferences |
Majority preferences can be cyclic — even when individual preferences are transitive:
| Voter | Ranking |
|---|---|
| Voter 1 | \(A \succ B \succ C\) |
| Voter 2 | \(B \succ C \succ A\) |
| Voter 3 | \(C \succ A \succ B\) |
No Condorcet winner exists! — a rock-paper-scissors cycle
Three desirable properties for any social welfare function:
Additionally, we assume unrestricted domain: any transitive preference ordering is admissible.
Important
Theorem (Arrow, 1951): For \(m \geq 3\) alternatives, no social welfare function can simultaneously satisfy:
under unrestricted domain.
Every practical voting system must sacrifice at least one fairness criterion.
Arrow (1951)
The proof proceeds by contradiction:
Key driver: Condorcet cycles force the aggregation to “break ties” by deferring to one voter.
| Voting Rule | Violates | Why |
|---|---|---|
| Dictatorship | Non-dictatorship | One voter decides everything |
| Plurality | IIA | Adding a “spoiler” changes the winner |
| Borda Count | IIA | Removing an alternative changes point totals |
| Pairwise Majority | Transitivity | Condorcet cycles |
Takeaway: There is no free lunch — every voting rule makes trade-offs.
Important
Theorem (Gibbard, 1973; Satterthwaite, 1975): For \(m \geq 3\) alternatives, any social choice function \(f\) that is:
must be dictatorial.
Every non-dictatorial voting rule is manipulable: some voter can gain by voting insincerely.
Gibbard (1973); Satterthwaite (1975)
Plurality — “Lesser of two evils”:
Borda Count — Strategic ranking:
Practical deterrence: While STV can always be manipulated in theory, finding a beneficial strategic vote can be NP-hard in worst cases.
Arrow’s and Gibbard–Satterthwaite’s theorems apply to any preference aggregation:
Modern approaches:
Gordon et al. (2022)
Arrow’s and Gibbard–Satterthwaite assume unrestricted domain: any transitive ordering is admissible.
Key idea: If we restrict which preferences can occur, we can escape impossibility!
In many real-world settings, preferences have natural structure we can exploit.
Note
Definition: A preference ordering \(\succ\) over a totally ordered set \(Y\) is single-peaked if there exists a peak \(p(\succ) \in Y\) such that:
Intuition: Each voter has an “ideal point” (peak), and utility decreases as alternatives move away from the peak in either direction.
This rules out “I prefer the extremes to the middle” — which creates cycles.
Three colleagues choosing the office thermostat (65°F to 75°F):
All three have single-peaked preferences on the temperature line.
Median voter outcome: 70°F (Carol’s peak) — and no one can profitably manipulate!
Note
Definition: Fix phantom votes \(a_1, \ldots, a_{n-1} \in \mathbb{R} \cup \{\pm\infty\}\). The generalized median voter scheme is:
\[ f(\succ_1, \ldots, \succ_n) = \text{median}\big(p(\succ_1), \ldots, p(\succ_n), a_1, \ldots, a_{n-1}\big) \]
The \(n-1\) phantom votes act as anchor points that shift the median:
Different phantom choices yield different rules:
| Rule | Phantom Votes | Effect |
|---|---|---|
| Pure median | All \(\pm\infty\) | Outcome = median of voter peaks |
| Dictatorial | All equal to voter \(i\)’s peak | Voter \(i\) always wins |
| Status quo | All equal to status quo \(q\) | Change requires consensus |
The phantom votes allow tuning the rule between fully responsive and highly conservative.
Important
Theorem (Moulin, 1980): On the domain of single-peaked preferences \(\text{SP}(Y)\), a social choice function \(f\) satisfies:
if and only if \(f\) is a generalized median voter scheme.
By restricting to single-peaked preferences, we escape Arrow’s impossibility and achieve both strategy-proofness and efficiency!
Moulin (1980)
Why can’t voters manipulate the median?
Key insight: You can only move the median if your peak crosses it — but then the outcome moves away from your true peak.
Another escape from Arrow: relax IIA instead of restricting the domain.
Note
Definition (Borda Count): The Borda score of alternative \(y\) counts pairwise wins:
\[ \text{Borda}(y) = \sum_{i=1}^{n} |\{y' \neq y : y \succ_i y'\}| \]
The Borda winner is the alternative with the maximum Borda score.
Equivalently: with \(m\) alternatives, voter gives \(m-1\) points to top, \(m-2\) to second, …, 0 to last.
Borda violates IIA but satisfies a weaker version:
Note
Definition (IIA’): If two profiles have, for every voter:
then the social choice should not flip between \(y\) and \(y'\).
Important
Theorem: Borda satisfies Unrestricted Domain, Pareto Efficiency, Non-dictatorship, and IIA’. By relaxing IIA to IIA’, we escape Arrow’s impossibility.
A remarkable result connects Borda to modern RLHF:
Important
Theorem (DPO-Borda Equivalence): Assume responses \(y, y'\) are drawn from \(\pi_{\text{ref}}(\cdot \mid x)\). The DPO-optimal policy satisfies:
\[ \frac{\pi_{\text{DPO}}(y \mid x)}{\pi_{\text{ref}}(y \mid x)} \propto \text{(weighted Borda score of } y \text{)} \]
DPO upweights responses proportionally to their Borda scores — it finds the response that wins the most head-to-head matchups.
Rafailov et al. (2023)
The DPO loss is: \[ \mathcal{L}_{\text{DPO}}(\pi) = -\mathbb{E}_{x,y,y'}\Big[\bar{\sigma}(\Delta r^*) \cdot \log \sigma\big(\beta \log \tfrac{\hat{\pi}(y' \mid x)}{\hat{\pi}(y \mid x)}\big) + \cdots\Big] \]
Taking the gradient and setting to zero: \[ \mathbb{E}_{y' \sim \hat{\pi}}\Big[\sigma\big(\beta \log \tfrac{\pi(y \mid x)}{\pi(y' \mid x)}\big)\Big] = \underbrace{\mathbb{E}_{y' \sim \mathcal{D}}\big[\bar{\sigma}(\Delta r^*(x, y', y))\big]}_{\text{Borda score of } y} \]
The RHS is the expected win rate of \(y\) against a random alternative — exactly its Borda score.
Social choice interpretation of DPO:
Implications:
Real-world decisions often involve multiple independent issues.
Example in RLHF: Optimize for helpfulness, harmlessness, and honesty simultaneously.
Question: Can we aggregate each criterion independently?
Note
Definition: A voting scheme is voting by committees if for each object \(x \in K\), there exists a committee \(C_x\) with winning coalitions \(W_x\) such that:
The outcome includes \(x\) \(\iff\) \(\{i : x \in B(\succ_i)\} \in W_x\)
where \(B(\succ_i)\) is voter \(i\)’s top-ranked subset.
Each issue is decided independently by its own committee — a natural decomposition.
Note
Definition: A preference \(\succ\) on \(2^K\) is separable if for all \(A \subseteq K\) and \(x \notin A\):
\[ A \cup \{x\} \succ A \quad \Longleftrightarrow \quad x \in G(\succ) \]
where \(G(\succ) = \{x \in K : \{x\} \succ \emptyset\}\) is the set of “good” objects.
Separability means: whether you want to add \(x\) to a bundle doesn’t depend on what’s already there.
Important
Theorem: A voting scheme satisfies surjectivity, strategy-proofness, and separability if and only if it is voting by committees.
Caveat: Voting by committees generally does not satisfy Pareto efficiency.
Application to RLHF: Preferences over “helpful” and “harmless” are often not separable — a highly helpful response may necessarily involve some risk of harm, creating dependencies.
Note
Definition: A preference is nosy if the individual cares about outcomes affecting others, not just themselves. A preference is private if the individual only cares about their own allocation.
Examples of nosy preferences:
Important
Theorem (Sen, 1970): The following three properties are inconsistent:
When preferences are nosy, even weak requirements conflict!
Sen (1970)
Two individuals and a controversial book. Alternatives: \(a\) (Prude reads), \(b\) (Lewd reads), \(c\) (no one reads).
Prude: \(c \succ_P a \succ_P b\)
Prefers no one reads it, but would rather read it themselves than let Lewd read it (nosy!)
Lewd: \(a \succ_L b \succ_L c\)
Wants Prude to read it most of all (also nosy!)
\[ c \succ a \succ b \succ c \quad \text{— a cycle! No consistent social choice.} \]
Implication for AI: Content moderation involves exactly this tension — one user’s preference for free expression conflicts with another’s preference for a safe environment.
Community Notes (formerly Birdwatch) aggregates ratings about content helpfulness across ideological divides.
Problem with majority voting: The largest ideological group would dominate.
Solution: Find bridging notes — rated positively by users who disagree ideologically.
\[ u(y; \alpha, p, \varepsilon) = \mu + \alpha_j + \beta_j + p^\top q_j + \varepsilon_j \]
| Term | Interpretation |
|---|---|
| \(\mu\) | Global intercept |
| \(\alpha_j\) | Rater intercept (some raters more positive) |
| \(\beta_j\) | Note intercept (note quality) |
| \(p^\top q_j\) | Ideological alignment factor |
| \(\varepsilon_j\) | Residual noise |
Key insight: \(\beta_j\) captures note quality after controlling for ideology.
Connections:
Important
Core insight: Observed behavior \(\neq\) underlying preferences or utility.
Standard revealed preference assumes choices reveal preferences. This can fail due to:
A smart pantry observes eating behavior:
But: The user might prefer healthier options — they just succumb to availability and habit.
Lesson: Optimizing for observed “preferences” (engagement) may not optimize for true welfare.
This is the engagement vs. satisfaction problem in recommender systems.
The inversion problem directly affects AI training from human feedback:
Potential solutions:
Preference learning inherently involves collecting personal data.
Tension: Better personalization requires more data, but privacy demands less.
Note
Contextual Integrity (Nissenbaum): Privacy is preserved when information flows match context-specific norms. Five parameters:
A privacy violation occurs when information flows against contextual norms, even with consent.
Nissenbaum (2009)
A fitness tracker collects heart rate data:
| Flow | Recipient | Transmission Principle | Appropriate? |
|---|---|---|---|
| To running coach | Coach | Training optimization | Yes |
| To advertiser | Ad network | Targeted advertising | No |
Same data, same consent — but different transmission principles violate expectations about the fitness context.
Differential privacy (DP) provides formal guarantees, but has fundamental limits for preference learning:
Contextual Integrity as middle ground: Allow data use that matches expectations (personalization within a service) while preventing unexpected flows (selling to third parties).
When should an AI system override a user’s stated preferences?
Key distinction:
Four conditions that might justify intervention:
AI systems that exercise paternalism should:
Example: When an AI assistant refuses a request — is it paternalism (protecting the user) or nosy (protecting others)? Often both.
While voting aggregates ordinal preferences, mechanism design aggregates cardinal valuations (with money).
Central concept: Incentive compatibility — design rules so that rational agents reveal true preferences.
Key question: Can we align individual self-interest with social welfare?
Two objectives:
Mechanism:
Example: Bids = \((2, 6, 4, 1)\)
Vickrey (1961)
Bidding \(b_i = v_i\) is a dominant strategy (DSIC):
Result: Allocates to highest valuer \(\Rightarrow\) welfare-maximizing.
First-Price Auction
Second-Price (Vickrey)
By decoupling the price from the winner’s bid, Vickrey removes the incentive to shade.
Goal: Maximize seller’s expected revenue (not welfare).
Setup: Bidders’ values \(v_i \sim F\) i.i.d. Define the virtual valuation:
\[ \varphi(v) = v - \frac{1 - F(v)}{f(v)} \]
Myerson’s theorem: Allocate to the bidder with the highest non-negative virtual value. If all virtual values are negative, don’t sell.
For i.i.d. regular distributions: this is a second-price auction with an optimal reserve price \(r\).
Myerson (1981)
For \(v \sim \text{Uniform}[0,1]\): \(\varphi(v) = 2v - 1\)
Setting \(\varphi(v) \geq 0\): optimal reserve price \(r = 0.5\)
Revenue comparison (two bidders):
| Scenario | Probability | Revenue |
|---|---|---|
| Both below 0.5 | \(1/4\) | \(0\) (no sale) |
| Both above 0.5 | \(1/4\) | \(\approx 2/3\) (second-highest value) |
| One above, one below | \(1/2\) | \(0.5\) (reserve price) |
Expected revenue: \(0 + \tfrac{1}{6} + \tfrac{1}{4} = \tfrac{5}{12} \approx 0.417\) vs. \(\tfrac{1}{3} \approx 0.333\) without reserve.
Important
Theorem (Bulow & Klemperer, 1996): For i.i.d. regular \(F\):
\[ \mathbb{E}[\text{Rev}^{\text{(second-price)}}(n+1)] \geq \mathbb{E}[\text{Rev}^{\text{(optimal)}}(n)] \]
A simple second-price auction with one extra bidder outperforms the optimal auction with fewer bidders!
Practical takeaway: Use simple, transparent mechanisms and focus on attracting more participants rather than complex optimal designs.
Bulow and Klemperer (1996)
Generalization of Vickrey’s auction to multiple items and complex outcomes.
Setting: Outcomes \(\omega \in \Omega\); agent \(i\) has valuation \(v_i(\omega)\); quasilinear utility.
Allocation rule — maximize total reported value:
\[ \omega^* = \arg\max_{\omega \in \Omega} \sum_{i=1}^n b_i(\omega) \]
Each agent pays the externality they impose on others:
\[ p_i(b) = \underbrace{\max_{\omega \in \Omega} \sum_{j \neq i} b_j(\omega)}_{\text{Others' welfare without } i} - \underbrace{\sum_{j \neq i} b_j(\omega^*)}_{\text{Others' welfare with } i} \]
Intuition: You pay the “damage” your presence causes to everyone else.
Despite theoretical elegance, VCG faces practical hurdles:
Application: Spectrum auctions — billions of dollars at stake; multi-round simultaneous auctions used in practice.
Setting: Students grade each other’s work. Design a mechanism that incentivizes careful grading.
The lazy grader problem: Always giving 80% can yield 96% accuracy under naive scoring rules — the grader “cheats” by predicting the class average.
Solution: Optimize the scoring rule to maximize the gap between diligent grading and lazy strategies.
Result: Incentive compatibility aligns grader incentives with accurate assessment — “payments” are grade points.
Hartline et al. (2020)
Setting: A planner (system) interacts with strategic agents (users) who arrive sequentially.
Challenge: Without monetary transfers, how can the planner induce exploration?
Key tool: Information asymmetry — users only see their own recommendations.
Idea: Hide exploration in a pool of exploitation.
Users don’t know if they’re the guinea pig, so following the recommendation is optimal!
The expected gain from deviating (ignoring the recommendation):
\[ \mathbb{E}[\mu_1 - \mu_2 \mid I_t = 2] \leq \tfrac{1}{L}(\mu_1 - \mu_2) + (1 - \tfrac{1}{L})\mathbb{E}[\mu_1 - \mu_2 \mid \mu_1 \lt \mu_2] \cdot P[\mu_1 \lt \mu_2] \]
This is \(\leq 0\) when \(L \geq 12\).
Interpretation: The small chance of being the guinea pig is outweighed by the chance that the exploration action is actually better.
General algorithm: Turn any bandit algorithm into an incentive-compatible one.
Recipe: Wrap every decision that the bandit algorithm \(A\) makes with \(L-1\) recommendations of the best-known arm.
Result: Simulates \(T\) steps of \(A\) in \(cT\) steps, achieving \(O(\sqrt{T})\) regret — the same rate as non-strategic settings!
Incentive compatibility comes “for free” (up to a constant factor).
Mansour, Slivkins, and Syrgkanis (2019)
Problem: How to incentivize truthful reporting when there’s no verifiable ground truth?
MIP (Kong & Schoenebeck, 2019): Reward agents based on mutual information between their report and a reference agent’s report:
\[ \text{Payment}_i = MI(\hat{\Psi}_i;\; \hat{\Psi}_j) \]
where \(j \neq i\) is randomly selected.
Kong and Schoenebeck (2019)
An information-monotone MI measure satisfies:
Two important families:
Important
Theorem: When the MI measure is strictly information-monotone, the resulting mechanism is:
Why it works: Any manipulation (noise, partial reporting) can only decrease mutual information with the reference agent — so truthful reporting maximizes payment.
Social Choice Theory:
Beyond Classical Voting:
Challenges in Practice:
Mechanism Design:
Incentives Without Money:

Chapter 5: Aggregation
Social Choice Theory
The central question: Can we design an aggregation rule that faithfully represents individual preferences while satisfying fairness axioms?
Common voting rules: