Line 7:
Line 7:
== Abstract ==
== Abstract ==
+
We give mechanisms in which each of n players in a game is given their
+
component of an (approximate) equilibrium in a way that guarantees
+
differential privacy --- that is, the revelation of the equilibrium
+
components does not reveal too much information about the utilities of
+
other players. More precisely, we show how to compute an approximate
+
correlated equilibrium (CE) under the constraint of differential privacy
+
(DP), provided $n$ is large and any player's action affects any other's
+
payoff by at most a small amount. Our results draw interesting connections
+
between noisy generalizations of classical convergence results for
+
no-regret learning, and the noisy mechanisms developed for differential
+
privacy. Our results imply the ability to truthfully implement good
+
social-welfare solutions in many games, such as games with small Price of
+
Anarchy, even if the mechanism does not have the ability to enforce
+
outcomes. We give two different mechanisms for DP computation of
+
approximate CE. The first is computationally efficient, but has a
+
suboptimal dependence on the number of actions in the game; the second is
+
computationally inefficient, but allows for games with exponentially many
+
actions. We also give a matching lower bound, showing that our results are
+
tight up to logarithmic factors.
+
Joint work with Michael Kearns, Mallesh Pai, and Jonathan Ullman.