1 Introduction

Bayesian auction design has been extremely flourishing in the areas of artificial intelligence, computational economics and operations research, since the seminal work of [1]. One of the main focuses is to generate revenue, by selling m heterogenous items to n players. Each player has a private valuation function describing how much she values each subset of the items, and the valuations are drawn from prior distributions. An important assumption in Bayesian mechanism design is that the distributions are commonly known by the seller and the players —the common prior assumption. However, as pointed out by another seminal work [2], such common knowledge is “rarely present in experiments and never in practice”, and “only by repeated weakening of common knowledge assumptions will the theory approximate reality.”

In this paper, we weaken the information assumption about the seller and the players by adopting an information elicitation approach [3]. We consider a framework for auctions where the knowledge about the players’ value distributions is arbitrarily scattered among the players and the seller. The seller must aggregate pieces of information from all players to gain a good understanding about the distributions, so as to decide how to sell the items.

As in information elicitation, the players get rewards for reporting their knowledge. However, different from classic information elicitation where a player’s utility is exactly her reward, in our model a player’s utility comes not only from her knowledge, but also from participating in the auction (i.e., from buying items). Moreover, information elicitation usually assumes the prior distribution is correlated: each player observes a private signal and reports their private information. This means every player has information about every other player. In our model, following the convention in multi-item auctions, the players’ value distributions for individual items are assumed to be independent. A player may be totally ignorant about some players and only partially know some other players’ distributions.

We focus on unit-demand auctions and additive auctions —two valuation types widely studied in the literature [4, 5]. In such auctions, a player’s valuation function is specified by m values, one for each item. For each player i and item j, the value \(v_{ij}\) is independently drawn from a distribution \(\mathcal {D}_{ij}\). When all players are unit-demand (respectively, additive), we call such an auction a unit-demand auction (respectively, an additive auction) for short. Each player privately knows her own values and some (or none) of the distributions of some other players for some items, like long-time competitors in the market. There is no constraint about who knows which distributions. The seller may also know some of the distributions, but she does not know which player knows what. We introduce directed knowledge graphs to succinctly describe the players’ knowledge. Each player knows the distributions of her neighbors, different items’ knowledge graphs may be totally different, and the structures of the graphs are not known by anybody.Footnote 1

Below we briefly state our main results.

1.1 Main results

Our goal is to design 2-step dominant strategy truthful (2-DST) information elicitation mechanisms whose expected revenue approximates that of the optimal Bayesian incentive compatible (BIC) mechanism, denoted by OPT.Footnote 2 In order for the seller to aggregate the players’ knowledge about the distributions, it is natural for the mechanism to ask each player to report her knowledge to the seller, together with her own values. A 2-DST mechanism [7] is such that, (1) no matter what knowledge the players may report about each other, it is dominant for each player to report her true values; and (2) given that all players report their true values, it is dominant for each player to report her true knowledge about others.

First, by exploring the knowledge graph’s combinatorial structure, we have the following result for single-good auctions.

Main result 1.

When the knowledge graph is 2-connected,Footnote 3. there is a 2-DST information elicitation mechanism for single-good auctions with revenue \(\ge (1-\frac{1}{n})OPT\).

When everything is known by somebody

For multi-item auctions, we first consider a setting when every distribution is known by at least \(k\ge 1\) players. We show the approximation ratios of the revenue generated by information elicitation mechanisms improve gracefully together with the amount of knowledge. More precisely, for any integer \(k\ge 1\), let \(\tau _k = \frac{k}{(k+1)^{\frac{k+1}{k}}}\). Note \(\tau _1 = \frac{1}{4}\) and \(\tau _k \rightarrow 1\) when k gets larger. We have the following result, formalized in Section 4.

Main result 2.

\(\forall k\in [n-1]\), when each distribution is known by at least k players, there is a 2-DST information elicitation mechanism for unit-demand auctions with revenue \(\ge \frac{\tau _k}{24}\cdot OPT\), and such a mechanism for additive auctions with revenue \(\ge \max \{\frac{1}{11}, \frac{\tau _k}{6+2\tau _k}\} OPT\).

Our mechanisms build on top of existing Bayesian mechanisms, such as the post-price mechanisms in [8] and the bundling mechanism in [9]. But it is worth pointing out that, although these mechanisms are used as a black-box, our mechanisms are not reductions from arbitrary Bayesian mechanisms.

Under arbitrary knowledge graphs

When the knowledge graphs are such that some distributions are not known by anybody, it is easy to see that no information elicitation mechanism can be a bounded approximation to OPT. Thus it is natural to consider the following benchmark: the optimal BIC mechanism applied to players and items for whom the distributions are indeed known by somebody, denoted by \(OPT_K\). This is a natural benchmark when considering players with limited knowledge and, if every distribution is known by somebody, then it is exactly OPT. We have the following result, formalized in Section 5.

Main result 3.

For any knowledge graph, there is a 2-DST information elicitation mechanism for unit-demand auctions with revenue \(\ge \frac{OPT_K}{96}\), and such a mechanism for additive auctions with revenue \(\ge \frac{OPT_K}{70}\).

To prove this result, for unit-demand auctions, we actually show a general result: any Bayesian mechanism for unit-demand auctions that is a good approximation in the COPIES setting (formally defined in Section 5.2) can be converted to information elicitation mechanisms. This applies to a large class of Bayesian mechanisms, including the ones in [4, 8, 10].

For additive auctions, we have developed a novel approach by using the adjusted revenue [11]. Although this concept is very useful in Bayesian auctions, it was unexpected that we found an interesting and highly non-trivial way of using it to analyze information elicitation mechanisms.

The use of scoring rules

Since our mechanisms elicit the players’ knowledge about each other’s value distributions, we can use scoring rules (see, e.g., [12]) to reward the players for their reported knowledge, as typical in information elicitation. However, the use of scoring rules does not solve the main problems in our auctions. Indeed, because a player’s utility comes both from the reward and from participating in the auction, the difficulties in designing information elicitation mechanisms are to guarantee that, even without rewarding the players for their knowledge, (1) it is dominant for each player to report her true values, (2) reporting her true knowledge never hurts him, and (3) the resulting revenue approximates the desired benchmark.

Accordingly, in Sections 3 to 5, we focus on designing information elicitation mechanisms without rewarding the players. Scoring rules are used later to break the utility-ties and make it strictly better for a player to report her true knowledge. In Section 6, we show how to add scoring rules to our mechanisms under the assumption of no-bluff and remove the assumption when everything is known.

1.2 Related work

Information elicitation

Following [3], information elicitation has become an important research area of artificial intelligence in the past decade [13,14,15]. where the mechanism asks each player to report her private signal about the prior distribution. The decision maker wants the mechanism to be BIC, and a player is rewarded based on her reported information and the other players’ reported signals. Proper scoring rules are widely used [12, 16,17,18,19]. The set of proper scoring rules was initially characterized in [16] and [20], and the proper scoring rules for characterizing various statistics of the subjective distributions are characterized in [21]. In our paper, we apply their techniques for eliciting the agents’ knowledge and then incorporate information elicitation into auction design by carefully designing the mechanisms to maximize the expected revenue without violating the properness constraints in scoring rules. Similar studies on the combination of information elicitation and mechanism design include [22,23,24,25,26]. The main difference is that those works focus on the direct optimization of scoring rules for incentivizing costly information acquisition without the additional concern of selling items for revenue maximization. Most studies on information elicitation require a common prior. Mechanisms without this assumption are considered by [27], and our work is information elicitation in auctions without a common prior. Moreover, information elicitation does not consider the players to have any cost for revealing their knowledge. We refer to a recent survey [28] for some recent trends in the field of information elicitation. It would be interesting to include such costs in the general model as well as in ours, to see how the mechanisms will change accordingly. In Section 6, we will formally define scoring rules and use them to reward the players for their knowledge.

Bayesian auction design

In his seminal work [1], Myerson introduced the first optimal Bayesian mechanism for single-good auctions, which also applies to many single-parameter settings [29]. Since then, there has been a huge literature on designing (approximately) optimal Bayesian mechanisms that are either BIC or dominant-strategy truthful (DST); see [30] for an introduction to this literature. Regarding mechanisms for multi-parameter settings, in [31], the authors characterize optimal BIC mechanisms for combinatorial auctions. For unit-demand auctions, [4, 8,9,10] construct DST Bayesian mechanisms that are constant approximations, and recently [32] designed benchmark-tight simple mechanism for a single unit-demand buyer. For additive auctions, [5, 9, 11, 33] provide logarithmic or constant approximations under different conditions. Finally, for general XoS and subadditive auctions, logarithmic and constant approximations are provided in [34,35,36,37,38,39].

Removing the common prior assumption

Following [2], a lot of effort has been made to remove the common prior assumption. In DST Bayesian mechanisms it suffices to assume that the seller knows the prior distribution. In prior-free mechanisms [40,41,42,43,44] the distribution is unknown and the seller learns it from the values of randomly selected players. In [45,46,47] and [48, 49] the seller observes independent or targeted samples from the distribution before the auction begins. In [50, 51] the players have arbitrary possibilistic belief hierarchies about each other. There is another line of research that assumes certain statistics of distributions are known, [52,53,54,55]. In robust mechanism design [56] the players have arbitrary probabilistic belief hierarchies. In crowdsourced Bayesian auctions [7] each player privately knows all the distributions (or their refinements), which is a special case of our model. Indeed, all knowledge graphs will be complete graphs under their setting (that is, everybody knows everything), while we allow arbitrary knowledge graphs. In Section 7 and Appendix B, we further discuss how to elicit the players’ knowledge refinements, and how to handle correlated distributions in a setting that is a special case of our model but is still more general than that of [7].

2 Preliminaries

In this work, we focus on multi-item auctions with n players (denoted by N) and m items (denoted by M). A player i’s value for an item j, \(v_{ij}\), is independently drawn from a distribution \(\mathcal {D}_{ij}\). Let \(v_i = (v_{ij})_{j\in M}\), \(\mathcal {D}_i = \times _{j\in M} \mathcal {D}_{ij}\) and \(\mathcal {D}=\times _{i\in N} \mathcal {D}_i\). Player i’s value for a subset S of items is \(\max _{j\in S} v_{ij}\) in unit-demand auctions, and is \(\sum _{j\in S} v_{ij}\) in additive auctions. The players’ utilities, denoted by \(u_i\), are quasi-linear, and the players are risk-neutral.

Knowledge graphs

It is illustrative to model the players’ knowledge graphically.Footnote 4 We consider a vector of knowledge graphs, \(G = (G_j)_{j\in M}\), one for each item. Each \(G_j\) is a directed graph with n nodes, one for each player. For any \(i\ne i'\), an edge \((i, i')\) is in \(G_j\) if and only if player i knows \(\mathcal {D}_{i'j}\). There is no constraint about the knowledge graphs: the same player’s distributions for different items may be known by different players, different players’ distributions for the same item may also be known by different players, and some distributions may not be known by anybody. Each player knows her own out-going edges, and neither the players nor the seller knows the whole graph.

We measure the amount of knowledge in the system by the number of players knowing each distribution. For any \(k\in \{0,1,\dots,n-1\}\), a knowledge graph is k-informed if each node has in-degree at least k: a player’s distribution is known by at least k other players. The vector G is k-informed if all knowledge graphs are so. Note that every knowledge graph is 0-informed, and “everything is known by somebody” when \(k\ge 1\). A common prior would imply all knowledge graphs are complete directed graphs, or \((n-1)\)-informed, which is the strongest condition in our model. The seller’s knowledge can be naturally incorporated into the knowledge graphs by considering her as a special “player 0”. All our mechanisms can easily utilize the seller’s knowledge, and we will not further discuss this issue.

Information elicitation mechanisms

Let \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) be a Bayesian auction instance and \(\mathcal {I}= (N, M, \mathcal {D}, G)\) a corresponding information elicitation instance, where G is a knowledge graph vector. Different from Bayesian mechanisms, which has \(\mathcal {D}\) as input, an information elicitation mechanism has neither \(\mathcal {D}\) nor G as input. Instead, it asks each player i to report a valuation \(b_i = (b_{ij})_{j\in M}\) and a knowledge \(K_i = \times _{i'\ne i, j\in M} \mathcal {D}^i_{i'j}\) —a distribution for the valuation subprofile \(v_{-i}\). \(K_i\) may contain “\(\bot\)” at some places, indicating i does not know the corresponding distributions. \(K_i\) is i’s true knowledge if \(\mathcal {D}^i_{i'j} = \mathcal {D}_{i'j}\) whenever \((i, i')\in G_j\), and \(\mathcal {D}^i_{i'j} = \bot\) otherwise. An information elicitation mechanism maps a strategy profile \((b_i, K_i)_{i\in N}\) to an allocation and a price profile, and may be randomized. To distinguish whether a mechanism \(\mathcal {M}\) is a Bayesian or an information elicitation mechanism, we may explicitly write \(\mathcal {M}(\hat{\mathcal {I}})\) or \(\mathcal {M}(\mathcal {I})\). The (expected) revenue of \(\mathcal {M}\) is denoted by \(Rev(\mathcal {M})\), and sometimes by \(\mathbb {E}_{\mathcal {D}} Rev(\mathcal {M})\) to emphasize the distribution.

An information elicitation mechanism is 2-step dominant strategy truthful (2-DST) if

  1. (1)

    For any player i, true valuation \(v_i\), valuation \(b_i\), knowledge \(K_i\), and strategy subprofile \(s_{-i} = (b_j, K_j)_{j\ne i}\) of the other players, \(u_i((v_i, K_i), s_{-i}) \ge u_i((b_i, K_i), s_{-i})\).

  2. (2)

    For any player i, true valuation \(v_i\), true knowledge \(K_i\), knowledge \(K'_i\), and knowledge subprofile \(K'_{-i}(v_{-i}) = (K'_j(v_j))_{j\ne i}\) of the other players, where each \(K'_j(v_j)\) is a function of player j’s true valuation \(v_j\), \(\mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K_i), (v_{-i}, K'_{-i}(v_{-i}))) \ge \mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K'_i), (v_{-i}, K'_{-i}(v_{-i})))\).Footnote 5

It is not hard to see that 2-DST is a refined solution concept of the Bayesian Nash equilibrium. When the inequalities are all strict in the two conditions of 2-DST, it is called two-step strictly dominant strategy truthful (2-SDST). As noted in [7], any 2-SDST mechanism has a unique equilibrium where players reveal their true valuations and their true beliefs. Moreover, truthfulness is the only strategy that survives two steps of iterated elimination of strictly dominated strategies.

3 Single-good auctions with 2-connected knowledge graphs

As a warm up, in this section, we construct an information elicitation mechanism that is nearly optimal for the special case of single-good auctions under a natural structure of the knowledge graph. More precisely, recall that a directed graph is strongly connected if there is a directed path from any node i to any other node \(i'\). Intuitively, in a knowledge graph this means that for any two players Alice and Bob, Alice knows a guy who knows a guy... who knows Bob. Also recall that a directed graph is 2-connected if it remains strongly connected after removing any single node and the adjacent edges. In a knowledge graph, this means there does not exist a crucial player as an “information hub”, without whom the players will split into two parts, with one part having no information about the other. It is easy to see that a knowledge graph being strong connected (or 2-connected respectively) implies it being 1-informed (or 2-informed respectively), but not vice versa. In fact, a graph of n nodes can be \((\lfloor \frac{n}{2}\rfloor -1)\)-informed without being connected.

When the knowledge graph is 2-connected, we construct the information elicitation Myerson mechanism \(\mathcal {M}_{IEM}\) in Mechanism Algorithm 1. Recall that Myerson’s mechanism maps each player i’s reported value \(b_i\) to the (ironed) virtual value, \(\phi _i(b_i; \mathcal {D}_i)\). It runs the second-price mechanism with reserve price 0 on virtual values and maps the resulting “virtual price” back to the winner’s value space, as her price.

Algorithm 1
figure a

\(\mathcal {M}_{IEM}\)

To help understand our mechanism, we illustrate in Fig. 1 the sets of players involved in the first round of Mechanism \(\mathcal {M}_{IEM}\). Informally, \(\mathcal {M}_{IEM}\) starts from a randomly selected player a and computes the virtual values of the players whose distributions are reported by a. In the following rounds, \(\mathcal {M}_{IEM}\) maintains a set of explored players S whose virtual values have been computed and a player \(i^*\) who has the highest virtual value among them. In each round, \(\mathcal {M}_{IEM}\) computes the virtual values of more players using the information reported by S excluding \(i^*\), and updates S and \(i^*\) accordingly. As we will prove, since the knowledge graph is 2-connected, eventually all the players’ virtual values can be reported and used to compute the winner, excluding a. We have the following theorem.

Fig. 1
figure 1

The sets of players involved in the first round of Mechanism \(\mathcal {M}_{IEM}\). The edges in the figure correspond to distributions reported by the players. In each round, the mechanism keeps in S the player with the highest virtual value so far, drops all the other players from S, and adds the players whose distributions are reported for the first time by the dropped ones

Theorem 1

For any single-good auction instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\) where G is 2-connected, \(\mathcal {M}_{IEM}\) is 2-DST and \(Rev(\mathcal {M}_{IEM}(\mathcal {I})) \ge (1-\frac{1}{n})OPT(\hat{{\mathcal {I}}})\).

Before proving the revenue bound, we first prove \(\mathcal {M}_{IEM}\) is 2-DST.

Lemma 1

\(\mathcal {M}_{IEM}\) is 2-DST.

Proof

By the definition of 2-DST, we prove the lemma in two steps.

Claim 1

For any player i, true value \(v_i\), value \(b_i\), knowledge \(K_i\), and strategy subprofile \(s_{-i} = (b_j, K_j)_{j\ne i}\) of the other players, \(\mathbb {E}_{\mathcal {M}_{IEM}}u_i((v_i, K_i), s_{-i}) \ge \mathbb {E}_{\mathcal {M}_{IEM}}u_i((b_i, K_i), s_{-i})\), where the expectation is taken over the mechanism’s random coins.

Proof

First, conditioning on \(a=i\), player i does not get the item and her reported value is not used by the mechanism. Thus \(u_i((v_i, K_i), s_{-i}) = u_i((b_i, K_i), s_{-i}) = 0\) in this case.

Second, we compare the two utilities conditional on \(a\ne i\). Notice that when \(a\ne i\), whether or not player i’s distribution is reported—that is, whether or not \(\mathcal {D}'_i\) is defined— only depends on \(s_{-i}\). Thus \(\mathcal {D}'_i\) is defined under \((v_i, K_i)\) if and only if it is defined under \((b_i, K_i)\).

If \(\mathcal {D}'_i\) is not defined, then \(i\in N'\) at the end of the mechanism, she does not get the item, and her reported value is not used. Therefore \(u_i((v_i, K_i), s_{-i}) = u_i((b_i, K_i), s_{-i}) = 0\) again.

If \(\mathcal {D}'_i\) is defined, then it is defined in the same round of the mechanism under both \((v_i, K_i)\) and \((b_i, K_i)\), which we refer to as round r. Also, \(\mathcal {D}'_i\) is the same in both cases and the mechanism’s execution is the same till this round. Notice that

  1. (1)

    \(\phi _i(\cdot; \mathcal {D}'_i)\) is monotone in its input;

  2. (2)

    i gets the item if and only if \(i = i^*\) in all rounds \(\ell\) with \(\ell \ge r\) and her virtual value is at least 0; and

  3. (3)

    when i gets the item, \(K_i\) is never used by the mechanism and thus does not affect the execution of any round \(\ell\) with \(\ell \ge r\).

Accordingly, the mechanism is monotone in player i’s reported value: if i gets the item by reporting some value, then she still gets it by reporting a higher value. Moreover, when i gets the item, her price in Step 14 is the threshold payment. Following standard characterizations of single-parameter DST mechanisms, it is the best for player i to report her true value \(v_i\). That is, \(u_i((v_i, K_i), s_{-i}) \ge u_i((b_i, K_i), s_{-i})\) when \(\mathcal {D}'_i\) is defined.

Combining the above cases together, we have \(\mathbb {E}_{\mathcal {M}_{IEM}} u_i((v_i, K_i), s_{-i})\ge \mathbb {E}_{\mathcal {M}_{IEM}} u_i((b_i, K_i), s_{-i})\) and Claim 1 holds. \(\square\)

Claim 2

For any player i, true value \(v_i\), true knowledge \(K_i\), knowledge \(K'_i\), and knowledge subprofile \(K'_{-i}(v_{-i}) = (K'_j(v_j))_{j\ne i}\) of the other players, where each \(K'_j(v_j)\) is a function of player j’s true value \(v_j\), \(\mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K_i), (v_{-i}, K'_{-i}(v_{-i}))) \ge \mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K'_i), (v_{-i}, K'_{-i}(v_{-i})))\).

Proof

Similar to Claim 1, conditional on \(a=i\), player i does not get the item no matter what knowledge she reports. Thus

$$\begin{aligned}&\mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} [u_i((v_i, K_i), (v_{-i}, K'_{-i}(v_{-i}))) \mid a=i] \\ =&\mathbb {E}_{v_{-i} \sim \mathcal {D}_{-i}} [u_i((v_i, K'_i), (v_{-i}, K'_{-i}(v_{-i}))) \mid a=i] = 0. \end{aligned}$$

Next, we compare the two utilities conditional on \(a\ne i\). Again similar to Claim 1, \(\mathcal {D}'_i\) is the same under both strategies of i, and the mechanism’s execution is also the same till the round r where \(\mathcal {D}'_i\) is defined (or till the end if \(\mathcal {D}'_i\) is not defined). There are three cases:

  • If \(\mathcal {D}'_i\) is not defined, then neither \(K_i\) nor \(K'_i\) is used by the mechanism, and i has utility 0 under both strategies.

  • If \(\mathcal {D}'_i\) is defined and \(i = i^*\) from round r to the end of the mechanism, then again \(K_i\) and \(K'_i\) are not used. Thus i has the same utility (maybe non-zero) under both strategies.

  • If \(\mathcal {D}'_i\) is defined and \(i\ne i^*\) starting from some round \(r'\ge r\), then i does not get the item under either strategy, thus her utility is 0 under both of them.

In sum, \(\mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K_i), (v_{-i}, K'_{-i}(v_{-i}))) = \mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K'_i), (v_{-i}, K'_{-i}(v_{-i})))\) and reporting her true knowledge does not hurt player i. \(\square\)

Lemma 1 follows directly from Claims 1 and 2. \(\square\)

In Section 6, we add scoring rules to mechanism \(\mathcal {M}_{IEUD}\) to reward the players’ knowledge, so that a player’s utility will be strictly larger when she reports her true knowledge than when she lies. To prove Theorem 1, it remains to show \(\mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{IEM}(\mathcal {I})) \ge (1-\frac{1}{n})OPT(\hat{{\mathcal {I}}})\) under the players’ truthful strategies.

Proof of Theorem 1

The key is to explore the structure of the knowledge graph to make sure that the player with the highest virtual value is found by the mechanism with high probability.

More specifically, arbitrarily fix the player a chosen by the mechanism. Notice that throughout the mechanism, \(N'\) is the set of players \(i\in N\setminus \{a\}\) such that \(\mathcal {D}'_i\) is not defined. We show that \(N'= \emptyset\) at the end of the mechanism. Indeed, since G is 2-connected, the out-degree of a in G is at least 2: otherwise, either a cannot reach any other node in G, or this becomes the case after removing the unique node j with \((a, j)\in G\), contradicting 2-connectedness. Since a reports her true knowledge \(K_a\), we have \(\mid S \mid \ge 2\) in Step 3 and the mechanism does not stop there. Moreover, at the beginning of each round, we have \(\mid S \mid \ge 2\) and thus \(S\setminus \{i^*\}\ne \emptyset\): otherwise \(S'=\emptyset\) in the previous round, and the mechanism would not have reached this round.

Assume, for the sake of contradiction that the mechanism finally reaches a round r where \(N'\ne \emptyset\) at the beginning but \(S' = \emptyset\) in Step 7. Since all players report their true knowledge, by the definition of \(S'\) we have that, in graph G, all neighbors of \(S\setminus \{i^*\}\) are in \(N\setminus N'\). Furthermore, for any player \(i\in (N\setminus N')\setminus S\), all neighbors of i are also in \(N\setminus N'\): indeed, i has been moved from \(N'\) to S and then dropped from S (except player a, whose neighbors are in \(N\setminus N'\) by definition); and when i is dropped from S, all her neighbors in \(N'\) are moved to S. Accordingly, all the edges going from \(N\setminus N'\) to \(N'\) are from player \(i^*\), and G becomes disconnected after removing \(i^*\), again contradicting 2-connectedness. Thus \(S'\ne \emptyset\) in all rounds and \(N'=\emptyset\) in the end, as we wanted to show.

Because all players report their true values and true knowledge, we have \(\mathcal {D}'_{-a} = \mathcal {D}_{-a}\) and \(\phi _i(v_i; \mathcal {D}'_i) = \phi _i(v_i; \mathcal {D}_i)\) for all \(i\ne a\). Letting \(\hat{\mathcal {I}}_a = (N\setminus \{a\}, M, \mathcal {D}_{-a})\), we claim

$$\begin{aligned} \mathbb {E}_{v\sim \mathcal {D}} [Rev(\mathcal {M}_{IEM}(\mathcal {I})) \mid a] = OPT(\hat{\mathcal {I}}_a). \end{aligned}$$
(1)

To see why this is true, note that by construction, in each round the mechanism keeps the player with the highest virtual value in S. Thus, the final player \(i^*\) has the highest virtual value in \(N\setminus \{a\}\), and \(\phi _{second}\) is the second highest virtual value in \(N\setminus \{a\}\). Accordingly, the outcome of Step 14 is the same as that of Myerson’s mechanism on \(\hat{\mathcal {I}}_a\), so is the revenue. Therefore (1) holds.

Finally, it remains to show that, by throwing away a random player a, the mechanism does not lose much revenue. For each player i, letting \(P_i(OPT(\hat{\mathcal {I}}))\) be the expected price paid by i in Myerson’s mechanism under \(\hat{\mathcal {I}}\), we have \(OPT(\hat{\mathcal {I}}) = \sum _{i\in N} P_i(OPT(\hat{\mathcal {I}}))\). Next, consider the following Bayesian mechanism \(\mathcal {M}'\) on \(\hat{\mathcal {I}}_a\): it runs Myerson’s mechanism on \(\hat{\mathcal {I}}\) and then projects the outcome to players \(N\setminus \{a\}\), i.e., the item is unsold and the payment is not collected if a gets the good; otherwise, the outcome is kept the same. It is easy to see that \(\mathcal {M}'\) is DST, thus it cannot generate more revenue than \(OPT(\hat{\mathcal {I}}_a)\). As the expected revenue of \(\mathcal {M}'\) is \(\sum _{i\ne a} P_i(OPT(\hat{\mathcal {I}}))\), then

$$\begin{aligned} OPT(\hat{\mathcal {I}}_a) \ge \mathbb {E}_{\mathcal {D}_{-a}} Rev(\mathcal {M}'(\hat{\mathcal {I}}_a)) = \sum _{i\ne a} P_i(OPT(\hat{\mathcal {I}})). \end{aligned}$$
(2)

Combining (1) and (2), we have

$$\begin{aligned} & \mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{IEM}(\mathcal {I})) = \sum _{a\in N} \frac{1}{n} \mathbb {E}_{v\sim \mathcal {D}} [Rev(\mathcal {M}_{IEM}(\mathcal {I})) \mid a] = \sum _{a\in N} \frac{1}{n} \left( OPT(\hat{\mathcal {I}}_a)\right) \\\ge & \sum _{a\in N} \frac{1}{n} \left( \sum _{i\ne a} P_i(OPT(\hat{\mathcal {I}}))\right) = \left( \sum _{i\in N} \frac{n-1}{n} P_i(OPT(\hat{\mathcal {I}}))\right) = (1-\frac{1}{n}) OPT(\hat{\mathcal {I}}). \end{aligned}$$

Combining Lemma 1, Theorem 1 holds. \(\square\)

4 When everything is known by somebody

When the knowledge graph vector G is k-informed with \(k\ge 1\), both mechanisms in Section 5 of course apply, but we can do better when k gets larger. In fact, we show that, for both unit-demand auctions and additive auctions, as k gets larger, the approximation ratio approaches the best known approximation to OPT by DST Bayesian mechanisms [8, 9, 11].

4.1 Unit-demand auctions

For unit-demand auctions, sequential post-price mechanisms have been constructed by [8, 10]. For information elicitation, if the seller asks the players to report both their values and knowledge, and directly uses the reported distributions in these mechanisms, then a player may want to withhold her knowledge about the other players. By doing so, a player may prevent the seller from selling the items to the others, so that the items are still available when it is her turn to buy.

A simple idea is to partition the players into two groups: a set of reporters who will not receive any item and is only asked to report their knowledge; and a set of potential buyers whose knowledge is never used. It is possible that the reported knowledge may not contain a potential buyer’s value distributions on all items, thus the technical part is to prove that the seller generates a good revenue even though the players’ knowledge is only partially recovered.

Our mechanism \(\mathcal {M}^{k}_{IEUD}\) is simple and intuitive; see Mechanism Algorithm 2, where \(\mathcal {M}_{UD}\) is the Bayesian mechanism of [8] (see Section 6 therein). The probability that each player is assigned to \(N_1\) (reporters) is \(q = 1-(k+1)^{-\frac{1}{k}}\), and the probability to \(N_{2}\) (potential buyers) is \(1-q\). The probability q is chosen to achieve the maximum probability for each distribution to be reported, and the latter is exactly \(\tau _k = \frac{k}{(k+1)^{\frac{k+1}{k}}}\). It’s worth pointing out that, although mechanism \(\mathcal {M}_{UD}\) is used as a black-box, Mechanism Algorithm 2 is not a reduction from arbitrary Bayesian mechanisms. We summarize our result in Theorem 2.

Algorithm 2
figure b

. \(\mathcal {M}^{k}_{IEUD}\)

Theorem 2

\(\forall k\in [n-1]\), any unit-demand auction instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\) where G is k-informed, mechanism \(\mathcal {M}^{k}_{IEUD}\) is 2-DST and \(Rev(\mathcal {M}^{k}_{IEUD}(\mathcal {I})) \ge \frac{\tau _k}{24} \cdot OPT(\hat{\mathcal {I}}).\)

Note that \(\tau _k\) is increasing in k, \(\tau _1 = \frac{1}{4}\) and \(\tau _k\rightarrow 1\). As k gets larger, the approximation ratio approaches 24, the best known approximation to OPT by DST Bayesian mechanisms [8, 9].

To prove Theorem 2, the 2-DSTness follows from Lemma 2 below. Regarding the revenue bound, in the next section, we will prove a more general result where k may be 0 (see Theorem 4), and omit a formal proof of the bound in Theorem 2. To understand the approximation ratio, blow we show that under the players’ truthful strategies, the probability for each distribution \(\mathcal {D}_{ij}\) to be reported in mechanism \(\mathcal {M}'_{IEUD}\) is at least \(\tau _{k}\). Then following the proof of Theorem 4, we have the revenue bound of the mechanism in Theorem 2.

Indeed, for any player i and item j,

$$\begin{aligned} & \Pr (\mathcal {D}_{ij} \text { is reported in the mechanism}) \\= & \Pr (i\in N_2)\Pr (\exists i'\in N_1, (i, i')\in G_j \ \mid \ i \in N_2) \ge (1-q)(1-(1-q)^k), \end{aligned}$$

where the inequality is because \(\mathcal {D}_{ij}\) is known by at least k players other than i, and the players are partitioned independently. Taking derivatives of the last term, we have that it is maximized when \(q=1-(k+1)^{-\frac{1}{k}}\) as in the mechanism, in which case

$$\Pr (\mathcal {D}_{ij} \text { is reported in the mechanism})\ge \frac{1}{(k+1)^{\frac{1}{k}}}\cdot \frac{k}{k+1} = \tau _k.$$

It remains to prove the 2-DSTness.

Lemma 2

Mechanism \(\mathcal{M}^k_{IEUD}\) is 2-DST.

Proof

We start with the first requirement in the solution concept: it is best for a player to report her true values, no matter what knowledge she reports and what strategies the others use.

Claim 3

For any player i, true valuation \(v_i\), valuation \(b_i\), knowledge \(K_i\), and strategy subprofile \(s_{-i} = (b_j, K_j)_{j\ne i}\) of the other players, \(\mathbb {E}_{\mathcal {M}^k_{IEUD}}u_i((v_i, K_i), s_{-i}) \ge \mathbb {E}_{\mathcal {M}^k_{IEUD}}u_i((b_i, K_i), s_{-i})\), where the expectation is taken over the mechanism’s random coins.

Proof

If player i is not in \(N_3\) given the mechanism’s randomness and the reported knowledge of all players, then her reported valuation is never used to compute his allocation or price, and she gets the same utility for reporting any valuation. Thus \(u_{i}((v_{i},K_{i}), s_{-i})=u_{i}((b_{i},K_{i}), s_{-i})\) conditional on \(i\notin N_3\).

If player i is in \(N_3\), then her utility is determined by \(\mathcal {M}_{UD}\). If \(\mathcal {D}'_{ij}\) is defined to be 0 with probability 1 in Step 7, then i’s reported value for item j is not given to \(\mathcal {M}_{UD}\) as input, and i gets the same utility for reporting any value for j, including \(v_{ij}\). Moreover, because i does not get such an item j, his utility is the same as a dummy player \(\hat{i}\) whose valuation is the same as i’s, except that the true value of \(\hat{i}\) for j is 0. Since \(\mathcal {M}_{UD}\) is DST, no matter what the distributions are and what values the other players report, it is the best for \(\hat{i}\) to report her true valuation. Accordingly, it is the best for i to report her true valuation \(v_i\) as well. That is, \(u_i((v_i, K_i), s_{-i}) \ge u_i((b_i, K_i), s_{-i})\) conditional on \(i\in N_3\).

Combining these two cases, Claim 3 holds. \(\square\)

We now prove the second requirement in the solution concept: given that all players report their true valuations, it is best for a player to report her true knowledge.

Claim 4

For any player i, true valuation \(v_i\), true knowledge \(K_i\), knowledge \(K'_i\), and knowledge subprofile \(K'_{-i}(v_{-i}) = (K'_j(v_j))_{j\ne i}\) of the other players, where each \(K'_j(v_j)\) is a function of player j’s true valuation \(v_j\), \(\mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K_i), (v_{-i}, K'_{-i}(v_{-i}))) \ge \mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K'_i), (v_{-i}, K'_{-i}(v_{-i})))\).

Proof

If player i is in \(N_1\), then she is guaranteed to get no item and pay 0, so her utility is 0 no matter which knowledge she reports. If player i is in \(N_2\), then her knowledge is never used, and she again gets the same utility no matter which knowledge she reports. Thus \(\mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K_i), (v_{-i}, K'_{-i}(v_{-i}))) = \mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} u_i((v_i, K'_i), (v_{-i}, K'_{-i}(v_{-i})))\) and Claim 4 holds. \(\square\)

Lemma 2 follows directly from Claims 3 and 4. \(\square\)

4.2 Additive auctions

Information elicitation mechanisms for additive auctions are harder to construct and analyze than for unit-demand auctions. First, randomly partitioning the players as before may cause a significant revenue loss, as the revenue of additive auctions may come from selling a subset of items as a bundle to a player i. Even when i’s value distribution for each item is reported with constant probability, the probability that her distributions for all items in the bundle are reported may be very low, thus the mechanism may rarely sell the bundle to i at the optimal price. Second, the seller can no longer “throw away” player-item pairs whose distributions are not reported and focus on the projected instance. When the players are not partitioned into reporters and potential buyers, doing so may cause a player to lie and withhold her knowledge about others, so that they are thrown away.

To simultaneously achieve truthfulness and a good revenue guarantee, our mechanism is very stingy and never throws away any information. Following [9] (see Section 7 therein), we can consider a mechanism consisting of a “bundling part” and an “individual sale part”. The former is referred to as the Bundle VCG mechanism, denoted by BVCG; and the latter is the individual 1-lookahead mechanism, denoted by \(\mathcal {M}_{1LA}\), which sells each item separately using the 1-lookahead mechanism of [60]. Mechanism \(\mathcal {M}_{1LA}\) can also be replaced by the individual Myerson mechanism, denoted by IM, which sells each item separately using Myerson’s mechanism. By choosing the mechanism that generates a higher expected revenue between IM and BVCG, [9] provides a Bayesian mechanism that is an 8-approximation to OPT.

We design corresponding information elicitation mechanisms separately for mechanisms BVCG and IM. The resulting mechanisms are denoted by \(\mathcal {M}_{IEBVCG}\) and \(\mathcal {M}_{IEIM}\), which are formally defined later. Because the seller does not know the prior \(\mathcal {D}\), she cannot compute the expected revenue of the two information elicitation mechanisms and choose the better one. Instead, we let her choose between the two mechanisms randomly, according to a probability distribution depending on k. It is worth pointing out that when \(k\ge 1\), mechanism \(\mathcal {M}_{IEBVCG}\) is able to recover all distributions in \(\hat{\mathcal {I}}\), thus its revenue equals the corresponding Bayesian revenue, which is lower-bounded in [9]. For \(\mathcal {M}_{IEIM}\), information elicitation is done by randomized partition depending on the value of k.

However, we can do even better. Indeed, although in Bayesian auctions the mechanism IM is optimal for individual item-sale and outperforms mechanism \(\mathcal {M}_{1LA}\), in information elicitation auctions there is a tradeoff between the two. In order for the players to report their knowledge truthfully for mechanism IM, we need to randomly partition them into reporters and potential buyers, thus each distribution is only recovered with probability \(\tau _k\). In contrast, no partition is needed for aggregating the players’ knowledge in mechanism \(\mathcal {M}_{1LA}\), and we can recover all distributions simultaneously with probability 1. The resulting information elicitation mechanism, \(\mathcal {M}_{IE1LA}\), is also defined later. As mechanism \(\mathcal {M}_{1LA}\) is a 2-approximation to mechanism IM, sometimes it is actually more advantageous to use \(\mathcal {M}_{IE1LA}\) rather than \(\mathcal {M}_{IEIM}\), depending on the value of k.

Properly combining the above gadgets together, our mechanism \(\mathcal {M}'_{IEA}\) is defined as follows: when \(k\le 7\), it runs \(\mathcal {M}_{IEBVCG}\) with probability \(\frac{2}{11}\) and \(\mathcal {M}_{IE1LA}\) with probability \(\frac{9}{11}\); when \(k> 7\), it runs \(\mathcal {M}_{IEBVCG}\) with probability \(\frac{\tau _k}{3+\tau _k}\) and \(\mathcal {M}_{IEIM}\) with probability \(\frac{3}{3+\tau _k}\). The choice of the two cases is to achieve the best approximation ratio for each k. We have the following theorem.

Theorem 3

\(\forall k \in [n-1]\), any additive auction instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\) where G is k-informed, \(\mathcal {M}'_{IEA}\) is 2-DST and \(Rev(\mathcal {M}'_{IEA}(\mathcal {I})) \ge \max \{\frac{1}{11}, \frac{\tau _k}{6+2\tau _k}\} OPT(\hat{\mathcal {I}})\).

To prove Theorem 3, we first formally define mechanisms \(\mathcal {M}_{IEIM}\), \(\mathcal {M}_{IE1LA}\) and \(\mathcal {M}_{IEBVCG}\).

The information elicitation individual myerson mechanism

We start by introducing the information elicitation individual Myerson mechanism \(\mathcal {M}_{IEIM}\), which runs the following mechanism \(\mathcal {M}_{IEIM,j}\) for each item j separately. Mechanism \(\mathcal {M}_{IEIM,j}\) is similar to \(\mathcal {M}^{k}_{IEUD}\), thus we have omitted many details in the analysis.

Algorithm 3
figure c

\(\mathcal {M}_{IEIM,j}\)

For each item j, let \(v_j = (v_{ij})_{i\in N}\), \(\mathcal {D}_j = (\mathcal {D}_{ij})_{i\in N}\), \(\hat{\mathcal {I}}_j = (N, \{j\}, \mathcal {D}_j)\) be the corresponding single-good Bayesian instance, and \(\mathcal {I}_j = (N, \{j\}, \mathcal {D}_j, G_j)\) be the corresponding single-good information elicitation instance. Lemma 3 below is similar to Lemma 2 and we provide its statement only.

Lemma 3

For any additive auction instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\), mechanism \(\mathcal {M}_{IEIM,j}\) is 2-DST for \(\mathcal {I}_j\) for each \(j\in M\), and mechanism \(\mathcal {M}_{IEIM}\) is 2-DST for \(\mathcal {I}\).

Next, we consider the expected revenue of \(\mathcal {M}_{IEIM}\).

Lemma 4

\(\mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{IEIM}({\mathcal {I}})) \ge \tau _k \mathbb {E}_{v\sim \mathcal {D}} Rev(IM(\hat{{\mathcal {I}}}))\).

Proof

By definition,

$$\begin{aligned} \mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{IEIM}({\mathcal {I}})) = \sum _{j\in M} \mathbb {E}_{v_j \sim \mathcal {D}_j} Rev(\mathcal {M}_{IEIM, j}(\mathcal {I}_j)) \end{aligned}$$

and

$$\begin{aligned} \mathbb {E}_{v\sim \mathcal {D}} Rev(IM(\hat{{\mathcal {I}}})) = \sum _{j\in M} OPT(\hat{\mathcal {I}}_j). \end{aligned}$$

Accordingly, it suffices to show that for each item j, \(\mathbb {E}_{v_j \sim \mathcal {D}_j} Rev(\mathcal {M}_{IEIM, j}(\mathcal {I}_j)) \ge \tau _k OPT(\hat{\mathcal {I}}_j)\), and we have

$$\begin{aligned} & \mathbb {E}_{v_j \sim \mathcal {D}_j} Rev(\mathcal {M}_{IEIM, j}(\mathcal {I}_j)) \nonumber \\= & \underset{N_3}{\mathbb {E}} \ \underset{v_{N_3, j}\sim {\mathcal {D}_{N_{3}, j}}}{\mathbb {E}} OPT(\hat{{\mathcal {I}}}_{N_3, j}) \ge \underset{N_3}{\mathbb {E}}\ \underset{v_j\sim \mathcal {D}_j}{\mathbb {E}} OPT(\hat{{\mathcal {I}}}_j)_{N_3} \nonumber \\= & \underset{N_3}{\mathbb {E}}\ \underset{v_j\sim \mathcal {D}_j}{\mathbb {E}} \sum \limits _{i\in N_3}P_i(OPT(\hat{{\mathcal {I}}}_j)) = \underset{v_j\sim \mathcal {D}_j}{\mathbb {E}} \ \underset{N_3}{\mathbb {E}} \sum \limits _{i\in N_3} P_i(OPT(\hat{{\mathcal {I}}}_j)) \nonumber \\= & \underset{v_j \sim \mathcal {D}_j}{\mathbb {E}} \sum \limits _{i}\Pr (i\in N_3) \cdot P_i(OPT(\hat{{\mathcal {I}}}_j)) \ge \tau _k \underset{v_j \sim \mathcal {D}_j}{\mathbb {E}} \sum \limits _{i} P_i(OPT(\hat{{\mathcal {I}}}_j)) \nonumber \\= & \tau _k OPT(\hat{{\mathcal {I}}}_j), \end{aligned}$$

as desired. Thus Lemma 4 holds. \(\square\)

The information elicitation individual 1-lookahead mechanism

Next, we introduce the information elicitation 1-lookahead mechanism \(\mathcal {M}_{IE1LA}\), which runs the following mechanism \(\mathcal {M}_{IE1LA, j}\) for each item j separately. We will show that the revenue of \(\mathcal {M}_{IE1LA}\) matches that of mechanism \(\mathcal {M}_{1LA}\) for any \(k\ge 1\).

Algorithm 4
figure d

\(\mathcal {M}_{IE1LA, j}\)

Note that \(\mathcal {M}_{IE1LA, j}\) does not partition the players into two groups. Also, it is not exactly using the 1-lookahead mechanism as a blackbox, because it has to handle boundary cases where the players’ distributions are not all reported. When the players all tell the truth, all true distributions will indeed be reported. However, for the mechanism to be well defined, it has to know what to do in all possible cases. Moreover, running the 1-lookahead mechanism on the set of players whose distributions are reported is not 2-DST: for example, if the player with the second highest value is the only one who knows the distribution for the player with the highest value, then she may choose not to report her knowledge about the latter, so that she himself has the highest value in the 1-Lookahead mechanism and gets a high utility. That is why the mechanism only tries to sell to the player with the highest value. We have the following two lemmas.

Lemma 5

For any additive auction instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\), mechanism \(\mathcal {M}_{IE1LA,j}\) is 2-DST for each \(\mathcal {I}_j\), and mechanism \(\mathcal {M}_{IE1LA}\) is 2-DST for \(\mathcal {I}\).

Proof

As in mechanism \(\mathcal {M}_{IEA}\), in each mechanism \(\mathcal {M}_{IE1LA,j}\), the fact that it is dominant for each player i to report her true value no matter what knowledge the players report follows from the truthfulness of the second-price mechanism and that of the 1-lookahead mechanism. Given that all players report their true values, a player i’s reported knowledge does not affect whether she is \(i^*\) or not. It may affect the other players’ utilities, but not her own. Thus reporting her true knowledge never hurts him, and mechanism \(\mathcal {M}_{IE1LA,j}\) is 2-DST for \(\mathcal {I}_j\).

Since the players have additive valuations and \(\mathcal {M}_{IE1LA}\) runs each mechanism \(\mathcal {M}_{IE1LA, j}\) separately for item j, we have that \(\mathcal {M}_{IE1LA}\) is 2-DST for \(\mathcal {I}\) and Lemma 5 holds. \(\square\)

Lemma 6

\(\mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{IE1LA}({\mathcal {I}})) = \mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{1LA}(\hat{{\mathcal {I}}})) \ge \frac{1}{2} \mathbb {E}_{v\sim \mathcal {D}} Rev(IM(\hat{{\mathcal {I}}}))\).

Proof

When the players report their true values and true knowledge, the outcome of each \(\mathcal {M}_{IE1LA, j}\) on the information elicitation instance \(\mathcal {I}_j\) is the same as that of mechanism \(\mathcal {M}_{1LA}\) on the Bayesian instance \(\hat{\mathcal {I}}_j\), because the distribution for \(i^*\) is reported. Accordingly,

$$\begin{aligned} \mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{IE1LA}({\mathcal {I}}))= & \sum \limits _{j\in M} \mathbb {E}_{v_j\sim \mathcal {D}_j} Rev(\mathcal {M}_{IE1LA, j}({\mathcal {I}_j})) = \mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{1LA}(\hat{{\mathcal {I}}})) \\\ge & \sum \limits _{j\in M} \dfrac{1}{2}OPT(\hat{{\mathcal {I}}}_j) = \dfrac{1}{2} \mathbb {E}_{v\sim \mathcal {D}} Rev(IM(\hat{{\mathcal {I}}})), \end{aligned}$$

where the inequality is because the 1-lookahead mechanism is a 2-approximation to the optimal Bayesian mechanism for each item j [60]. Thus Lemma 6 holds. \(\square\)

Note that the approximation ratio of \(\mathcal {M}_{IE1LA}\) does not depend on the specific value of k, as long as \(k\ge 1\).

The information elicitation BVCG mechanism

The mechanism \(\mathcal {M}_{IEBVCG}\) is defined in Mechanism Algorithm 5. It is similar to \(\mathcal {M}_{IEA}\) and approximates mechanism BVCG in information elicitation settings. If a player i’s value distributions are not all reported, \(\mathcal {M}_{IEBVCG}\) throws i away and leaves his winning set unsold. This simplifies the instructions compared to \(\mathcal {M}_{IEA}\) and still ensures truthfulness. Doing so would seriously damage the revenue if the knowledge graphs can be totally arbitrary. However, when everything is known by somebody and when the players report their true knowledge, no player is actually thrown away. We have the following two lemmas.

Algorithm 5
figure e

\(\mathcal {M}_{IEBVCG}\)

Lemma 7

Mechanism \(\mathcal {M}_{IEBVCG}\) is 2-DST.

Proof

Arbitrarily fix a player i, a strategy subprofile of the other players, and a knowledge of i. If not all m distributions of i’s values are reported by the others, then i gets nothing and pays nothing, so it does not matter what valuation she reports about himself. Otherwise, \(\mathcal {M}_{IEBVCG}\) sells to player i in the same way as BVCG: using the other players’ highest reported value as the reserve price for each item, either player i gets the whole set of items for which her value passes the reserve price (i.e., her winning set), or she gets nothing and those items are unsold to anybody. Following [11], it is dominant for i to report her true values given any entry fee that does not depend on her reported values, so is it when the entry fee is computed based on \(\mathcal {D}'_i\) and \(b_{-i}\).

Moreover, a player i’s reported knowledge \(K_i\) about others affects neither \(M_i\) nor \(e_i\), nor the reserve prices for him. Thus reporting her true knowledge never hurts her and Lemma 7 holds. \(\square\)

Lemma 8

\(\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(\mathcal {M}_{IEBVCG}({\mathcal {I}})) = \underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(BVCG(\hat{{\mathcal {I}}}))\).

Proof

Since \(\mathcal {M}_{IEBVCG}\) retrieves the whole distribution \(\mathcal {D}\) from the players, its outcome is exactly the same as that of BVCG under the Bayesian instance \(\hat{\mathcal {I}}\). \(\square\)

Remark

Similar to \(\mathcal {M}_{IE1LA}\), the revenue of \(\mathcal {M}_{IEBVCG}\) does not depend on the specific value of k, as long as \(k\ge 1\). Indeed, notice the special structures of the two Bayesian mechanisms \(\mathcal {M}_{1LA}\) and BVCG: the winning set of a player i solely depends on the players’ values; the distribution \(\mathcal {D}_i\) is only used to compute better reserve prices or entry fee to increase revenue; and the distribution \(\mathcal {D}_{-i}\) is irrelevant to i. Therefore, in the information elicitation setting we can allow a player to be both a reporter about the others’ distributions and a potential buyer of some items. In some other mechanisms such as Myerson’s mechanism, all players’ distributions are used both to choose the potential winner and to set her price, thus in the information elicitation setting we must separate the knowledge reporters and the potential winners.

We are now ready to prove Theorem 3.

Proof of Theorem 3

Recall that mechanism \(\mathcal {M}'_{IEA}\) is defined as follows: when \(k\le 7\), it runs \(\mathcal {M}_{IEBVCG}\) with probability \(\frac{2}{11}\) and \(\mathcal {M}_{IE1LA}\) with probability \(\frac{9}{11}\); when \(k> 7\), it runs \(\mathcal {M}_{IEBVCG}\) with probability \(\frac{\tau _k}{3+\tau _k}\) and \(\mathcal {M}_{IEIM}\) with probability \(\frac{3}{3+\tau _k}\).

The mechanism \(\mathcal {M}'_{IEA}\) is clearly 2-DST, since all the sub-mechanisms are 2-DST, and which mechanism is chosen does not depend on the players’ strategies. When \(k\le 7\), we have \(\frac{\tau _k}{6+2\tau _k} < \frac{1}{11}\). By Lemmas 6 and 8,

$$\begin{aligned} & \underset{v\sim \mathcal {D}}{\mathbb {E}} Rev(\mathcal {M}'_{IEA}( {\mathcal {I}})) = \dfrac{2}{11}\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(\mathcal {M}_{IEBVCG}( {\mathcal {I}})) + \dfrac{9}{11}\underset{v\sim \mathcal {D}}{\mathbb {E}} Rev(\mathcal {M}_{IE1LA}( {\mathcal {I}})) \nonumber \\\ge & \dfrac{2}{11}\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(BVCG(\hat{{\mathcal {I}}})) + \frac{3}{11} \underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(IM( \hat{{\mathcal {I}}})) + \dfrac{3}{11}\underset{v\sim \mathcal {D}}{\mathbb {E}} Rev(\mathcal {M}_{1LA}( {\hat{\mathcal {I}}})). \end{aligned}$$
(3)

When \(k>7\), we have \(\frac{\tau _k}{6+2\tau _k}> \frac{1}{11}\). By Lemmas 4 and 8,

$$\begin{aligned} & \underset{v\sim \mathcal {D}}{\mathbb {E}} Rev(\mathcal {M}'_{IEA}( {\mathcal {I}})) \end{aligned}$$
(4)
$$\begin{aligned}= & \frac{\tau _k}{3+\tau _k}\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(\mathcal {M}_{IEBVCG}( {\mathcal {I}})) + \frac{3}{3+\tau _k}\underset{v\sim \mathcal {D}}{\mathbb {E}} Rev(\mathcal {M}_{IEIM}( {\mathcal {I}})) \nonumber \\\ge & \frac{\tau _k}{3+\tau _k}\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(BVCG(\hat{{\mathcal {I}}})) + \frac{3\tau _k}{3+\tau _k} \underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(IM( \hat{{\mathcal {I}}})). \end{aligned}$$
(5)

By [9],

$$\begin{aligned} OPT(\hat{\mathcal {I}})\le & 2\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(BVCG(\hat{{\mathcal {I}}})) + 3\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(IM(\hat{{\mathcal {I}}})) + 3\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(\mathcal {M}_{1LA}(\hat{{\mathcal {I}}})) \\\le & 2\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(BVCG(\hat{{\mathcal {I}}})) + 6\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(IM(\hat{{\mathcal {I}}})). \end{aligned}$$

Combined the first inequality with Inequality (3) and the second with Inequality (5), we have

$$\begin{aligned} \underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(\mathcal {M}'_{IEA}({\mathcal {I}})) \ge \max \left\{ \frac{1}{11}, \dfrac{\tau _k}{6+2\tau _k}\right\} \cdot OPT(\hat{{\mathcal {I}}}), \end{aligned}$$

and Theorem 3 holds. \(\square\)

Note that when the knowledge graphs are 2-connected, instead of using mechanism \(\mathcal {M}_{IE1LA}\) or \(\mathcal {M}_{IEIM}\), one can use \(\mathcal {M}_{IEM}\) for each item j. We thus have the following corollary, where the mechanism \(\mathcal {M}''_{IEA}\) runs \(\mathcal {M}_{IEM}\) with probability \(\frac{3}{4}\) and \(\mathcal {M}_{IEBVCG}\) with probability \(\frac{1}{4}\).

Corollary 1

For any additive auction instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\) where each \(G_j\) is 2-connected, mechanism \(\mathcal {M}''_{IEA}\) is 2-DST and \(\mathbb {E}_{\mathcal {D}}Rev(\mathcal {M}''_{IEA}({\mathcal {I}})) \ge \frac{1}{8}(1-\frac{1}{n}) OPT(\hat{{\mathcal {I}}})\).

5 Under arbitrary knowledge graphs

5.1 Knowledge-based revenue benchmark

When the knowledge graphs can be totally arbitrary, some distributions may not be known by anybody. It is not hard to see that in this case, no information elicitation mechanism can be a bounded approximation to OPT. Indeed, if all but one value distributions of the players are constantly 0, and if the only non-zero distribution, denoted by \(\mathcal {D}_{ij}\), is unknown by anybody, then a Bayesian mechanism can find the optimal reserve price based on \(\mathcal {D}_{ij}\), while an information elicitation mechanism can only set the price for player i based on the reported values of the other players, which are all 0.

Thus, for arbitrary knowledge graphs, we define a natural revenue benchmark: the optimal Bayesian revenue on players and items for which the distributions are known in the information elicitation setting. More precisely, let \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) be a Bayesian instance and \(\mathcal {I}= (N, M, \mathcal {D}, G)\) a corresponding information elicitation instance. Let \(\mathcal {D}'=\times _{i\in N, j\in M} \mathcal {D}'_{ij}\) be such that \(\mathcal {D}'_{ij} = \mathcal {D}_{ij}\) if there exists a player \(i'\) with \((i', i)\in G_j\), and \(\mathcal {D}'_{ij}\) is constantly 0 otherwise. We refer to \(\mathcal {D}'\) as \(\mathcal {D}\) projected on G. Letting \(\mathcal {I}' = (N, M, \mathcal {D}')\) be the resulting Bayesian instance, the knowledge-based revenue benchmark is \(OPT_K(\mathcal {I}) \triangleq OPT(\mathcal {I}')\), the optimal BIC revenue on \(\mathcal {I}'\). This is a demanding benchmark in information elicitation settings: it takes into consideration the knowledge of all players, no matter who knows what. When everything is known by somebody, even if G is only 1-informed, we will have \(\mathcal {I}' = \hat{\mathcal {I}}\) and \(OPT_K(\mathcal {I}) = OPT(\hat{\mathcal {I}})\).

5.2 Unit-demand auctions

When some distributions may not be known by anybody, we design mechanism \(\mathcal {M}_{IEUD}\) which is very similar as \(\mathcal {M}^k_{IEUD}\); see Mechanism Algorithm 6, where \(\mathcal {M}_{UD}\) is also the Bayesian mechanism of [8]. It’s worth pointing out that, although mechanism \(\mathcal {M}_{UD}\) is used as a black-box, Mechanism Algorithm 6 is not a reduction from arbitrary Bayesian mechanisms. Instead, we will prove a projection lemma that allows such a reduction from an important class of Bayesian mechanisms, where mechanism \(\mathcal {M}_{UD}\) is an important example. We have the following theorem.

Algorithm 6
figure f

\(\mathcal {M}_{IEUD}\)

Theorem 4

Mechanism \(\mathcal {M}_{IEUD}\) for unit-demand auctions is 2-DST and, for any instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\), \(Rev(\mathcal {M}_{IEUD}(\mathcal {I})) \ge \frac{OPT_K(\mathcal {I})}{96}\).

The 2-DSTness follows from Lemma 2, and in the following we prove the revenue bound.

The COPIES setting

Before analyzing the revenue of mechanism \(\mathcal{M}_{IEUD}\), we first recall the COPIES setting for reducing multi-parameter settings to single-parameter settings [10]. Given a unit-demand auction instance \(\hat{\mathcal{I}}=(N, M, \mathcal {D})\), the corresponding COPIES instance is constructed as follows. We make m copies for each player, called player copies, and denote the resulting player set by \(N^{CP}=N\times M\). We make n copies for each item, called item copies, and denote the resulting item set by \(M^{CP}=M\times N\). Each player copy (ij) has value \(v_{ij}\sim \mathcal {D}_{ij}\) for the item copy (ji), and 0 for all the other item copies.

The set of feasible allocations in the original unit-demand auction naturally defines the set of feasible allocations in the COPIES auction: for any feasible allocation A in the original setting, if player i gets item j, then in the corresponding allocation in the COPIES setting, player i’s copy j gets item j’s copy i, and all other copies of i get nothing. Since in the original setting each item is sold to at most one player, in the COPIES setting, for all copies of the same item, at most one of them is sold. Moreover, since each player gets one item in the original setting, in the COPIES setting, for all copies of the same player, at most one of them gets an item copy. We denote by \(\hat{\mathcal {I}}^{CP} = (N^{CP}, M^{CP}, \mathcal {D}^{CP})\) the COPIES instance, with \(\mathcal {D}^{CP} = \times _{(i,j)\in N^{CP}} \mathcal {D}_{ij}\).

The projected setting

Next, we consider the optimal Bayesian revenue for the COPIES setting when “projected” to smaller instances. Let \(NM\subseteq N\times M\) be a subset of player-item pairs and \(N'\) be NM projected to N. The unit-demand instance \(\hat{\mathcal {I}}_{NM} = (N', M, \mathcal {D}'_{N'})\), referred to as \(\hat{\mathcal {I}}\) projected to NM, is such that \(\mathcal {D}'_{ij} = \mathcal {D}_{ij}\) if \((i, j)\in NM\), and \(\mathcal {D}'_{ij}\equiv 0\) otherwise.

Let \(\hat{\mathcal {I}}^{CP}_{NM} = (N'^{CP}, M^{CP}, \mathcal {D}'^{CP}_{N'})\) be the COPIES instance corresponding to \(\hat{\mathcal {I}}_{NM}\). It can also be considered as \(\hat{\mathcal {I}}^{CP}\) projected to NM. That is, when projecting a COPIES instance to a set of player-item pairs, we still want the resulting instance to be a COPIES instance, thus we patch it up with the missing player-item pairs but with values constantly 0. The relations of these instances are illustrated in Fig. 2. We are interested in the optimal Bayesian revenue under \(\hat{\mathcal {I}}^{CP}_{NM}\), \(OPT(\hat{\mathcal {I}}^{CP}_{NM})\).

Fig. 2
figure 2

Relations of the COPIES and the projected instances

Let \(OPT(\hat{\mathcal {I}}^{CP})_{NM}\) be the revenue of the optimal Bayesian mechanism for \(\hat{\mathcal {I}}^{CP}\) obtained from NM: that is, for any pair \((i, j)\notin NM\), the price paid by the j-th copy of player i is not counted, even though this player copy may get the i-th copy of item j according to the optimal mechanism for \(\hat{\mathcal {I}}^{CP}\). We have the following lemma, where we explicitly write out the distributions for different Bayesian instances.

Lemma 9

(The projection lemma) For any \(\hat{\mathcal {I}}\) and \(NM\subseteq N\times M\), \(OPT(\hat{\mathcal {I}}^{CP}_{NM}) \ge OPT(\hat{\mathcal {I}}^{CP})_{NM}\).

Proof

Given the instance \(\hat{\mathcal{I}}_{NM}^{CP}\), consider a Bayesian mechanism \(\mathcal {M}'\) as follows:

  • Patch up the instance to be exactly \(\hat{\mathcal {I}}^{CP}\), including changing the distribution \(\mathcal {D}'_{ij}\) from 0 to \(\mathcal {D}_{ij}\) for any (ij) with \(i\in N'\) and \((i, j)\notin NM\);

  • For any \((i, j)\in NM\), let i’s copy j report her value, while for any \((i, j)\not \in NM\), sample the value of i’s copy j from \(\mathcal {D}_{ij}\);

  • Run the optimal DST Bayesian mechanism on \(\hat{\mathcal {I}}^{CP}\), with the reported and the sampled values;Footnote 6

  • Project the resulting outcome to NM.

The key is to show that \(\mathcal {M}'\) is a DST Bayesian mechanism for \(\hat{\mathcal{I}}_{NM}^{CP}\). Indeed, for any (ij) such that \(i\in N'\) and \((i, j)\notin NM\), the value of i’s copy j in \(\hat{\mathcal{I}}_{NM}^{CP}\) is 0 with probability 1, his reported value is not used by \(\mathcal {M}'\), and at the end this player copy gets nothing and pays 0. Therefore it is dominant for this player copy to report her true value 0. For any \((i, j)\in NM\), the utility of i’s copy j under \(\mathcal {M}'\) is the same as that under the optimal DST Bayesian mechanism on \(\hat{\mathcal {I}}^{CP}\). As it is dominant for this player copy to report her true value in the latter, so is it in \(\mathcal {M}'\). Accordingly,

$$\begin{aligned} \underset{\mathcal {D}'^{CP}_{N'}}{\mathbb {E}} OPT(\hat{\mathcal{I}}_{NM}^{CP}) \ge \underset{\mathcal {D}'^{CP}_{N'}}{\mathbb {E}} Rev(\mathcal {M}'(\hat{\mathcal{I}}_{NM}^{CP})), \end{aligned}$$

by the definition of OPT. By construction we have

$$\begin{aligned} \underset{\mathcal {D}'^{CP}_{N'}}{\mathbb {E}} Rev(\mathcal {M}'(\hat{\mathcal{I}}_{NM}^{CP})) = \underset{\mathcal {D}^{CP}}{\mathbb {E}} OPT(\hat{\mathcal {I}}^{CP})_{NM}, \end{aligned}$$

thus Lemma 9 holds. \(\square\)

Note that the projection lemma is not true for non-COPIES settings in general, because the corresponding mechanism \(\mathcal {M}'\) is not DST. Indeed, when the same player i has \((i, j)\in NM\) and \((i, j')\notin NM\) for some items j and \(j'\), she prefers receiving j to receiving \(j'\) in the patched-up auction, even if the former leads to a smaller utility: her projected utility will be 0 under the latter.

Now we are ready to finish the proof of Theorem 4.

Proof of Theorem 4

Letting \(\mathcal {I}' = (N, M, \mathcal {D}')\) be the Bayesian instance where \(\mathcal {D}'\) is \(\mathcal {D}\) projected on the knowledge graph G, the COPIES instance \(\mathcal {I}'^{CP}\) is respectively defined. Following Lemma 2 and by the definition of \(OPT_K\), it remains to show

$$\mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{IEUD}(\mathcal {I})) \ge \frac{OPT(\mathcal {I}')}{96}.$$

For any player i and item j with \(\mathcal {D}_{ij}\) known by some other players, we have

$$\begin{aligned} & \Pr (\mathcal {D}_{ij} \text { is reported in the mechanism})\\= & \Pr (i\in N_2)\Pr (\exists i'\in N_1 \text{ s.t. } (i', i)\in G_j \ \mid \ i \in N_2) = \frac{1}{4}, \end{aligned}$$

where the equality is because there exists at least one player \(i'\) with \((i', i)\in G_j\), and players i and \(i'\) are partitioned independently. Below we use \(NM_3\) to denote the set of player-item pairs whose distribution is reported in the mechanism: that is, the set of players \(N_3\) together with their reported items. Accordingly,

$$\begin{aligned} \Pr ((i, j)\in NM_3) \ge \frac{1}{4}. \end{aligned}$$
(6)

Also, we use \(\mathcal {D}_{NM_3} = \times _{(i,j)\in NM_3} \mathcal {D}_{ij}\) and \(v_{NM_3} = (v_{ij})_{(i,j)\in NM_3}\) to denote the vector of distributions and the vector of true values for \(NM_3\), respectively. Let \(\hat{\mathcal {I}}_{NM_3} = (N_3, M, \mathcal {D}'_{N_3})\) be the unit-demand instance given to \(\mathcal {M}_{UD}\) in Step 8. Note that it is exactly \(\mathcal {I}'\) projected to \(NM_3\): that is, \(\mathcal {D}'_{ij} = \mathcal {D}_{ij}\) for any \((i, j)\in NM_3\) and \(\mathcal {D}'_{ij}\equiv 0\) for any \((i, j)\notin NM_3\). Accordingly,

$$\begin{aligned} \underset{v\sim \mathcal {D}}{\mathbb {E}}\ Rev(\mathcal{M}_{IEUD}(\mathcal{I}))= & \underset{NM_{3}}{\mathbb {E}} \ \underset{v_{NM_3}\sim \mathcal {D}_{NM_3}}{\mathbb {E}}Rev(\mathcal{M}_{UD}(\hat{\mathcal{I}}_{NM_{3}})). \end{aligned}$$
(7)

Let \(\hat{\mathcal {I}}^{CP}_{NM_3} = (N_3^{CP}, M^{CP}, \mathcal {D}'^{CP}_{N_3})\) be the COPIES instance corresponding to \(\hat{\mathcal {I}}_{NM_3}\). Following [8], given any set \(NM_3\), the revenue of \(\mathcal {M}_{UD}\) under the unit-demand Bayesian instance \(\hat{\mathcal {I}}_{NM_3}\) is a 6-approximation to the optimal revenue for the COPIES instance. That is, for any \(NM_3\),

$$\begin{aligned} \underset{v_{NM_3} \sim \mathcal {D}_{NM_3}}{\mathbb {E}}Rev(\mathcal{M}_{UD}(\hat{\mathcal{I}}_{NM_{3}}))\ge & \frac{1}{6}\underset{\mathcal {D}'^{CP}_{N_3}}{\mathbb {E}} OPT(\hat{\mathcal{I}}_{NM_{3}}^{CP}). \end{aligned}$$
(8)

Combining Inequalities (7) and (8), we have

$$\begin{aligned} & \underset{v\sim \mathcal {D}}{\mathbb {E}}\ Rev(\mathcal{M}_{IEUD}(\mathcal{I})) \ge \frac{1}{6} \underset{NM_{3}}{\mathbb {E}}\ \underset{\mathcal {D}'^{CP}_{N_3}}{\mathbb {E}}OPT(\hat{\mathcal{I}}_{NM_{3}}^{CP}). \end{aligned}$$
(9)

By Lemma 9,

$$\begin{aligned} \underset{\mathcal {D}'^{CP}_{N_3}}{\mathbb {E}} OPT(\hat{\mathcal {I}}_{NM_{3}}^{CP}) \ge \underset{\mathcal {D}^{CP}}{\mathbb {E}} OPT(\mathcal {I}'^{CP})_{NM_3}, \end{aligned}$$

thus

$$\begin{aligned} \underset{NM_{3}}{\mathbb {E}}\ \underset{\mathcal {D}'^{CP}_{N_3}}{\mathbb {E}} OPT(\hat{\mathcal{I}}_{NM_{3}}^{CP}) \ge \underset{NM_{3}}{\mathbb {E}}\ \underset{\mathcal {D}^{CP}}{\mathbb {E}} OPT(\mathcal {I}'^{CP})_{NM_3}. \end{aligned}$$
(10)

Let \(P_{ij}(OPT(\mathcal {I}'^{CP}))\) be the price paid by player i’s copy j under the optimal mechanism for \(\mathcal {I}'^{CP}\). We can rewrite the right-hand side of (10) as follows.

$$\begin{aligned} \hspace{30pt} & \underset{NM_{3}}{\mathbb {E}}\ \underset{\mathcal {D}^{CP}}{\mathbb {E}} OPT(\mathcal {I}'^{CP})_{NM_3} = \underset{NM_{3}}{\mathbb {E}}\ \underset{\mathcal {D}^{CP}}{\mathbb {E}} \sum \limits _{(i,j)\in NM_{3}} P_{ij}(OPT(\mathcal {I}'^{CP})) \nonumber \\= & \underset{\mathcal {D}^{CP}}{\mathbb {E}}\ \underset{NM_{3}}{\mathbb {E}} \sum \limits _{(i, j)\in NM_{3}} P_{ij}(OPT(\mathcal {I}'^{CP})) \nonumber \\= & \underset{\mathcal {D}^{CP}}{\mathbb {E}} \sum \limits _{(i, j) \in N\times M} \Pr ((i, j) \in NM_{3})\cdot P_{ij}(OPT(\mathcal {I}'^{CP})), \end{aligned}$$
(11)

where the first equality is by the definition of the projection, the second is because sampling from \(\mathcal {D}^{CP}\) is done independently from \(NM_3\), and the third is because \(P_{ij}(OPT(\hat{\mathcal{I}}^{CP}))\) does not depend on \(NM_3\). We can further lower-bound the last term of (11) as follows:

$$\begin{aligned} \hspace{20pt} & \underset{\mathcal {D}^{CP}}{\mathbb {E}} \sum \limits _{(i, j) \in N\times M} \Pr ((i, j) \in NM_{3})\cdot P_{ij}(OPT(\mathcal {I}'^{CP})) \nonumber \\= & \underset{\mathcal {D}^{CP}}{\mathbb {E}} \sum \limits _{(i, j) \in N\times M: \exists i' \text{ s.t. } (i',i)\in G_j} \Pr ((i, j) \in NM_{3})\cdot P_{ij}(OPT(\mathcal {I}'^{CP})) \nonumber \\\ge & \frac{1}{4} \underset{\mathcal {D}^{CP}}{\mathbb {E}} \sum \limits _{(i, j) \in N\times M: \exists i' \text{ s.t. } (i',i)\in G_j} P_{ij}(OPT(\mathcal {I}'^{CP})) \nonumber \\= & \frac{1}{4} OPT(\mathcal {I}'^{CP}) \ge \frac{1}{16} OPT(\mathcal {I}'). \end{aligned}$$
(12)

Here the first equality is because \(P_{ij}(OPT(\mathcal {I}'^{CP})) = 0\) for every (ij) such that \(\mathcal {D}_{ij}\) is unknown, the first inequality is by (6), the second equality is by the definition of revenue, and the second inequality is because \(OPT(\mathcal {I}'^{CP}) \ge \frac{OPT(\mathcal {I}')}{4}\) by [9].

Combining (9), (10), (11) and (12), we have

$$\begin{aligned} \mathbb {E}_{v\sim \mathcal {D}} Rev(\mathcal {M}_{IEUD}(\mathcal {I})) \ge \frac{1}{6}\cdot \frac{1}{16} \cdot OPT(\mathcal {I}') = \frac{OPT(\mathcal {I}')}{96}, \end{aligned}$$

and Theorem 4 holds. \(\square\)

Note that Lemma 9 is only concerned with COPIES instances. Using this lemma and similar to our proof of Theorem 4, any Bayesian mechanism \(\mathcal {M}\) whose revenue can be properly lower-bounded by the COPIES instance can be converted to an information elicitation mechanism in a black-box way. We have the following theorem, with the proof omitted.

Theorem 5

Let \(\mathcal {M}\) be any DST Bayesian mechanism such that \(Rev(\mathcal{M}(\hat{\mathcal{I}})) \ge \alpha OPT(\hat{\mathcal{I}}^{CP})\) for some \(\alpha>0\). There exists a 2-DST information elicitation mechanism that uses \(\mathcal {M}\) as a black-box and is a \(\frac{\alpha }{16}\)-approximation to \(OPT_K\).

By Theorem 5, the mechanisms in [4] and [10] automatically imply information elicitation mechanisms. For single-good auctions, replacing mechanism \(\mathcal {M}_{UD}\) with Myerson’s mechanism, the information elicitation mechanism is a 4-approximation to \(OPT_K\).

5.3 Additive auctions

For additive auctions, recall that we can never throw away any information. If a player i’s value distribution for an item j is reported by others, then j may be sold to i via the \(\beta\)-Bundling mechanism of [11] (see Sections 2 and 3 therein), denoted by Bund. If i’s distribution for j is not reported, then j may still be sold to i via the second-price mechanism. Indeed, our mechanism handles the players neither solely based on the original Bayesian instance \(\hat{\mathcal {I}}\) nor solely based on the projected instance \(\mathcal {I}'\). Rather, it works on a hybrid of the two.

Our mechanism \(\mathcal {M}_{IEA}\) is still simple; see Mechanism Algorithm 7. However, significant effort is needed to analyze its revenue. Indeed, note that in Mechanism Algorithm 7, each \(M_i\)– the winning set of agent i– is defined according to the original Bayesian instance \(\hat{\mathcal {I}}\), while the partition of M is done according to the knowledge graphs in the information elicitation instance \(\mathcal {I}\), where \(M_i^1\) contains the items for which i’s distribution is reported and \(M_i^2\) contains the ones that have not been reported by anyone. The mechanism Bund is run on a hybrid instance, where \(\beta _i\) is based on \(\hat{\mathcal {I}}\) and \(\mathcal {D}_i'\) is based on \(\mathcal {I}\). The mechanism Bund then returns the optimal entry fee \(e_i\) and reserve prices \((p'_{j})_{j\in M^1_{i}}\). If \(e_i>0\), the items in \(M_i^1\cap M_i\) are sold to player i with price \(e_i + \sum _{j\in M^1_{i}\cap M_i} p'_j\) if \(\sum _{j\in M^1_{i}\cap M_i} b_{ij}\ge e_i + \sum _{j\in M^1_{i}\cap M_i} p'_j\); otherwise the items in \(M^1_{i}\cap M_i\) are not sold. If \(e_i=0\), for each item \(j\in M^1_{i}\cap M_i\), player i gets item j with price \(p'_j\) if \(b_{ij}\ge p'_j\); otherwise item j is not sold. Finally, the items in \(M_i^2\cap M_i\) are sold using second price.

Algorithm 7
figure g

\(\mathcal {M}_{IEA}\)

Although running Bund on \(\mathcal {I}'\) achieves a constant approximation to \(OPT(\mathcal {I}')\), some items sold by Bund under \(\mathcal {I}'\) may end up being sold by \(\mathcal {M}_{IEA}\) using second-price, and the revenue of \(\mathcal {M}_{IEA}\) cannot be lower-bounded by that of Bund on \(\mathcal {I}'\). To overcome this difficulty, we develop a novel way to use the adjusted revenue [11] in our analysis; see Lemmas 11 and 13, where we also recall the related definitions. As we show there, the adjusted revenue in a hybrid information setting, combined with the revenue of the second-price sale, eventually provides a desirable lower-bound to the revenue of \(\mathcal {M}_{IEA}\). We have the following theorem.

Theorem 6

Mechanism \(\mathcal {M}_{IEA}\) for additive auctions is 2-DST and, for any instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\), \(Rev(\mathcal {M}_{IEA}(\mathcal {I})) \ge \frac{OPT_K(\mathcal {I})}{70}\).

Lemma 10

Mechanism \(\mathcal {M}_{IEA}\) is 2-DST.

Proof

The structure of a detailed proof for Lemma 10 will be following the two requirements in the solution concept and similar to that of Lemma 2. Thus we only highlight the key points here.

For each player i, given the distributions and values reported by the other players, the subsets \(M^1_i\) and \(M^2_i\) do not depend on player i’s strategy, neither do the entry fee \(e_i\) and reserve prices \((p'_j)_{j\in M^1_i}\). As far as player i is concerned, the other players’ values are always taken to be \(b_{-i}\), even if some of their value distributions are not reported. Computing i’s winning sets \(M_i\cap M^1_i\) and \(M_i\cap M^2_i\) is just part of the two mechanisms, Bund and second-price, again with the other players’ values taken to be \(b_{-i}\). Thus, it is dominant for i to report her true values for \(M^1_i\), following the truthfulness of Bund; and it is dominant for her to report her true values for \(M^2_i\), following the truthfulness of the second-price mechanism.

Moreover, given that all players truthfully report their values, for each player i, reporting her true knowledge never hurts him, no matter what knowledge the other players report. Indeed, the winning set \(M_i\) only depends on the players’ reported values. Player i’s reported knowledge may affect how M is partitioned into \(M^1_{i'}\) and \(M^2_{i'}\) for another player \(i'\), but does not affect the sets \(M^1_i\) and \(M^2_i\), or whether she gets some items or not, or the prices she pays. Thus \(\mathcal {M}_{IEA}\) is 2-DST as desired. \(\square\)

The adjusted revenue

To lower-bound the expected revenue of \(\mathcal {M}_{IEA}\), we first introduce several important concepts following [11].

For any single-player Bayesian instance \(\hat{\mathcal {I}_i}=(\{i\},M,\mathcal {D}_i)\) and any non-negative reserve-price vector \(\beta _i=(\beta _{ij})_{j\in M}\), a single-player DST Bayesian mechanism is \(\beta _i\)-exclusive if it never sells an item j to i whenever her bid for j is no larger than \(\beta _{ij}\). Denote by \(Rev^X(\hat{\mathcal {I}}_i,\beta _i)\) the optimal \(\beta _i\)-exclusive revenue for \(\hat{\mathcal {I}}_i\): that is, the supremum over the revenue of \(\beta _i\)-exclusive mechanisms.

For any single-player DST Bayesian mechanism, its \(\beta _i\)-adjusted revenue on \(\hat{\mathcal {I}}_i\) is its revenue minus its social welfare generated from player-item pairs (ij) such that i’s bid for j is no larger than \(\beta _{ij}\). Denote by \(Rev^A(\hat{\mathcal {I}}_i,\beta _i)\) the optimal \(\beta _i\)-adjusted revenue for \(\hat{\mathcal {I}}_i\): that is, the supremum over the \(\beta _i\)-adjusted revenue of all single-player DST Bayesian mechanisms. Note that if a mechanism is \(\beta _i\)-exclusive, then its \(\beta _i\)-adjusted revenue is exactly its revenue.

Given a Bayesian instance \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\), for each player i and valuation subprofile \(v_{-i}\sim \mathcal {D}_{-i}\), let \(\beta _i(v_{-i}) = (\beta _{ij}(v_{-i}))_{j\in M}\) be such that \(\beta _{ij}(v_{-i})= \max _{i'\ne i}v_{i'j}\) for each item j. Note that we are slightly abusing notations here: each \(\beta _{ij}\) is now a function rather than a value. The optimal \(\beta\)-adjusted revenue for \(\hat{\mathcal {I}}\) is

$$\begin{aligned} \mathbb {E}_{\mathcal {D}} Rev^A(\hat{\mathcal {I}}, \beta )&= \mathbb {E}_{v\sim \mathcal {D}} \sum _i Rev^A(\hat{\mathcal {I}}_i, \beta _i(v_{-i})) \\&= \sum _i \mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} \mathbb {E}_{v_i\sim \mathcal {D}_i} Rev^A(\hat{\mathcal {I}}_i,\beta _i(v_{-i})). \end{aligned}$$

When \(v_{-i}\) is clear from the context, we may simply write \(\beta _i\) and \(\beta _{ij}\).

Because we also consider the projected Bayesian instance \(\mathcal {I}' = (N, M, \mathcal {D}')\), let \(v'\) be v projected on the knowledge graph G: that is, \(v'_{ij} = v_{ij}\) if there exists a player \(i'\) with \((i',i)\in G_j\), and \(v'_{ij} = 0\) otherwise. As v is distributed according to \(\mathcal {D}\), \(v'\) is distributed according to \(\mathcal {D}'\). Thus we sometimes directly sample \(v'\sim \mathcal {D}'\) rather than sample v first and then map it to \(v'\). Let \(\beta'_{ij}(v'_{-i}) = \max _{i'\ne i}v'_{i'j}\) for each player i and item j. The optimal \(\beta'\)-adjusted revenue for \(\mathcal {I}'\) is defined respectively:

$$\begin{aligned} \mathbb {E}_{\mathcal {D}'}Rev^A(\mathcal {I}',\beta')&= \mathbb {E}_{v'\sim \mathcal {D}'} \sum _i Rev^A(\mathcal {I}'_i, \beta'_i(v'_{-i})) \\&= \sum _i \mathbb {E}_{v'_{-i}\sim \mathcal {D}'_{-i}} \mathbb {E}_{v'_i\sim \mathcal {D}'_i} Rev^A(\mathcal {I}'_i,\beta'_i(v'_{-i})). \end{aligned}$$

It is important to emphasize that, given \(v_{-i}\) and \(\beta _i(v_{-i})\), the optimal \(\beta _i\)-exclusive revenue and the optimal \(\beta _i\)-adjusted revenue are well defined for any Bayesian instance for player i, whether it is \(\hat{\mathcal {I}}_i\) or \(\mathcal {I}'_i\). In particular, we will consider \(Rev^X(\mathcal {I}'_i, \beta _i)\) and \(Rev^A(\mathcal {I}'_i,\beta _i)\), which are on the hybrid of \(\hat{\mathcal {I}_i}\) and \(\mathcal {I}'_i\). The optimal \(\beta\)-adjusted revenue on the hybrid of \(\hat{\mathcal {I}}\) and \(\mathcal {I}'\) are similarly defined:

$$\begin{aligned} \mathbb {E}_{\mathcal {D}} Rev^A(\mathcal {I}', \beta ) = \sum _i \mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} \mathbb {E}_{v'_i\sim \mathcal {D}'_i} Rev^A(\mathcal {I}'_i,\beta _i(v_{-i})). \end{aligned}$$

To highlight that in the inner expectation, player i’s value is \(v'_i\) even if it is obtained by first sampling \(v_i\sim \mathcal {D}_i\) and then projecting on G, we may also write it as \(\mathbb {E}_{v_i\sim \mathcal {D}_i} Rev^A(\mathcal {I}'_i, \beta _i; v'_i)\) and

$$\begin{aligned} \mathbb {E}_{\mathcal {D}} Rev^A(\mathcal {I}', \beta )&= \sum _i \mathbb {E}_{v_{-i}\sim \mathcal {D}_{-i}} \mathbb {E}_{v_i\sim \mathcal {D}_i} Rev^A(\mathcal {I}'_i,\beta _i(v_{-i}); v'_i)\\&= \underset{v \sim \mathcal {D}}{\mathbb {E}}\ \sum _i Rev^A(\mathcal {I}'_i, \beta _i(v_{-i}); v'_i). \end{aligned}$$

Finally, denote by IM the individual Myerson mechanism, which sells each item separately using Myerson’s mechanism [1]. The revenue of IM under the projected Bayesian instance \(\mathcal {I}'\), denoted by \(IM(\mathcal {I}')\), is thus the optimal revenue by selling each item separately.

Having defined the notions and notations needed in our proof, we prove Theorem 6 via the following two technical lemmas, whose proofs can be found in Appendix C. For each lemma, note that on the left-hand side the values are drawn from the original Bayesian instance, and on the right-hand side the values are drawn from the projected instance.

Lemma 11

\(\underset{v\sim \mathcal {D}}{\mathbb {E}}Rev(\mathcal {M}_{IEA}(\mathcal {I})) \ge \frac{1}{68} \underset{v'\sim \mathcal {D}'}{\mathbb {E}}Rev^A(\mathcal {I}',\beta').\)

Lemma 12

\(\underset{v\sim \mathcal {D}}{\mathbb {E}}\ Rev(\mathcal {M}_{IEA}(\mathcal {I})) \ge \frac{1}{2} \underset{v'\sim \mathcal {D}'}{\mathbb {E}}\ IM(\mathcal {I}').\)

Proof of Theorem 6

By Theorem 8.1 of [11], \(\mathbb {E}_{v'\sim \mathcal {D}'}Rev^{A}(\mathcal {I}', \beta')+ \underset{v'\sim \mathcal {D}'}{\mathbb {E}}\ IM(\mathcal {I}') \ge OPT(\mathcal {I}')\). Combining this inequality with Lemmas 11 and 12, we have \(\underset{v\sim \mathcal {D}}{\mathbb {E}}\ Rev(\mathcal {M}_{IEA}(\mathcal {I})) \ge \frac{OPT(\mathcal {I}')}{70} = \frac{OPT_K(\mathcal {I})}{70}\), and Theorem 6 holds. \(\square\)

6 Using scoring rules to buy knowledge from players

In this section we use proper scoring rules to reward the players for their knowledge, so that it is strictly better for them to report truthfully. More precisely, a scoring rule is a function f that takes as inputs a distribution \(\mathcal {D}'\) over a state space \(\Omega\) and a random sample \(\omega\) from an underlying true distribution \(\mathcal {D}\) over \(\Omega\), and outputs a real number. Scoring rule f is proper if

$$\begin{aligned} \mathbb {E}_{\omega \sim \mathcal {D}} f(\mathcal {D}, \omega )\ge \mathbb {E}_{\omega \sim \mathcal {D}} f(\mathcal {D}', \omega ) \end{aligned}$$

for any \(\mathcal {D}\) and \(\mathcal {D}'\), and strictly proper if the inequality is strict for any \(\mathcal {D}'\ne \mathcal {D}\). Moreover, f is bounded if there exist constants \(c_1, c_2\) such that \(c_1\le f(\mathcal {D}', \omega ) \le c_2\) for any \(\mathcal {D}'\) and \(\omega\). Our mechanisms can use any strictly proper scoring rules that are bounded. For concreteness, we use Brier’s scoring rule [12]:

$$\begin{aligned} BSR(\mathcal {D}',\omega ) = 2-(\sum _{s\in \Omega }(\delta _{\omega,s}-\mathcal {D}'(s))^2) = 2\mathcal {D}'(\omega ) - \mid \mid \mathcal {D}'\mid \mid _{2}^{2} +1, \end{aligned}$$

where \(\mathcal {D}'(s)\) is the probability of s according to \(\mathcal {D}'\), and \(\delta _{\omega, s}\) is the indicator for \(\omega = s\). Note that \(BSR(\mathcal {D}', \omega ) \in [0,2]\) for any \(\mathcal {D}'\) and \(\omega\).Footnote 7 We remark that Brier’s scoring rule can be replaced by any strictly proper and bounded scoring rules. However, scoring rules like the logarithmic scoring rule [18] are not suitable, as they are unbounded.

In all our mechanisms, when player i reports \(\mathcal {D}^i_{i'j}\ne \bot\) for player \(i'\) and item j, the seller rewards i based on \(BSR(\mathcal {D}^{i}_{i'j}, b_{i'j})\). If there are more than one reporters for the same distribution, the seller can either reward all of them or randomly choose one. We can scale the reward for each distribution so that the total reward given to the players is at most some constant \(\epsilon\), which is an \(\epsilon\) additive loss to the revenue. For example, in Mechanism \(\mathcal {M}_{IEUD}\), the seller could reward each player i with

$$\begin{aligned} r^{i}_{i'j} = \frac{\epsilon }{2mn^{2}}BSR(\mathcal {D}^i_{i'j}, b_{i'j}) \end{aligned}$$

for reporting the value distribution of player \(i'\) on item j.

Although scoring rules help breaking utility-ties, they cause another problem: a player who does not know a distribution may report something she made up, just to receive a reward. Therefore we start by considering our mechanisms under the no-bluff assumption: that is, a player will not report anything about a distribution that she does not know. More precisely, a player i in an information elicitation auction is no-bluff if, for any knowledge graph \(G_j\) and player \(i'\) with \((i, i')\notin G_j\), and for any strategy \((b_i, K_i)\) of i, i reports \(\bot\) for the corresponding distribution of \(i'\). Note that for a player \(i'\) with \((i, i')\in G_j\), i may report any distribution about \(i'\), including \(\bot\). In some sense, the no-bluff assumption is the analogy of the no-overbidding assumption adopted in budget-constrained auctions: a player will not bid higher than her true value or budget, even if doing so may not lead to a price higher than the latter.

For all our mechanisms, it is easy to see that the reward will only affect the players’ incentives for reporting their knowledge, not their incentives for reporting their values. Accordingly, it is still dominant for the players to report their true values, no matter what knowledge they report. Given that the players all report their true values, the reported values are distributed according to the prior. Thus reporting her true knowledge is now strictly better than lying for a player, because it maximizes her reward. Rather than restating all our previous theorems, we summarize them in the theorem below.

Theorem 7

Under the no-bluff assumption, for any information elicitation mechanism in previous sections, the revised mechanism with proper scoring rules is 2-DST, and reporting her true knowledge is strictly better than lying for each player i. Moreover, the mechanism’s revenue is the same as before with an \(\epsilon\) additive loss.

Next, we show how to remove the no-bluff assumption when everything is known. Without this assumption, player i may report a distribution for another player \(i'\)’s value for an item j, even if \((i, i')\notin G_j\). However, if there exists a third player \(\hat{i}\) who knows \(i'\)’s distribution \(\mathcal {D}_{i'j}\), and if player i is also rewarded for player \(\hat{i}\)’s report, then intuitively player i would have no incentive to bluff about \(i'\). That is, not only a player is paid for reporting the distributions she knows, but she is also paid for keeping quiet about the distributions she does not know and letting the experts speak. Surely reporting \(\mathcal {D}_{i'j}\) maximizes player i’s expected reward, but she does not have the information to decide what \(\mathcal {D}_{i'j}\) is.Footnote 8 Therefore, as long as reporting “\(\bot\)” gives player i the same utility as the unknown strategy of reporting \(\mathcal {D}_{i'j}\), and as long as reporting any distribution other than \(\mathcal {D}_{i'j}\) gives her a strictly smaller utility,

player i will report “\(\bot\)” about \(\mathcal {D}_{i'j}\).

Taking mechanism \(\mathcal {M}'_{IEUD}\) in Section 4 as an example, the players are rewarded as follows:

  • For each player \(i'\) and item j, let \(R_{i'j}\) be the set of players who did not report “\(\bot\)” about the value of \(i'\) for j. Randomly select a player \(\hat{i}\) from \(R_{i'j}\) and let \(r_{i'j} = BSR(\mathcal {D}^{\hat{i}}_{i'j}, b_{i'j})\). Reward every player \(i\ne i'\) using \(r_{i'j}\), properly scaled.

Note that the reward \(r_{i'j}\) is given to player i even if she has reported \(\mathcal {D}^i_{i'j} = \bot\).

It is easy to see that, for any k-informed information elicitation instance with \(k\ge 1\), if all players except i report their true values and true knowledge, then player i’s best strategy is to tell the truth about her own. Indeed, reporting her true values is still dominant no matter what knowledge the players’ report. Moreover, for each player \(i'\) and item j, there are two ways for player i to maximize the reward \(r_{i'j}\) she receives: (1) reporting \(\mathcal {D}^i_{i'j} = \bot\), so that \(\hat{i}\) is chosen with probability 1 from the set of players who actually know \(\mathcal {D}_{i'j}\); or (2) successfully guessing \(\mathcal {D}_{i'j}\) and reporting it, so that \(\hat{i}\)’s report is still \(\mathcal {D}_{i'j}\) with probability 1. Indeed, the resulting information elicitation mechanism is Bayesian incentive compatibility (BIC): that is, all players reporting their true values and true knowledge is a Bayesian Nash equilibrium. Note that option (2) is not a well-defined Bayesian strategy, because i does not have enough information to carry it out. In fact, option (1) is the only Bayesian Nash equilibrium in the mechanism, besides the unachieveable ones where each player reports the unknown true distributions. We again summarize our results in the theorem below.

Theorem 8

For any information elicitation mechanism in previous sections where the knowledge graph is at least 1-informed, the revised mechanism does not rely on the no-bluff assumption, and all players reporting their true values and true knowledge is the unique Bayesian Nash equilibrium. Moreover, the mechanism’s revenue is the same as before with an \(\epsilon\) additive loss.

When not everything is known and the players may bluff, a player may not report “\(\bot\)” about a distribution she does not know, because she may still get some reward in case nobody knows that distribution. It is an interesting open problem to design information elicitation mechanisms when not everything is known and without the no-bluff assumption.

7 Information elicitation mechanisms for combinatorial auctions

In our main results, the knowledge graphs for different items can be totally different from each other: player 1’s value distribution for item 2 may be known by player 3, while her value distribution for item 4 may be known by player 5, etc. When all the knowledge graphs are the same, we say that the auction has player-wise information: all value distributions of a player have the same “knower”.

With player-wise information, only one knowledge graph G is needed: an edge \((i, i')\) is in G if and only if player i knows the distribution \(\mathcal {D}_{i'} = \mathcal {D}_{i'1}\times \cdots \times \mathcal {D}_{i'm}\). Such an information setting can also model arbitrary combinatorial auctions, where each player i’s valuation function \(v_i\) maps each subset of items to a non-negative real, with \(v_i(\emptyset ) = 0\). Thus the distribution \(\mathcal {D}_i\) is over such functions, and i’s values for items within a subset can be arbitrarily correlated. Given a combinatorial Bayesian auction instance \(\hat{\mathcal {I}} = (N, M, \mathcal {D}= \times _{i\in N} \mathcal {D}_i)\), a corresponding information elicitation instance is denoted by \(\mathcal {I}= (N, M, \mathcal {D}, G)\), where G is a single knowledge graph rather than a vector of graphs.

Player-wise information is a much stronger assumption than (player, item)-wise information, the information settings considered in Sections 3 to 4 of this paper. We mention this model here mainly for completeness and to facilitate the comparison of our results with the literature. Indeed, with player-wise information and by randomly partitioning the players into reporters and potential buyers as we have done in mechanisms \(\mathcal {M}_{IEUD}\) and \(\mathcal {M}'_{IEUD}\), we have a simple blackbox reduction from Bayesian mechanisms to information elicitation mechanisms. Our mechanism \(\mathcal {M}_{IEB}\) is shown in Mechanism Algorithm 8, where \(\mathcal {M}_B\) can be any Bayesian mechanism. We have the following theorem, whose proof is similar to that of Theorem 2 and thus omitted.

Algorithm 8
figure h

\(\mathcal {M}_{IEB}\)

Theorem 9

For any \(k\in [n-1]\), for any combinatorial auction instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\) where \(\mathcal {I}\) has player-wise information and G is k-informed, if \(\mathcal {M}_B\) is DST then \(\mathcal {M}_{IEB}\) is 2-DST; and if \(\mathcal {M}_B\) is BIC then \(\mathcal {M}_{IEB}\) is BIC. Moreover, if \(\mathcal {M}_B\) is a \(\sigma\)-approximation to OPT, then \(\mathcal {M}_{IEB}\) is a \(\tau _k \sigma\)-approximation to OPT.

With player-wise information, a player i’s valuation distribution \(\mathcal {D}_i\) is either “completely known” or “completely unknown”. Thus for unit-demand auctions there is no need to adopt the COPIES setting to handle the scenario where only part of \(\mathcal {D}_i\) is reported. As before, the approximation ratio of \(\mathcal {M}_{IEB}\) increases as k gets larger and converges to that of the Bayesian mechanism. When \(k=1\), it is a 4-approximation to OPT using the optimal Bayesian mechanism as a black-box. Note that the model studied in [7] is a very special case even compared with the player-wise information setting: that is, G is the complete graph and \(k=n-1\). Since \(\tau _{n-1} = \frac{n-1}{n^{n/(n-1)}} \rightarrow 1-\frac{1}{n}\) when n gets larger, the revenue of our mechanism essentially matches that of [7] for combinatorial auctions. Moreover, if the players can observe private signals and refine their knowledge about the prior, then our mechanism can also aggregate such refinements as in [7]. Finally, we have the following corollary for arbitrary knowledge graphs (i.e., k may be 0).

Corollary 2

For any combinatorial auction instances \(\hat{\mathcal {I}} = (N, M, \mathcal {D})\) and \(\mathcal {I}= (N, M, \mathcal {D}, G)\) with player-wise information, mechanism \(\mathcal {M}_{IEB}\) with \(q=\frac{1}{2}\) is a \(\frac{\sigma }{4}\)-approximation to \(OPT_K(\mathcal {I})\) if \(\mathcal {M}_B\) is a \(\sigma\)-approximation to OPT.

8 Discussion and conclusion

In our main results, the seller asks the players to report the distributions in their entirety, without being concerned with the communication complexity for doing so. This is common in information elicitation and allows us to focus on the main difficulties in aggregating the players’ knowledge, but maybe unrealistic in practice. In Appendix A, we consider the communication complexity from a theoretical point of view and show how to modify our mechanisms so that the players only report a small amount of information about the distributions. A systematic and more practical study of efficient communication is left for further research. Another interesting future direction is to incorporate the cost of reporting into the model and design incentives for the agents to report.

As Bayesian auctions require the seller (and the players under common-prior assumption) has correct knowledge about all distributions, in our main results we do not consider scenarios where players have “insider” knowledge. If the insider knowledge is correct (i.e., is a refinement of the prior), then our mechanisms’ revenue increases; see Appendix B. Still, how to aggregate even the incorrect information that the players may have about each other is a very interesting question for future studies.

Another important direction is to elicit players’ information for BIC mechanisms. For example, the BIC mechanisms in [31, 61] are optimal in their own settings, and it is unclear how to convert them to information elicitation mechanisms.