MAC Advice for facility location mechanism design
We define a notion of Mostly Approximately Correct predictions, and use them to get better strategyproof mechanisms for facility location.
摘要
评审与讨论
This paper studied facility location games with mostly approximately correct (MAC) predictions. In this setting, there are n agents in a metric space, and k facilities to be located in that space. The cost of each agent is defined as the minimum distance from the facilities to their location. Each agent location is private, and the mechanism designer can predict the agent locations. In a MAC prediction, there are two parts of data: some (more than half) are approximately correct up to an additive error, and the others are arbitrary. The goal is to approximately minimize the total cost of all agents while eliciting truthfulness, given the predictions.
-For a 2-facility location on a line space, they designed a randomized mechanism that guarantees an expected approximation ratio of 3.6+, which is better than the result which is derived from the no-prediction setting.
-For single-facility location in a more general metric space, they designed a deterministic mechanism which is better than the result which is derived from the no-prediction setting when the part of arbitrary prediction is small.
-They also studied k-facility location in a general metric space where at least a fraction of the agents must be assigned to each of the k-facilities. A truthful mechanism with a constant approximation ratio is presented.
优点
Overall, this is an interesting and new topic with a few papers having studied facility location with prediction. The leverage of the prediction helps to improve the approximation ratio when we deal with multiple facilities and multi-dimensional space.
缺点
One concern is that I am not sure whether the MAC model is natural enough in this problem. I find the mechanisms designed in this paper are not very interesting in the sense that the predictions are exploited in a very heavy way. For most of the mechanisms, they go like this: if some conditions are satisfied, the mechanism uses the predictions with the agents’ reports completely ignored; otherwise, the predictions are ignored and traditional facility location mechanisms are used. This makes me feel that the mechanisms are artificial, and I also feel like the conceptual contribution of this paper is incremental due to this.
Moreover, the structure of the paper can be improved. The model proposed is quite general while many special cases are studied. For example, there are many settings such as single/2/k facilities, deterministic/randomized mechanisms, and line/general spaces. It would be better to give a table to tell which combinations are addressed and which are not.
问题
No question.
局限性
N.A.
Thank you for your feedback. We respond to your comments and questions below. If our response is satisfactory, we would greatly appreciate it if you would consider raising your score.
-
Concern: “One concern is that I am not sure whether the MAC model is natural enough in this problem.”
Reply: We would like to try and demonstrate that the MAC model is quite natural as a way to capture predictions for facility location mechanism design problems. For example, in such problems, historical data can give predicted locations for each of the agents, which we can then use (along with agent reports) to get better mechanisms. Another “real life” motivational example is the following:
- Example: A large tech company (such as Google or Apple) wants to set up a facility (such as a physical store or data center) as close as possible to its users' home location in some city. They can predict home locations of people via a learning algorithm, trained on features such as GPS data, which WiFi hotspots they are connected to, which stores they often buy at (via wallet application data), and more. When a prediction is correct, it might be close to correct (approximately correct), e.g., due to GPS location not being accurate so the neighboring house location might be predicted rather than the real one. Also, it might be the case that a prediction is far from being correct (such as faulty GPS data, or a place of work being identified as a place of living). For most people, the features contain enough information to approximately predict their home location.
Please also see the discussion in the introduction section (lines 41-80) for why this kind of model is needed to support handling the important issue of outliers, and the discussion in appendix B laying out more potential sources for such MAC predictions.
-
Concern: “I find the mechanisms designed in this paper are not very interesting in the sense that the predictions are exploited in a very heavy way. For most of the mechanisms, they go like this: if some conditions are satisfied, the mechanism uses the predictions with the agents’ reports completely ignored; otherwise, the predictions are ignored and traditional facility location mechanisms are used. This makes me feel that the mechanisms are artificial, and I also feel like the conceptual contribution of this paper is incremental due to this.”
Reply: The mechanisms that only use predictions (and ignore the agent reports) are the ones in the “warm up” Section 6.1: these results are to develop our tools and build towards the main results.
The main technical core of the paper is in Section 6.2, which incorporates both the predictions and the input (agent-reported locations) in quite a non-trivial way. Indeed, we strongly believe it goes beyond the common approach of interpolating between the best “no predictions” algorithm and the approach of completely trusting the predictions:
- The first stage for choosing the first facility via Big-Cluster-Center using the robust statistics results we developed in Appendices F and H, has different guarantees for balanced and unbalanced instances.
- The second stage chooses the second facility using both the predictions (by using the first facility) and the input (reported locations).
- Finally, it is only the characterization we develop for balanced vs unbalanced optimal clustering cases that allows us to show that the two stages together return a good solution, as in the proof of Theorem 9 (Appendix J).
This is why we believe the resulting mechanism is interesting and non-trivial. We hope you will agree with us!
-
Concern: “Moreover, the structure of the paper can be improved. The model proposed is quite general while many special cases are studied. For example, there are many settings such as single/2/k facilities, deterministic/randomized mechanisms, and line/general spaces. It would be better to give a table to tell which combinations are addressed and which are not.”
Reply: One contribution of the paper is to introduce the new model which is general, and to show how it can be used to achieve beyond worst-case analysis in several examples (of facility location mechanism design). We thank the reviewer for suggesting a table for the results and we completely agree it will make things clearer. We intend to add it to the next version.
The authors study variants of the facility location problem from a mechanism design perspective, in which the mechanism receives predictions on the agents' locations. A percentage of the predictions may have an unbounded error, whilst the remainder of the predictions can be incorrect up to a certain bound. This is in contrast to most existing work on facility location with predictions, in which the designer is given predictions for the optimal facility placements. Specifically, the paper looks at the single facility location problem in multiple dimensions, and the 2-facility location problem on a line, designing strategyproof mechanisms with bounded approximation ratios.
优点
The paper is written and structured very well, and I had almost no issues with understanding the paper. I found very few typos or grammatical errors (see Minor comments). The setting is novel, well-motivated, and is applicable to a wide range of researchers (as the paper is relevant to the fields of both machine learning and computational social choice/mechanism design). I did not find any errors in the proofs, which were technically sound. The approximation ratios achieved by the algorithms improve on existing results under reasonable conditions (though this is achieved with the help of predictions). The use of the Big-Cluster-Center mechanism and its analysis is quite interesting.
缺点
My main concern is on the omission of the epsilon term throughout the paper and slightly vague discussion in the appendix, which is a bit suspicious. This could be improved by being more precise with the effect of epsilon on the results. (see the first question) Minor comments: Question 1: “despite large” -> “despite a large” Line 74: add a comma -> “using agent reports, it is…” Line 299: add a comma after the approximation ratio expression Line 358: “sine”-> “since” The references in lines 55-57, 84-86, 194 should be \citep instead of \citet Theorem 9 begins with “For a for a” Definition 13 could be written slightly more formally: instead of “the goal is”, the authors could write something like “the objective is”
问题
Could the authors give some precise examples on how the algorithms’ approximation ratios would change for both small and large epsilon>0? If my understanding is correct, would each approximation ratio have an additive epsilon*n term for small epsilon, and be unbounded for large epsilon? Would “winner-imposing mechanisms” such as from the paper “Winner-Imposing Strategyproof Mechanisms for Multiple Facility Location Games” by Fotakis and Tzamos (2013) be applicable in this setting? If so, could this concept be used to extend the “second-proportional mechanism” to multiple facilities? Regarding the phrase “While we have not optimized the constants,…” on line 129, specifically which constants are being referred to here? And can the authors comment on how much room there is for improvement? Are the existing deterministic mechanisms for the 2-facility location problem with predictions (in the paper by Xu and Lu (2022)) immediately applicable in this setting?
局限性
N.A.
We thank the reviewer for their feedback! We are glad that the reviewer appreciated many aspects of our work. In what follows, we attempt to address the remaining concern and questions.
-
Concern: “My main concern is on the omission of the epsilon term throughout the paper and slightly vague discussion in the appendix, which is a bit suspicious. This could be improved by being more precise with the effect of epsilon on the results. (see the first question)”
Reply: In the reply to the first question later in this rebuttal, we explain the mixed-additive-and-multiplicative nature of the cost that we get by introducing epsilon, showing why it is in fact reasonable to not include it in the calculation. We note that while we considered dropping epsilon from this work (and calling it Mostly Correct Predictions), we ultimately decided to keep it for the following reasons:
-
We wanted to present a general/flexible model that would be useful for problems beyond those considered here. To the best of our knowledge, mechanism design subject to our definition of MAC predictions has not been studied even for the epsilon=0 case, so we could not revert to existing notions from the literature. Moreover, we anticipate that the case of non-zero epsilon may be interesting in other settings, and we wanted to present a comprehensive model to the community.
-
We think that in many instances of facility location it might be unrealistic to expect perfect predictions (𝜖=0), but is realistic to expect a small 𝜖. This notion emphasizes that this is not a problem, and is handled just as well by the model.
-
-
Regarding the minor comments: Thank you, we will fix all these.
-
Question: “Could the authors give some precise examples on how the algorithms’ approximation ratios would change for both small and large epsilon>0? If my understanding is correct, would each approximation ratio have an additive epsilon*n term for small epsilon, and be unbounded for large epsilon?”
Reply: Since the cost increases additively by at most 𝜖*n for all values, we would get a mixed-additive-and-multiplicative result in this setting. Hence, as 𝜖 gets large compared to OPT/n (i.e., large compared to the average optimal cost for each client), our approximation would indeed be unbounded. In this case, one can say that the predictions are not worth using, since the noise (the additive error in the predictions) is more than the signal (the average cost for clients).
- For instance, consider the (degenerate) single facility location problem with n input agents, all located on the real line at x=0, and all MAC predictions are x=𝜖. Our mechanism would return x=𝜖 as the location for the facility. The mechanism’s cost is 𝜖*n while the optimal cost is 0, resulting in an unbounded approximation ratio. While each agent only pays an 𝜖 more, the (purely multiplicative) approximation ratio does not capture the fact that this is still a good solution for a small 𝜖. This kind of example can be extended to k facilities.
We are happy to add this discussion to the next version.
-
Question: “Would “winner-imposing mechanisms” such as from the paper “Winner-Imposing Strategyproof Mechanisms for Multiple Facility Location Games” by Fotakis and Tzamos (2013) be applicable in this setting? If so, could this concept be used to extend the “second-proportional mechanism” to multiple facilities?”
Reply: This is a very interesting question. As Fotakis and Tzamos (2013) point out, giving the mechanism power to force agents to only use a single facility does indeed improve the approximation ratio to a k-dependent constant. Without this additional power, the problem is harder: for they get a linear approximation ratio.
We focused on the problem without this additional power, but we believe it likely that utilizing this additional power can yield better approximation ratios with MAC predictions as well. One promising strategy could be to use the matching between predictions and agent-reported values (as we mention in lines 383-385). This way, it might be possible to force an agent to only use a facility near its predicted value if the reported value is close to the predicted value. This will be a great future direction!
-
Question: “Regarding the phrase “While we have not optimized the constants,…” on line 129, specifically which constants are being referred to here? And can the authors comment on how much room there is for improvement?”
Reply We refer mostly to the constants in Theorems 4 and Theorem 13. It is tight asymptotically in terms of the balancedness constant b: given a constant , the approximation ratio of our mechanism is (we have such an example for which we omitted from the paper). There is room for improvement here for large values of , specifically in extending the theorem to larger values of .
-
Question: “Are the existing deterministic mechanisms for the 2-facility location problem with predictions (in the paper by Xu and Lu (2022)) immediately applicable in this setting?”
Reply: Indeed, we can use the optimal solution for the predicted points as the predicted centers, and then use the mechanism of Xu and Lu (2022). However, the instance from Example 1 shows that even a single outlier can cause the predicted solution’s optimum to be very bad and thus their mechanism will also perform poorly given this prediction.
Thank you very much for the detailed reply. I have no further questions.
This work considers facility location mechanism design in the presence of (pretty good) advice on locations. In this setting, agents report their locations to impact the places facilities are installed or built. In this model, the designer has access to advice that is Mostly and Approximately Correct, a notion introduced by this work whereby at least of the advice is approximately correct with additive error at most .
The authors show that they can use that advice to beat the theoretical results without advice. In particular, for 2 facility location on a line, they can achieve approximation ratio of , a significant improvement over the best known without advice.
The algorithm for merging the advice and reports is a simple combination of the two: first, use the advice to place the first facility, then use the agent reports to place the second facility. The analysis is quite related to some analysis from Lu et al 2010 for facility location in metric spaces without advice, but they are able to improve the analysis since they are on the line.
优点
This work presents a general model, MAC, for advice that would seem to be useful in many other situations. In particular the model includes a small chance of arbitrarily bad advice.
The presented algorithm details how to use that advice in order to improve on the classic 2-facility location on a line problem.
缺点
While the MAC notion is a clean one that would seem to be a good vehicle for exploring advice, the additive error here plays a minimal role - the authors immediately drop the everywhere... "we drop the from the calculations to avoid "dragging" the additive -dependent term along each result". As a result of that... is it really better to include it in the first place if you're going to ignore it? Or better to say that the advice is mostly correct and then describe how behavior scales with the additive error? It feels overall like a simpler notion from literature could have been used for advice if the additive error wasn't going to play any sort of role in the analysis.
The underlying approach - while beneficial that it is simple, is a little dependent on the fact that it is somewhat easy to split this problem into two: the first facility location, and the second facility location. The approach and analysis for the second facility location draws very heavily on prior work from Lu et al (2010), and is able to improve on that primarily because of the restriction to real line rather than metric.
Relative to prior work, this work assumes a prediction for each agent. This helps quite a bit in minimizing error, as the use of medians and clustering can mitigate the risk of some outliers. As a result, I wouldn't say that the impact is an improvement on an existing result, but more of a change in the type of error and prediction.
问题
Relative to the work without predictions, there is the requirement that the number of agents assigned to each facility is at least a minimum value. How sensitive is this? Can this be dropped or is there a tight upper bound?
局限性
N/a
We thank the reviewer for their feedback! We are glad that the reviewer appreciated some aspects of our work like its usefulness in many situations. We respond to your comments and questions below. If our response is satisfactory, we would greatly appreciate it if you would consider raising your score.
-
Concern: “While the MAC notion is a clean one that would seem to be a good vehicle for exploring advice, the additive error here plays a minimal role - the authors immediately drop the 𝜖 everywhere... "we drop the 𝜖 from the calculations to avoid "dragging" the additive 𝜖-dependent term along each result". As a result of that... is it really better to include it in the first place if you're going to ignore it? Or better to say that the advice is mostly correct and then describe how behavior scales with the additive error? It feels overall like a simpler notion from literature could have been used for advice if the additive error wasn't going to play any sort of role in the analysis.”
Reply: We agree it is an option to drop 𝜖 for much of this work (and call it Mostly Correct Predictions), but we see the following advantages in keeping it:
-
We wanted to present a general/flexible model that would be useful for problems beyond those considered here. To the best of our knowledge, mechanism design subject to our definition of MAC predictions has not been studied even for the 𝜖=0 case, so we could not revert to existing notions from the literature. Moreover, we anticipate that the case of non-zero 𝜖 may be interesting in other settings, and we wanted to present a comprehensive model to the community.
-
We think that in many instances of facility location it might be unrealistic to expect perfect predictions (𝜖=0), but is realistic to expect a small 𝜖. This notion emphasizes that this is not a problem, and is handled just as well by the model.
-
-
Concern: “while beneficial that it is simple, is a little dependent on the fact that it is somewhat easy to split this problem into two: the first facility location, and the second facility location”.
Reply: While our Robust Half technique approach does proceed in two stages, the stages delicately complement each other, rather than being independent. The first stage for choosing the first facility via Big-Cluster-Center using the robust statistics results we developed in Appendices F and H, has different guarantees for balanced and unbalanced instances. The second stage chooses the second facility using both the predictions (by using the first facility) and the input (reported locations). Finally, it is only the characterization we develop for balanced vs unbalanced optimal clustering cases that allows us to show that the two stages together return a good solution, as in the proof of Theorem 9 (Appendix J).
-
Concern: “The approach and analysis for the second facility location draws very heavily on prior work from Lu et al (2010), and is able to improve on that primarily because of the restriction to real line rather than metric.”
Reply: The approximation ratio of 4 for the algorithm of Lu et al. is in fact tight for the line metric space (see Section 4.3 of their paper showing a lower bound of 4 for the performance of their algorithm for points on the line). So the access to the MAC advice is crucial for improving the approximation (to ) for this tight case. Thus our result separates between what can currently be done with/without advice!
Our results show that MAC predictions are useful for other problems too (like the single facility, and the balanced k-facilities), and we hope they will be useful for other settings --- including the approximation for two-facility location on more general metrics --- in the near future.
-
Concern: “Relative to prior work, this work assumes a prediction for each agent. This helps quite a bit in minimizing error, as the use of medians and clustering can mitigate the risk of some outliers. As a result, I wouldn't say that the impact is an improvement on an existing result, but more of a change in the type of error and prediction.”
Reply: We completely agree: we get an improvement over the “no predictions” setting, but in relation to the previous models, the MAC model is incomparable. That said, it presents a different and useful (and also important, we believe) perspective: by allowing a delta fraction of the predictions to be arbitrarily bad, it is robust to outliers, and gives a finer-grained control over the kinds of errors that may arise in applications.
-
Question: “Relative to the work without predictions, there is the requirement that the number of agents assigned to each facility is at least a minimum value. How sensitive is this? Can this be dropped or is there a tight upper bound?”
Reply: To avoid any possible misunderstanding, let us emphasize that our main results do not require that the number of agents assigned to each facility is at least a minimum value. The only results with this requirement are Theorem 4/Theorem 13 on the balanced k-facility location (for k>1) with predictions.
About the sensitivity: The bounds we know on the sensitivity are in Theorem 13 (line 898, Appendix G). At a high level, this says that our approximations improve as we consider more balanced solutions. Moreover, we cannot remove the balancedness assumption for this type of mechanism as Example 1 (lines 63-68) shows. We do not have a bound in the general case.
Finally, we note that the balanced setting is closely related to the capacitated variant of the problem — studied without predictions by Aziz et al. (2020) who showed a linear (O(n)) approximation ratio for k=2. For k=2 the minimum cluster size setting implies a maximum size (capacitated) and vice versa; therefore, Theorem 13 can be viewed as an improvement of the capacitated variant via MAC predictions to a constant approximation ratio.
This paper studies a learning-augmented version of the facilitation location mechanism design problem. In particular, the authors consider a model for predictions they call “mostly approximately correct” in which most points have an estimate close to the true value and a small fraction can be arbitrarily wrong. In the paper, they first recap that the standard statistics are indeed robust with respect to distance and approximation. Then, they apply these results to solve 1-facility location in and 2-facility location on a line. The latter they do by breaking the problem down into picking the location of the second facility conditioned on the first one being fixed (which is solved by Lu et al. 2020) and separately estimating a good location for the first one.
优点
- defines MAC predictions in a sensible way
- uses results from robust statistics to show that MAC predictions can be used for the original problem
- provides results for a handful of relevant versions of the problem
- they claim results show that the problem strictly benefits from having this kind of MAC advice, demonstrating a separation between what can be done without the advice and what can be done with it. If this is true, this is a great strength of the paper.
缺点
I am unsure if I quite believe the claim that the problem strictly benefits from having this kind of MAC advice for the following reason. It seems in line 331-2 that the better approximation factor comes from analyzing the algorithm in a more restricted metric space, not from having access to the MAC advice.
问题
- Why have definitions 9 and 10 been defined in generality for a pair of location estimators that might be the same? It seems like they are only used in the case where \hat{f} = f, so why not just define them that way to begin with (which is perhaps more familiar to the audience anyway)?
- In Theorem 2, is the cost for the optimizer of the 1-median cost function?
- Theorem 4, looks like (up to constants) is necessary. Is this a limitation?
- I would have found it clearer to define a third set (which may be dependent on time if the mechanism is iterative) that represents the reported locations of the agents. Then, strategyproofness is just saying that That way, it is clear the algorithm gets the reported locations (which confused me at first), but the strategyproof property of the mechanism means agents have no reason to deviate from the truth.
- It would be helpful to have a short explanation of why the coordinatewise median gets approximation factor.
- In lines 294-295, what is the randomness over? Is it randomness over the state of the world? So with probability, e.g., 1/n^2, all of the points could be bad?
- Generally, it would be helpful as a reader to receive some prose descriptions of the algorithms, why they work, and the results as they are presented in technical detail.
- it would be helpful to have a summary table of the problem with different values of k, d and with/without randomization, with/without MAC advice (your results) for at-a-glance parsing the strengths of your paper.
- Relatedly, are there computational hardness results?
- Can you please briefly comment on other abstractions people have used for learned auxiliary inputs to an algorithm? And how MAC is similar / different to them? The nature of the prediction in [Agrawal 2022] is quite different.
Small things:
- I believe lines 222-224 need to be moved to before Definition 6
- Please add a definition of “strategyproof”
- line 332, I believe the citation year should be 2020
局限性
theorem statements detail the settings in which they apply, so the scope of the theoretical results are well-addressed.
We thank the reviewer for their detailed and helpful feedback! We are glad that the reviewer appreciated many aspects of our work. We respond to your comments and questions below. If our response to the suspected concern is satisfactory, we would greatly appreciate it if you would consider raising your score.
- Concerns:
-
Concern: “It seems in line 331-2 that the better approximation factor comes from analyzing the algorithm in a more restricted metric space, not from having access to the MAC advice.”
Reply: The approximation ratio of for the algorithm of Lu et al. (2010) [1] is in fact tight for the line metric space. (See Section 4.3 of their paper showing a lower bound of 4 for the performance of their algorithm for points on the line.) So the access to the MAC advice is indeed crucial for improving the approximation (to ) for this tight case.
Thus our result does separate between what can currently be done with/without advice, and we thank you for noting this is a strength of our paper.
Our results show that MAC predictions are useful for other problems too (like the single facility case, and the balanced k-facilities), and we hope that these will be useful for other settings --- including the approximation for two-facility location on more general metrics --- in the near future.
- Questions:
-
Question: “Why have definitions 9 and 10 been defined in generality for a pair of location estimators that might be the same? It seems like they are only used in the case where , so why not just define them that way to begin with (which is perhaps more familiar to the audience anyway)?”
Reply: In Theorem 4 we use different , which is why we defined it that way. If the reviewer thinks it is too confusing, we are open to changing that.
-
Question: “In Theorem 2, is 𝑚𝑒𝑑1(𝑋) the cost for the optimizer of the 1-median cost function?”
Reply: Yes, we will clarify this.
-
Question: “Theorem 4, looks like (up to constants) is necessary. Is this a limitation?”
Reply: This is an interesting direction, we don’t think getting delta larger than is a barrier in general, but we don’t have better bounds yet.
-
Question: “I would have found it clearer to define a third set (which may be dependent on time if the mechanism is iterative) that represents the reported locations of the agents. Then, strategyproofness is just saying that . That way, it is clear the algorithm gets the reported locations (which confused me at first), but the strategyproof property of the mechanism means agents have no reason to deviate from the truth.”
Reply: We considered that option, but worried that it would add more notation to the paper. We can definitely add a sentence clarifying this point.
-
Question: “It would be helpful to have a short explanation of why the coordinatewise median gets 𝑑 approximation factor.”
Reply: We will add this. In short it is because the coordinatewise median is the optimal solution w.r.t the L1 norm, which is at most sqrt(d) times L2 norm.
-
Question: “In lines 294-295, what is the randomness over? Is it randomness over the state of the world? So with probability, e.g., , all of the points could be bad?”
Reply: Both your statements are correct. In essence, this captures a PAC-style setting, where there is a small probability (e.g. ) that all points could be bad. We hedge against this by using the -approximate Min-Bounding-Box mechanism in this (low-probability) bad event, as we explain in appendix F.3.
-
Proposition: add descriptions of the algorithms and a summary table of the problem.
Reply: We will add descriptions and a summary as suggested (also see the tables in the pdf file attached in the global response).
-
Question: “Relatedly, are there computational hardness results?”
Reply: We don’t have any yet, but that’s a great question and we will continue to think about this.
-
Question: “Can you please briefly comment on other abstractions people have used for learned auxiliary inputs to an algorithm? And how MAC is similar / different to them? The nature of the prediction in [Agrawal 2022] is quite different.”
Reply: We have a comparison of related abstractions people use for learned auxiliary inputs in the related work section 3 (lines 179-191), the model section (lines 83-103), as well as Appendix A (lines 558-591), but let us comment here on the most related work. The predictions in [Agrawal 2022] (and relatedly [Xu&Lu 2022]) are indeed quite different: they predict only the optimal facility location(s) for facility location, for one and two facilities respectively. Unlike these works (and many others), we don’t predict the optimal solution to the problem, but instead predict the input. Moreover, their measure of the prediction error is the distance between the predicted solution and the real one, while our measure is the fraction of errors, regardless of how bad each error is. This allows us to capture the property that some predictions may be arbitrarily bad (outliers), but most of the predictions are good — this fine-grained notion is difficult to achieve using a single prediction.
-
Small things: Reply: We will address these issues. (One clarification, though: the Lu et paper we refer to is from 2010, not 2020:
[1] Pinyan Lu, Xiaorui Sun, Yajun Wang, and Zeyuan Allen Zhu. 2010. Asymptotically optimal strategy proof mechanisms for two-facility games. In Proceedings of the 11th ACM conference on Electronic commerce. 315–324.
Did you have a different paper in mind?)
-
I follow your point re the approximation factor. Thanks for the clarification. But if I understand correctly, the factor you show is for second-facility location, whereas the factor Lu et al. show is for the full 2-Facility Location problem? In which case the apples-to-apples comparison that should appear in line 331 should be Also, please add theorem and section references for the Lu result into your paper so it is clear to the reader exactly what you are citing + comparing to.
Edited to add: I see that another reviewer also had a similar confusion to me about this issue so please clarify it in the paper.
Your responses for the rest sound good! Couple points:
- I think adding as mentioned above would actually help greatly in clarification despite the fact that it adds notation.
- I don't see a pdf attached in the global response. Am I looking in the wrong place?
- I noticed the discussion in the related work about the abstractions for learned auxiliary inputs! I was particularly curious about the input learning-augmentation -- have there been other attempts for this kind of augmentation for related problems, e.g. clustering? If not, is studying such problems in the MAC setting interesting? If so, would suggest mentioning that in the Future Directions section!
- Re Lu 2010 -- you're right; I was looking at the wrong entry, sorry!
Thank you for your reply, we really appreciate it and are very glad the rest of the responses sound good!
Regarding the main concern:
Indeed, we confirm your understanding: the factor of 3 is for the different problem of the second facility location. For an apples-to-apples comparison, we compare the ratio of 4 without predictions to the one of with MAC predictions - this appears in lines 110-113 in Section 2.2 "Our Results".
The goal of lines 331-332 was to emphasize the biggest technical difference in the analysis (compared to the analysis of Lu et al.). However, it is not the difference in the analysis which impacts the approximation ratio the most. We completely agree that lines 331-332 are confusing in this sense and will change them to avoid this confusion and be clearer. We are also happy to add the theorem and section references of the Lu et al. result as you suggested.
We hope this clarifies the issue of showing that the improvement is indeed due to MAC predictions rather than a different metric, and thank you for your suggestions that will greatly improve the clarity of our exposition.
Regarding the remaining points:
-
Question: “I noticed the discussion in the related work about the abstractions for learned auxiliary inputs! I was particularly curious about the input learning-augmentation -- have there been other attempts for this kind of augmentation for related problems, e.g. clustering? If not, is studying such problems in the MAC setting interesting? If so, would suggest mentioning that in the Future Directions section!”
-
Reply: We are not entirely sure if we are interpreting "input learning-augmentation" correctly (if we’ve misinterpreted, it would be much appreciated if you could refer us to the correct lines in the related work section). If we understood correctly, you are asking about related work where the predictions are for the problem input, rather than for the optimal solution of the problem instance. Our answer is that we don’t know of other studies considering problems like offline clustering from a perspective similar to the MAC setting. We do think this is a potentially exciting direction, because even for classic algorithm design a small part of the input may contain errors. Two such examples are: using historical location data as the input, even though it accumulated some inaccuracies; and companies like Google using user location data as input, even though it can be noisy. Theorem 4 (lines 267-272) can be viewed as such results for the “balanced” k-medians clustering problem. An intriguing future direction would be to consider other clustering problems — we will add this to the Future Directions section as you suggest, thanks! We are also more than happy to continue the discussion if we did not answer your question or if additional questions arise.
-
Regarding the other comments: Our reply should have been “We will add descriptions and a summary table as suggested”, since we agree that this will be a great improvement to clarity! The PDF file reference was from an earlier version of the response, sorry about that. We will incorporate the remaining suggestion ( notation), thank you!
- Thank you for the clarification on the approximation factor -- indeed those lines were a bit confusing, and I appreciate that you will fix them in the next version.
- You interpreted "input learning-augmentation" exactly as I meant it! Apologies for the lack of clarity. Thanks for the discussion on it, and I would be excited to see future work looking at this style of "noisy" data.
- Got it, okay.
I appreciate the clarifications and trust the authors will make updates to the paper based on the discussion here. I am happy to increase my score from 6 to 8.
Edited to add: the particular reason for skipping 7 is that if indeed there is not other work that tries to model "noisy input" to this class of problems in this way, that is a significant contribution to the community. The specific problem studied here and algorithms provided are interesting, to be sure, but the novelty and sensibility of the MAC model are perhaps the more interesting contributions.
Thank you so much for increasing your score and for the engaged discussion.
We thank the reviewers for taking the time to carefully read our work and for their valuable feedback.
We reply to specific points of each reviewer separately.
This paper provides a new prediction model for algorithms with predictions called MAC advice where a fraction of the predictions can be arbitrarily wrong and the other predictions must be approximately correct. The authors study this new model in the context of facility location mechanism design. This new MAC advice model is interesting and well-motivated. It might be of broader interest and lead to follow-up work that uses this model for other problems