PaperHub
4.8
/10
Rejected4 位审稿人
最低3最高6标准差1.1
5
3
5
6
3.3
置信度
正确性2.5
贡献度2.5
表达2.5
ICLR 2025

Learnability of Discrete Dynamical Systems under High Classification Noise

OpenReviewPDF
提交: 2024-09-27更新: 2025-02-05
TL;DR

We establish the efficient learnability of discrete dynamical systems under high classification noise

摘要

关键词
Efficient learning under noiseDynamical systemsPAC modelSample complexity

评审与讨论

审稿意见
5

The model being considered is the following: a undirected graph G=(V.E)G=(V.E) is given where VV are vertices (nn in number) and EE decides the neighbors of a vertex vV.v\in V. Each vertex vv can be in a state 00 or 1.1. Each node in vv has a function fvf_v associated with it which is appplied sequentially with t=1,2,.t=1,2,\ldots.

For each iteration tt, fvf_v updates the value at the vertex according to the following rule: if the sum of states of neighbors of vv at time tt is greater than a node dependent threshold τv\tau_v, (not changing with tt) then the node value at t+1t+1 for vv will be set to 11, otherwise, it is set to 0.0.

The problem is to learn the threshold values τv\tau_v correctly from training data for each vv. The training data assumes that initial state is available and a corrupted version of the transitioned data is available. Non-asymptotic sample complexity bounds are obtained for three algorithms.

The authors move the above dynamic setup to stationary setup wherein the problem is converted to the following. The vertices take values in {0,1}\{0,1\} with CC representing the configuration according to a distribution DD (which is unknown), a map hh^* acts on the the initial state CC to provide updated state CC' using threshold τv\tau_v as described erlier. The updated state CC' is measured; the measurement adds uncertainty and flips the state (which is either 00 or 11) with probability ηv.\eta_v. The training set consists of qq tuples (Wi,W^)(W_i,\hat{W}) which denote the initial state and the measured transitioned state (which is corrupted). The problem is then to find the optimal mapping hh that minimizes the error with respect to the hh^* and obtain the sample complexity of training data needed to provide a needed level of accuracy.

优点

The underlying problem being considered is applicable to diverse set of domains. The mathematical approach taken is overall intuitive and seems correct eventhough there are concerns

缺点

The presentation and the main thrust of ideas can be better presented. Some lemmas are not well written or need attention on the mathematics. Please see the questions section of the review to obtain a detailed view of the concerns.

问题

Detailed comments

\bullet The notation for hh^* on line 120 is not in accordance with how its employed in most of the rest of the article.

\bullet PAC system is not described in needed detail.

\bullet Line 170 can be phrased better. Possibly the authors want to empahsize that the learner does not know which transitioned state is wrong. The initial state prior to transitioning is known.

\bullet In Remark 2, authors try and connect the original problem which is dynamic and the "stationary" problem being analyzed. The relationship of the dynamical updates of the original problem and the stationary problem that assumes a distribution of initial states being generated remains unclear and unconvincing; is it possible the dynamic setup cannot be described by a good distribution and thus data collected at diffferent phases of the dynamic evolution will lead to different results.

\bullet. In Remark 3, the authors provide rationale on why the measurment noise corruption leading to flipping the bit with probability less than 0.50.5 leads no loss of generality. It seems like knowledge whether the probability of flipping the bit is greater than 1/2 or less than 1/2 is needed. The other part of the remark on the difficulty of unravelling the structure of the graph is difficult needs to be better qualified. There is a large literature on estimating topology of agents interacting with each other via quite general relationships. This part of the remark overeaches considerably.

\bullet In line 225, the phrase number of labels is used. There is no explanaation of what labels mean. From later part, it seems like the number of labels is two, either zero or one. It is advisable to remove the use of labels and related development (for example in the Proof of Lemma 3.1) and set it to two.

\bullet On element wise ERM, the authors use the emperical loss with respect to node ii of any hypothesis as

minj=1q1(h(Wj)[i]W^j[i])\min\sum_{j=1}^q \mathbb{1} \left(h(W_j)[i]\not = \hat{W}_j[i]\right)

where the tuples (Wj,W^j)(W_j,\hat{W}_j) are provided to the learner. The above optimization will lead to a solution say hopt,ih^{opt,i} for the ithi^{th} node. The authors seem to make an assumption that hopt,i=hopt,j.h^{opt,i}=h^{opt,j}. Though authors state as much, this assumption is very restrictive.

\bullet Lemma 3.1 is quite confusing. The authors introduce partitions with respect to each node ii which are equivalence classes of hypothesis which result in the same empirical loss. Thus, the partition depends on the number qq of the available training data and thus are dictated by the probability distributions DD and ηv.\eta_v. Thus the cardinality of the maximum element of the collection of partitions tmaxt_max should be a random variable (dictated by the distribution on the training size, qq, DD and ηv.\eta_v.) Now the order relation of qq involves tmaxt_max which is itself dependent on qq, DD and ηv.\eta_v. This seems like a circular dependency. Authors needs to indicate clearly what the relationship implies.

\bullet The V-ERM algorithm generates a hypothesis by first setting τiopt\tau^{opt}_i the threshold for node ii that minimizes the empirical loss for node ii and then forming the hypothesis by using τiopt\tau^{opt}_i for node ii for i=1,,n.i=1,\ldots,n. Theorem 3.2 statement does not suffer from the ambiguity of Lemma 3.1 pointed earlier; however, the proof relies on Lemma 3.1. There might be a way to derive Theorem 3.2 directly with the concrete construction of the partition.

\bullet For VisScore and VisRange algorithms, the authors propose to use the training data of configurations and the transitioned measured configurations to evaluate the frequency of a score, of a node, which is number of neighbor states that possess the value 11 over the qq number of training data. For each of these scores the frequency of the transitioned configuration of the node as measured is also tracked. The statitics of the score and the transitioned value is leveraged to obtain estimates on the threshold to be used for setting the state to one. The analysis tracks scores that have high frequency for a node and the associated frequency of transitioned state being 00 or 11. Here standard techniques are employed to arrive at the sample-complexity of etsimates on the probability of scores and the error introduced by the measurement (considered to have a probability less than 1/21/2) to diminish with training size qq to arrive at the result.

\bullet Simulation results are presented that show VERMVERM showing gradual improvement whereas VisScore and VisRange algorithms show phase transition like behavior. More analysis with respect to the nature of the graph would provide more insights. Authors can clarify how the simulations were carried out; whether the dynamic version was simulated or was distribution of configurations DD used in some manner. If the dynamic version was employed then it would be interesting to understand which kinds of distributions arise.

伦理问题详情

None

审稿意见
3

The paper studies the problem of learning discrete dynamical systems under random classification noise. Here, given a graph GG with a collection of boolean functions {hv:vG}\{h_v: v\in G\}, defines the dynamics of the system. Formally, Given C{0,1}nC\in \{0,1\}^{n}, the next state is generated as C^[v]=hv(C)\hat{C}[v]=h_{v}(C) for all vGv\in G. The graph to be known and they get samples of the form C,C^C,\hat{C} where CDC\sim D for some distribution DD. The aim of the task is to predict C^\hat{C} given CC with high probability over DD.

In this paper, they study the problem when the vector C^\hat{C} is corrupted by random classification noise. That is, for each vertex v,C^[v]=hv(C)v,\hat{C}[v]=h_{v}(C) with probability 1η1-\eta and flipped otherwise. They consider hvh_v to be the class of 1 dimensional threshold, that is hv(C)=1h_v(C)=1 iff uN(v)C[u]tv\sum_{u\in N(v)}C[u]\geq t_v where tvt_v is a threshold associated to the node vv and N(v)N(v) is the neighbourhood of vv in GG. They give two algorithms, V-ERM with sample complexity Oη(n2logn/ϵ2)O_\eta(n^2\log n/\epsilon^2) and VisRange with sample complexity Oη(nlogn/ϵ)O_{\eta}(n\log n/\epsilon), the latter matches the sample complexity of the noiseless case.

They also perform experiments using the algorithm

优点

The problem of discrete dynamical systems seems to be an interesting one with practical applications. Robustness to noise in measurements is crucial for safe deployment of such algorithms. The authors take a step towards addressing this problem.

缺点

I might be mistaken but I don't see why the sample complexity of O(nlogn/ϵ(12η)2)O(n\log n/\epsilon(1-2\eta)^2) doesn't immediately follow as a corollary of prior work on learning with bounded noise in the supervised learning setting. I will sketch an argument here. I am willing to change my score if the authors highlight why the following argument doesn't work and why their theorem is not immediate.

For a VC class of dimension VV, [MN07] proved that the sample complexity of PAC learning with bounded noise rate η\eta scales with O(Vϵ(12η)log(1/δ))O(\frac{V}{\epsilon(1-2\eta)}\log(1/\delta)) (see section 1.3.1). Their algorithm is ERM.

Note that since hvh_{v} is a one dimensional threshold (since G is known), the VC dimension is O(1)O(1). Thus, running ERM node-wise should give a sample complexity of O(1ϵ(12η)log(1/δ))O(\frac{1}{\epsilon(1-2\eta)}\log(1/\delta)) and guarantees that with probability 1δ1-\delta, ERM finds a h^\hat{h} that agrees with hh on 1ϵ1-\epsilon fraction of inputs. Setting ϵ=ϵ/n\epsilon=\epsilon/n, δ=δ/n\delta=\delta/n and using a union bound gives the desired result. The algorithm is again node-wise ERM.

[MN07] Pascal Massart and Élodie Nédélec; Risk bounds for statistical learning; The Annals of Statistics, vol. 34, no. 5, 2006, pp. 2326–66.

问题

  1. Refer to the Weaknesses section
  2. The authors claim that the unknown graph case is hard. I don't see why there isn't a direct reduction to halfspace learning with random classification noise. There are many efficient algorithms for this problem. For the state of the art, see [DKT23]

[DKT23] Ilias Diakonikolas, Christos Tzamos, and Daniel Kane; A Strongly Polynomial Algorithm for Approximate Forster Transforms and Its Application to Halfspace Learning; STOC 2023

审稿意见
5

This paper studies the learning of discrete dynamical systems under random classification noise. They consider a dynamical system with a known underlying graph structure with binary values and unknown discrete threshold functions as interaction function. The paper try to answer if the efficient learning is possible under classification noise and establish the required sample complexity for ϵ\epsilon prediction error with probability 1δ1-\delta. They analyze three algorithm: V-ERM, VisScore, and VisRange. V-ERM is element-wise empirical risk minimization algorithm that assigns the threshold value that minimizes the error function over the corrupted training data. The sample complexity scales with O(n2log(n))\mathcal{O}(n^2log(n)) for this algorithm, which is factor nn larger than the theoretical upper bound for the noise free case. In addition, they analyze VisRange algorithm which is the extension of the VisScore algorithm. The authors define the notion of visiting time of a score to run this algorithm. For each range score of the vertex, they assign the output based on the majority voting in the training data set and they choose the threshold as the maximum of the left values of the ranges with the majority vote value 0. This algorithm requires O(nlog(n))\mathcal{O}(nlog(n)) number of samples which matches the theoretical upper bound for the noise-free upper bound. Besides the complexity w.r.t. to the dimension of the system, the sample complexities scales with O((12η)2)\mathcal{O}((1-2\eta)^{-2}) where η\eta is the maximum classification noise among the vertices. The author claims that V-ERM performs better in practice that VisRange despite the worse sample complexities. While the algorithm VisRange exhibits phase transition behavior, the loss function for the algorithm V-ERM decreases gradually. Last but not at least, the authors provided numerical experiments with synthetic and real-life data set to support the theoretical results. In their experiments, they tested the evolution of the loss ll and the number of required samples with dependence on the dimension nn, density of the graph davgd_avg, and the probability of random classification noise η\eta.

优点

  • This paper is the first to study dynamical systems under classification noise according to the literature review.
  • The problem studied is of significant importance. Although the case with a finite hypothesis class and a single interaction function is analyzed, this paper could be a pioneering work in understanding dynamical systems under classification noise.
  • The paper is well-structured and easy to follow in terms of the development of the results. Although there are some minor typos, it is grammatically well-written.
  • Remarks provided throughout the paper address potential questions that may arise while reading. I particularly liked the placement of the remarks, as they do not interrupt the flow of the paper.
  • The implications of the results and their relation to the existing upper and lower bounds are explained well.
  • The proofs are easy to follow and appear mathematically sound in most parts.

缺点

  • Many acronyms are used before written in full format. For example, the algorithm names V-ERM and VisRange should be given in full name and what the names stand for must be explained before using the acronyms. Similarly, on line 097, CNF is used without any explanation.
  • I think V-ERM is not different from empirical risk minimization, where the Hamming distance between the vectors h(Wj)h(\mathcal{W}_j) and Wj^\hat{\mathcal{W}_j} is minimized. V-ERM does not provide any additional insight into the existing literature. Can you clarify how V-ERM differs from the standard empirical risk minimization?
  • For line 316, the use of "visiting time" sounds awkward. It implies those are the times the score ss was visited, not the number of times the score ss was visited. I think it should be renamed as "visiting frequency" rather than "visiting time" because it measures the number of times the score ss was visited.
  • There are issues with the mathematical representation of a General Learning Model on line 219. You claim it is multi-class learning with kk classes for each vertex. However, the vector Wj\mathcal{W}_j takes values in 0,1n{0, 1}^n. It should be 0,1,,k1n{0, 1, \dots, k-1}^n. The same error occurs in Appendix A.3 as well. Nevertheless, I believe this does not affect the proofs or mathematical results presented in the paper.
  • Despite analyzing the general learning model in section 3.1, the result of Lemma 3.1 does not depend on kk. I think you must either point out that Lemma 3.1 is for binary values, or you need to include the bound that depends on k=(k2)/(k1)k' = (k-2)/(k-1), as you provided in the Appendix. Moreover, you can raise ηˉ\bar \eta up to (k1)/(k)(k-1)/(k) in the multi-class case. It might be worth mentioning.
  • On line 370, you say nn is the dominant term, but as η\eta approaches 1/21/2, the term O((12η)2)\mathcal{O}((1-2\eta)^{-2}) becomes dominant. I think it is better to say "dependence on the dimension" rather than "dominant term."
  • Using the letter \ell for the hypothesis class splits and the loss function value in experiments could be confusing.
  • In general, the algorithms seem too simple and appear to have been studied before in various types of classification problems. This problem is a parametric classification problem over a finite number of parameters using a loss function. Therefore, the authors need to explain what kind of new mathematical understanding these algorithms bring. Can you provide a detailed comparison with existing methods in parametric classification?

问题

Please see the weaknesses above in addition to my questions below.

  • Remark 2 is unclear in terms of how this dynamical system could be sampled over a trajectory. I think there could be some issues when the system updates are cyclic and the sampling only captures a single cyclic behavior. Let D\mathcal{D} be the trajectory of the system, and suppose the system is sampled every other time period. We can generate a system such that the score for a vertex ii is 0 during these time periods, but the threshold function could be arbitrarily large, say, the degree of ii. Then, you will not be able to learn the threshold function of vertex ii despite the minimal training error. In other words, any positive threshold value minimizes the training error, but the test error would be high when we randomly sample from trajectory D\mathcal{D}. Therefore, I believe there should be additional assumptions on the sampling behavior over a trajectory to avoid such ill cases and cyclic behavior. How can you address this?
  • If you are sampling data from a trajectory of the dynamic system, the observed data would be correlated over time. How do you address the correlation over time in your proofs and results? Does it pose an additional challenge? How is it different from samples drawn from independent and identically distributed multiple trajectories of the system?
  • If my understanding of Hamming distance ERM is correct, why do you claim ERM cannot be done efficiently unless P = NP on line 064? Do you mean ERM with the error loss function only?
  • In Figure 3 (b), the sample complexity does not increase quadratically for the algorithm V-ERM. The sample complexity is O(n2log(n))\mathcal{O}(n^2 \log(n)) for this algorithm. Rather, it looks like a linear increase. How can you express this behavior?
  • How do you explain the extension from VisScore to VisRange? What was the theoretical and empirical justification for focusing on this problem?
  • How can you generalize these results to hypothesis classes of infinite size, e.g., H=|\mathcal{H}| = \infty? For example, you could have an interaction function with a parameter that can take infinitely many possible values.
  • Similarly, how can these results be applicable to a system where vertex vv receives faulty observations from its neighbors? In other words, if we did not have random classification noise but had random observation noise from the neighbors due to partial information sharing or information asymmetry, can we use these algorithms for learning?
  • How can the result be extended to non-stationary threshold functions (what happens when the threshold changes over time)?
审稿意见
6

This authors study the problem of learning discrete dynamical systems from noisy data. They introduce two algorithms, V-ERM and VisRange, which are noise-tolerant and achieve efficient learning guarantees under PAC-bounds. The paper provides sample complexity bounds and demonstrates that the number of samples needed in the noisy scenario remains in the same order of those in noise-free settings. Experimental results on synthetic and real-world networks further validate the algorithms' effectiveness, revealing performance that favor V-ERM in practical applications and VisRange for theoretical bounds.

优点

Clarity of exposition: The paper is well-written and systematically introduces the problem setting, contributions, and theoretical derivations. Definitions and assumptions are clearly stated.

Intuitive and well-discussed results: The authors provide sound theoretical results with rigorous sample complexity bounds, effectively extending previous works on learning discrete dynamical systems to the noisy setting.

Theoretical contribution: Through empirical evaluations, the authors demonstrate the practical relevance of V-ERM in noisy environments and the sharp theoretical guarantees of VisRange, offering both a practical and a theoretically sound solution.

缺点

  1. In Section 2.2, the authors assume that the underlying graph structure is fully known, which simplifies the learning task. Additional discussion on the practical implications would be necessary, especially regarding the scenarios where graph information may only be partially known.

  2. The assumption on the noise rate may be too restrictive. The authors could expand more on the implications of ηv=1/2\eta_v = 1/2.

问题

How does the guarantees of VisRange vary with increased graph density? It is unclear how the performance scales with denser graphs, as higher connectivity could potentially increase noise propagation. Could the authors give more intuitions on that? Could the authors provide experimental results on how VisRange performs with varying graph densities?

Minor: Repeated ``there" in line 1103.

评论

Dear reviewers and ACs,

We acknowledge that significant revisions are needed at this stage. We are currently preparing a revised version of the paper that incorporates all the feedback provided by the reviewers.

Thank you for your valuable feedback! Once the revised version is completed, we will provide detailed responses to the questions raised by each reviewer.

Best,

Authors

AC 元评审

This work addresses the challenge of learning discrete dynamical systems from noisy data. It introduces efficient noise-tolerant algorithms with provable PAC guarantees and establishes tight sample complexity bounds. The required training samples in the noisy setting match the noise-free upper bound (up to a constant factor) and are only a logarithmic factor higher than the best-known lower bound. Empirical studies on synthetic and real-world networks validate the algorithms' performance.

There reviewers raised several concerns regarding the novelty and technical contribution of the paper. Unfortunately, the authors did not provide any response to these comments.

审稿人讨论附加意见

Four reviews were collected for this paper, with three recommending rejection and one recommending acceptance. The AC agrees with the majority vote, supporting a rejection due to the unaddressed critiques raised by the reviewers.

最终决定

Reject