Towards Understanding Memory buffer based Continual Learning
摘要
评审与讨论
This paper investigates the effect of memory buffer for linear models in continual learning. It considers two memory buffer settings, and theoretically derives the forgetting and generalization errors under linear models. Based on the theoretical results, the paper argues that a memory buffer can reduces forgetting and improves generalization under certain conditions, while may have the opposite effect under specific model size, task similarity and other factors.
优点
- Analyze the effect of memory buffers in continual learning is an interesting and important problem.
- Theoretically shows how the memory buffer size effects forgetting and generalization is clear. Meanwhile, analyzing simple cases like helps readers easily understand the relations among them.
- The paper is well structured.
缺点
- The paper is not self-contained. The authors should introduce important previous works like [1], for instance, in the appendix or the main paper.
- For full rehearsal memory buffer setting, the optimization problem is same as the joint (multi-task) training. This is because in the last task , the model is trained on all data and the problem is convex. The paper should discuss the relation between the results in Section 4.2 and joint training.
- The paper should define "similarity" between tasks formally, which I believe is related to . However, in Remark 1, it is unclear why the following statement holds true:
... This suggests that with a smaller memory buffer or a larger number of parameters, the model forgets less than without memory when all tasks are highly similar.
- The connection between theories and memory selection is unclear. The paper only discusses one memory selection method: reservoir sampling-based memory buffer, which limits the generalization of findings to other settings.
Reference
[1] Theory on forgetting and generalization of continual learning. ICML 2023
问题
- In Section 3.2, why the memory buffer does not store the labels, and the labels are generated instead. Meanwhile, it is not clear why for reservoir sampling-based memory buffer, the labels are generated by Eq. (1).
- The paper should have a notation table to sum up all the notations. It is hard to follow the paper since many notations are similar.
- It is not clear where does the paper get the following conclusion, as stated in the abstract:
... (1) a larger memory buffer must be paired with a larger model to effectively reduce forgetting;
Q1. The paper is not self-contained. The authors should introduce important previous works like [1], for instance, in the appendix or the main paper.
Response: Thank you for your suggestion. We clearly state the differences between our work and [1]. Additionally, we summarize the key results from [1] in Lemma 1 to ensure the paper is self-contained.
Q2. For full rehearsal memory buffer setting, the optimization problem is same as the joint (multi-task) training. This is because in the last task , the model is trained on all data and the problem is convex. The paper should discuss the relation between the results in Section 4.2 and joint training.
Response: Thank you for your suggestion. We have added the relationship between our work and [1] in Section 4.2 for clarity.
Q3. The paper should define "similarity" between tasks formally, which I believe is related to . However, in Remark 1, it is unclear why the following statement holds true: This suggests that with a smaller memory buffer or a larger number of parameters, the model forgets less than without memory when all tasks are highly similar.
Response: The task similarity between task and task is characterized by . We have added the definition of task similarity as Definition 1 in Section 4.1. In Remark 1, we discuss the impact of on forgetting in reservoir sampling-based memory buffers. Compared with the result in Lemma 1, the inequality consistently holds under the condition . When all tasks are highly similar (i.e., ), depends primarily on . The condition suggests that a smaller or a larger will make this inequality hold. Therefore, we arrive at the conclusion: “This suggests that with a smaller memory buffer or a larger number of parameters, the model forgets less than without memory when all tasks are highly similar.”
Q4. The connection between theories and memory selection is unclear. The paper only discusses one memory selection method: reservoir sampling-based memory buffer, which limits the generalization of findings to other settings.
Response: In this paper, we primarily investigate two types of memory buffers: limited-sized and unlimited-sized. For the limited-sized memory buffer, we adopt the reservoir sampling-based strategy, as described in Section 4.1. For the unlimited-sized memory buffer, we consider the full rehearsal strategy, as outlined in Section 4.2.
In fact, reservoir sampling-based memory buffers can be generalized to other types of memory buffers, where the memory buffer stores samples for each of the previous tasks. This is a special case of our result, which excludes the random term .
Q5. In Section 3.2, why the memory buffer does not store the labels, and the labels are generated instead. Meanwhile, it is not clear why for reservoir sampling-based memory buffer, the labels are generated by Eq. (1).
Response: We let be the feature matrix and the corresponding label vector for the -th task. We define as the ground truth for the -th task, which satisfies the equation . Therefore, label vector can be expressed in terms of . To improve clarity, we have rewritten the problem setup in Section 3.1. Therefore, the memory stores both features and labels, where the labels can be expressed in terms of the feature matrix and the ground truth vector.
Q6. The paper should have a notation table to sum up all the notations. It is hard to follow the paper since many notations are similar.
Response: Thank you for your suggestion. We have provided a notation table in the Appendix for clarity.
Q7. It is not clear where does the paper get the following conclusion, as stated in the abstract:(1) a larger memory buffer must be paired with a larger model to effectively reduce forgetting.
Response: As described in the response to Q3, the inequality consistently holds under the condition . Therefore, should decrease if increases to ensure that holds. This implies that a larger memory buffer must be paired with a larger model to effectively reduce forgetting.
Reference: [1] Theory on forgetting and generalization of continual learning. ICML 2023
I would like to thanks the authors for the response. All of my questions are answered. I remain positive to this paper.
This paper proposes a theoretical analysis of memory-based over-parameterized linear continual learners. More specifically, the authors explicitly express the forgetting and generalization errosr when training a linear model with a MSE and memory mechanism. The authors derive various relation between memory size, dimension of the input and overall forgetting and generalization. Based on this analysis, the authors make various claim regarding the choice of model and memory size to minimize forgetting and maximum generalization.
优点
- The motivation of proposing a theoretical framework to give a deeper understanding of memory-based methods is appreciated
- Some limitations have been rightfully addressed
缺点
Weaknesses
- The introduction is very hard to read and introduces many undefined variables such as . The introduced challenges l.48 are unclear; therefore, the context introduction of this theoretical analysis is extremely confusing. I would advise the authors to give a more intuitive understanding of such challenges (i) and (ii). The bullet point contributions at the end of the introduction are appreciated.
- In 3.1, I do not understand the definition of the ground truth vectors . Ground truth vectors should be but here they are vectors multiplied to the input to give the ground truth. I imagine that what the authors actually mean is that is the linear learned classifier? However, in that case is the prediction, not the ground truth.
- The definition of reservoir sampling is confusing. Traditionally, the probability of selecting a sample to be put in memory is with the memory size and the stream index. Therefore when the authors write l166 "the probability of any example from the previous tasks being in the buffer is equal", it does not correspond to reservoir sampling. The probability of being selected decreases over time.
- In 3.1 the diag operator is not defined
- According to 3.1, is a scalar, however, in equation (1) it is multiplied by and the output is a scalar. How is that? Same remark for eq (2)
- Table 1 is never used in the text
- in 3.3, I believe the assumption corresponds to the features. Therefore, each element of follows an isotropic Gaussian of mean 0 and variance 1. In that sense, each element of future tasks follows the exact same distribution as previous tasks, but the label is different. This seems unrealistic as in Continual Learning the distribution very likely changes over time from one task to the other.
- I believe the authors consider a classification problem, however the training loss in (3) is most certainly more suited for a regression problem. The authors should also discuss the limitation of studying only one specific loss function.
- In Theorem (1) the authors assume that . But the value of can be very large
- is sometimes the dimension of the input (section 3.1), sometimes the number of parameters (Table 1). Even if a larger dimension implies more parameters, it would be more interesting to study the impact of over-parameterization, which does not seem to be the case in this paper.
- How much is the overparameterized assumption important here? Could the same analysis apply otherwise? I believe this analysis makes a lot of sense when fine-tuning linear layers of pre-trained models, however this parallel is lacking in the current writing of the paper.
- An experimental validation of the proposed analysis seems necessary to me given the large amount of assumptions made throughout the entire paper.
Typos
- l.149 lacks the dimension of
- In Assumption 1, it should be
问题
See weaknesses
Q1. The introduction is very hard to read and introduces many undefined variables such as . The introduced challenges l.48 are unclear; therefore, the context introduction of this theoretical analysis is extremely confusing. I would advise the authors to give a more intuitive understanding of such challenges (i) and (ii). The bullet point contributions at the end of the introduction are appreciated.
Response: We define the variables in the caption of table 1 (line 56). represents the number of training tasks, denotes the number of samples per task, represents the number of training model parameters, and is the maximum size of the memory buffer.
Q2. In 3.1, I do not understand the definition of the ground truth vectors . Ground truth vectors should be but here they are vectors multiplied to the input to give the ground truth. I imagine that what the authors actually mean is that is the linear learned classifier? However, in that case is the prediction, not the ground truth.
Response: We let be the feature matrix and the corresponding label vector for the -th task. We define as the ground truth for the -th task, which satisfies the equation . Therefore, label vector can be expressed in terms of , which is why we refer to as the ground truth vector for the -th task. Additionally, in Section 3 of [1], is also referred to as the ground truth.
Q3. The definition of reservoir sampling is confusing. Traditionally, the probability of selecting a sample to be put in memory is wtih the memory size and the stream index. Therefore when the authors write l166 "the probability of any example from the previous tasks being in the buffer is equal", it does not correspond to reservoir sampling. The probability of being selected decreases over time.
Response: "the probability of any example from the previous tasks being in the buffer is equal" means that the probability that any of the examples seen has of being in the buffer is equal to , as described in Section 3.1 of [1]. Although decreases as increase, the probability of selection remains the same for each example. For instance, if we consider five tasks, the probability of any sample from task 1 being in the buffer is , and the probability for a sample from task 4 is also . Therefore, the definition of reservoir sampling used here is consistent.
Q4. In 3.1 the diag operator is not defined.
Response: The diag operator creates a square diagonal matrix where the elements of the vector become the diagonal entries of the matrix, and all off-diagonal elements are zero.
Q5. According to 3.1, is a scalar, however, in equation (1) it is multiplied by and the output is a scalar. How is that? Same remark for eq (2).
Response: We define as a matrix and as a -dimensional vector. Consequently, is an -dimensional vector. Additionally, the output of Equation (1) is a vector.
Q6. Table 1 is never used in the text
Response: Table 1 is included in the introduction to facilitate a convenient comparison between our results and those in [1].
Q7. In 3.3, I believe the assumption corresponds to the features. Therefore, each element of follows an isotropic Gaussian of mean 0 and variance 1. In that sense, each element of future tasks follows the exact same distribution as previous tasks, but the label is different. This seems unrealistic as in Continual Learning the distribution very likely changes over time from one task to the other.
Response: Firstly, we assume follows an isotropic Gaussian to simplify our analysis. This assumption is made for tractability and is a common choice in machine learning, as it provides a general framework. If we were to consider distributions other than the Gaussian for each task, it would introduce significantly more complex coefficients in Eq. (29). Secondly, our results are compared with those in [1], and therefore, we align with Assumption 3.2 from [1]. This alignment allows us to highlight the advantages of training with a memory buffer as compared to no memory in some aspects, as demonstrated in [1]. Finally,although the distribution of each task follows an isotropic Gaussian, the vector captures the variation between tasks. The right-hand-side of Eq (11, 12, 18, 19) also includes , which effectively quantifies the difference between two tasks. We also discuss the impact of memory under different task similarity scenarios in the remark.
Q8. I believe the authors consider a classification problem, however the training loss in (3) is most certainly more suited for a regression problem. The authors should also discuss the limitation of studying only one specific loss function.
Response: In this paper, we consider the linear regression problem. Therefore, we use MSE loss as our training loss.
Q9. In Theorem (1) the authors assume that . But the value of can be very large
Response: In Theorem (1), we consider the reservoir sampling-based memory buffer. We limit the buffer capacity to not exceed , as described in Line 169.
Q10. is sometimes the dimension of the input (section 3.1), sometimes the number of parameters (Table 1). Even if a larger dimension implies more parameters, it would be more interesting to study the impact of over-parameterization, which does not seem to be the case in this paper.
Response: In this paper, we analyze continual learning under an over-parameterized linear model, i.e. . We assume that the feature matrix . Therefore, is not only the diemension of the input, but also the dimension of parameters (i.e. the number of parameters).
Q11. How much is the overparameterized assumption important here? Could the same analysis apply otherwise? I believe this analysis makes a lot of sense when fine-tuning linear layers of pre-trained models, however this parallel is lacking in the current writing of the paper.
Response: Firstly, 'double descent' is a well-known phenomenon in machine learning, where overparameterized models often achieve lower test error than underparameterized models. This makes the overparameterized regime particularly valuable for further study. Secondly, in the underparameterized regime, there is a unique solution that minimizes the MSE loss (i.e. Eq. (3)), given by . This solution can then be substituted into Eq. (7) and Eq. (8) to easily derive the expected forgetting and overall generalization error. However, in the overparameterized regime, minimizing Eq. (3) results in infinitely many solutions that achieve zero error. Among these solutions, we are particularly interested in the one corresponding to the convergent point of SGD. Therefore, the analysis in the overparameterized regime is more complex than in the underparameterized regime, which is why we focus on the overparameterized case.
Reference: [1] Theory on forgetting and generalization of continual learning. ICML 2023
I thank the authors for taking the time to respond to my comments, which have clarified some of my concerns. However, I still believe that this work demonstrates a lot of weaknesses. Notably, the authors ignored the questions regarding the absence of experiments, despite being raised by several reviewers. For these reasons, I will keep my score unchanged.
Q1: So this is normal for you to use notations in the introduction that are defined in the caption of a table that is never called in the text?
Q2: Reviewer 4mXQ had similar comments on this regard and I really think you should reconsider the writing of this; a set of weights is not a ground truth.
Q6: Every introduced table or figure should be called in the text when you write a paper.
Q8: I understand, but this is not stated properly in the paper, and it should be, as it is a clear limitation.
Q10: Same as before, you should clarify your mathematical notations here. Your claim that can interchangeably be a dimension or the number of parameters is not obvious in the current writing.
This paper provides a theoretical investigation of replay based CL where each task is an overparameterized linear regression task, under two different scenarios, i.e., reservior sampling and a full replay where all previous data are stored. By deriving explicit forms of expected forgetting and generalization errors, the authors analyze the impact of memory size and model size on forgetting and generalization error, under the coupling with task similarity.
优点
-
The theory for replayed based CL is very limited, and this paper provides the first explicit forms of forgetting and generalization error for two replay strategies.
-
Analysis are provided to understand the impact of memory.
缺点
-
No experimental studies are provided to justify the theoretical results. In particular, it is not clear if similar phenomenons can be observed in practice with neural networks and real datasets, which challenges the usefulness and importance of the theoretical results in this paper. And it is not clear how the theoretical results can help in practice.
-
It is not convincing that analyzing the full replay case is important here, as this replay strategy is barely used in practice.
-
The presentation needs to be improved. For example, it is not clear what the distinct tasks are in line 172. Figure 1 also needs to be more clear. The last sentence in Remark 6 also seems questionable, as these two figures are talking about generalization error instead of forgetting.
问题
Besides the weakness above, I have some further questions:
-
In the replay buffer, why not put samples for the same old task together? This may be convenient for analysis, but seems not standard.
-
How did you handle the correlation between the replay samples and the model? For example, data stored in the replay buffer for task have already been seen during the training of . When using this data to update the model at task , the model should be corrected with . Address this challenge is important to analyze replay-based CL, but I didn't see how this was particularly addressed in the paper.
-
In section 4.1, the claims about the impact of memory size on forgetting are made only based on , which seems not very rigorous. The reason is that, when changing the memory, from Theorem 1, most time the coefficient of will also change. And unlink the second that depends on task similarity, this first term cannot be very small.
Q1. It is not convincing that analyzing the full replay case is important here, as this replay strategy is barely used in practice.
Response: In our paper, we theoretically analyze the impact of memory on training by comparing two cases: training with a limited-sized memory buffer and with an unlimited-sized memory buffer. Full rehearsal is a typical example of an unlimited-sized memory buffer. Therefore, we provide a theoretical analysis of its effects on both forgetting and generalization.
Q2. The presentation needs to be improved. For example, it is not clear what the distinct tasks are in line 172. Figure 1 also needs to be more clear. The last sentence in Remark 6 also seems questionable, as these two figures are talking about generalization error instead of forgetting.
Response: "the distinct tasks from the previous tasks" means that we randomly select tasks from the set of tasks without replacement. Therefore, any two tasks among the selected tasks will not overlap. Thank you very much for pointing out the error in the last sentence of Remark 6. We have now corrected it accordingly.
Q3. In the replay buffer, why not put samples for the same old task together? This may be convenient for analysis, but seems not standard.
Response: In this paper, storing samples from the same old task is a special case of our result. "The same old task" implies that the ground truths of these samples are identical. Therefore, the term holds.
Q4. How did you handle the correlation between the replay samples and the model? For example, data stored in the replay buffer for task have already been seen during the training of . When using this data to update the model at task , the model should be corrected with . Address this challenge is important to analyze replay-based CL, but I didn't see how this was particularly addressed in the paper.
Response: The convergence point of SGD for minimizing the training loss (MSE) corresponds to the solution of the following optimization problem: . Using the method of Lagrange multipliers, the solution to this problem can be directly obtained as: , where includes the samples from both the -th task and memory buffer, as described in Section 3.3. This formulation accounts for the correlation between the replay samples and the model. Additionally, we provide a more detailed derivation in Eq. (29) in Appendix A.2.
Q5. In section 4.1, the claims about the impact of memory size on forgetting are made only based on , which seems not very rigorous. The reason is that, when changing the memory, from Theorem 1, most time the coefficient of will also change. And unlink the second that depends on task similarity, this first term cannot be very small.
Response: We analyze the impact of memory size on forgetting based on both and . In Remark 1, by comparing and , we conclude that always holds when . Under this condition, in Remarks 2, 3, and 4, we explore the conditions under which . By combining the insights from Remarks 1, 2, 3, and 4, we can determine under which conditions training with a memory buffer results in less forgetting compared to training with no memory buffer. Therefore, when considering the impact of memory size on forgetting, we account for both and .
Thanks the authors for the rebuttal and clarifying some of my questions. But my main concerns regarding the weakness still remain: 1) lack of experimental verifications; 2) I understand that full replay is a special case of unlimited memory, but I am still not convinced that analyzing this case is meaningful for standard CL, because this is far from what has been used in practice. In terms of Q4, I understand how the closed form is obtained, but analyzing the forgetting based on these closed form solutions can be challenging due to the sample-model correlation, i.e., is strongly correlated with , which however has not been clarified. Given these, I decide to maintain my score.
The paper seeks to theoretically analyze the impact of memory on continual learning in overparameterized linear models. To do this, the paper starts from error terms for forgetting and generalization defined in the ICML 2023 paper by Lin et al. for continual learning without a memory buffer. The present paper then extends these error terms for continual learning methods that use one of two memory buffers: 1) a buffer created using reservoir sampling (a partial rehearsal buffer) or 2) a buffer that stores all previous samples (a full rehearsal buffer). The paper claims that these error terms lead to three conclusions:
- “A larger memory buffer must be paired with a larger model to reduce forgetting effectively”;
- “A sufficiently large model can lead to zero forgetting”;
- “A larger memory buffer may improve generalization when tasks are highly similar but can degrade generalization when tasks are highly dissimilar.”
优点
Originality: Theoretically analyzing the error bounds for forgetting and generalization in continual learning (overparameterized linear) models that use a memory buffer is novel. To the best of my knowledge, there do not currently exist any studies that examine this problem. The most related work is the ICML 2023 paper by Lin et al.
Significance: It is useful to study these error bounds for continual learning models that use memory buffers since this has become a predominant strategy for mitigating catastrophic forgetting in experimental works in the field. Theoretical bounds pave the way for a better understanding of experimental findings in the field. Although the error bounds are only derived for overparameterized linear models, future work could explore these bounds for other models like deep neural networks or non-linear models more broadly.
Quality & Clarity: While providing theoretical error bounds for the forgetting and generalization of memory-based continual learning methods is useful, the paper is lacking clarity and is difficult to follow. For example, there are several design choices that do not have clear justifications and the mathematical notation is inconsistent, which is expanded upon in the “Weaknesses” section below.
缺点
My biggest issue with the paper is the lack of clarity, which makes it difficult to evaluate the correctness and impact of the presented results. I will expand on several points related to this next.
-
It is unclear what the meaning of a “task” is in the paper. The data vectors are said to be sampled from Normal(0,1), but then what is changing from one task to another? This is important for understanding the results. Moreover, the notion of “task similarity” is mentioned several times, including in the concluding findings from the theoretical analysis. However, “task similarity” is never mathematically defined. Are we assuming some bound on the difference between two matrices as their “similarity”? If so, this is not mentioned anywhere in the paper and remains unclear.
-
The paper could benefit from more consistent mathematical notation. For example, in Section 3.1, W_T is defined as the set of linear ground truth vectors of all T tasks, but then the vectors of W_T are used to produce the output y=X^T w_t. However, y appears to be a prediction, not a ground truth, so this is unclear. Moreover, in Section 3.2.1, it is stated that “the memory buffer stores m_{t-1} samples” and also that we have “\hat{m}_{t-1}” tasks. The inconsistency of variables makes the math extremely difficult to follow. These are just a few examples.
-
Another benefit could come from plain English explanations of theoretical results. The Remarks in the paper are useful for understanding the theoretical results, however, they are written in terms of math and it is unclear what their implications are for the study. These plain English explanations could help the reader better understand the lemmas, theorems, and remarks in the context of the broader continual learning field.
-
There are several justifications for design choices missing. For example, there is an assumption that the number of samples for each task is equal (Assumption 1). Is this just to simplify notation? How do we know these results will hold for the more general case when each task does not contain the same number of samples? In Section 3.2.1, it is stated that “When t \geq 2, the memory buffer stores m_{t-1} samples for each of the previous t-1 tasks and 1 sample for each of the \hat{m}_{t-1} distinct tasks from the previous t-1 tasks.” Does this mean the buffer only stores 1 sample per task? If so, this is a very strong assumption and it is unclear why this was chosen.
In the study by Lin et al., there were empirical results with deep neural networks that demonstrated similar findings to the theoretical analyses performed for the overparameterized linear model. This was powerful to demonstrate that their theoretical results might hold for more complex models (like non-linear neural networks). It would be helpful to see a similar type of experimental analysis in the present paper to strengthen the theoretical findings.
问题
- How is a “task” being mathematically defined? Why was this specific definition chosen?
- How is “task similarity” being defined? Why was this specific definition chosen?
- It would be helpful to see experimental results similar to those with a deep neural network from Lin et al. to enhance the theoretical findings in the present paper.
- It would be helpful to have plain English explanations of the “Remarks” throughout the paper to better understand the findings.
- It would be helpful to have more context about the study from Lin et al. so that the reader does not have to read excerpts from Lin et al.’s paper to understand the present paper.
- In the Introduction, it is mentioned that the challenges for the present study consist of “the coupling of the data matrix and label vector”. What exactly does this mean?
- What exactly was the setup used to produce the plots in Figure 1? Were Gaussians sampled multiple times to create different “tasks” and then the results averaged for the plots?
Q7. There are several justifications for design choices missing. For example, there is an assumption that the number of samples for each task is equal (Assumption 1). Is this just to simplify notation? How do we know these results will hold for the more general case when each task does not contain the same number of samples? In Section 3.2.1, it is stated that “When , the memory buffer stores samples for each of the previous t-1 tasks and 1 sample for each of the distinct tasks from the previous t-1 tasks.” Does this mean the buffer only stores 1 sample per task? If so, this is a very strong assumption and it is unclear why this was chosen.
Response: As shown in the response to Q7, the buffer stores at least samples per task, not 1. This assumption is consistent with Assumption 3.3 in [1].
Reference: [1] Theory on forgetting and generalization of continual learning. ICML 2023
Thank you to the authors for the responses to my questions and questions from other reviewers. However, several of my concerns are still not fully addressed (see below) and there remain some issues regarding the clarity of the method and its presentation in the paper. Moreover, I still think the paper could benefit from experimental validation (similar to Lin et al.) as suggested by myself, along with Reviewer sdbe and Reviewer ds34. For these reasons, I am maintaining my original score.
Q1. I am still confused about the definition of a task. It seems that every "task" follows a normal distribution, so I am confused about what the non-stationarity is concerning data X. What changes in the data from one task to the next? Moreover, similar to Reviewer ds34, I find defining weights as "ground truth" confusing.
Q2. Similar to Q1, it also seems weird to define task similarity by the similarity of weights. Shouldn't the similarity of tasks also consider the data X?
Q3. While the current paper does state some of the differences with the ICML study by Lin et al., the present paper could still benefit from more details about the setup and assumptions from Lin et al. I found myself having to reference the ICML paper several times to better understand the current paper.
Q4. Thank you for the clarification. The paper could benefit from a more expanded description of this "coupling". However, I am still confused about why the data X needs to be decoupled by task? Also is it assumed that we have task labels to know how to decouple X by task? If so, this assumption is not obvious to the reader. Moreover, this goes back to Q1 about the definition of a task as it relates to the data X.
Q6. My comment was more of a suggestion to improve the notation in the paper for clarity.
Q1. How is a “task” being mathematically defined? Why was this specific definition chosen?
Response: We let be the feature matrix and the corresponding label vector for the -th task. We define as the ground truth for the -th task, which satisfies the equation . Thus, characterizes the differences between tasks. To simplify our analysis, we assume follows for all . Without this assumption, additional parameters would be introduced, leading to significantly more complex results. This assumption is also consistent with Assumption 3.2 of [1].
Q2. How is “task similarity” being defined? Why was this specific definition chosen?
Response: We define the task similarity between any task and as . serves as a representative of task , allowing us to use the norm of their difference to characterize the similarity between two tasks. This approach is a common setting in linear regression and is consistent with Section 4 in [1]. We have added the definition of task similarity as Definition 1 in Section 4.1.
Q3. It would be helpful to have more context about the study from Lin et al. so that the reader does not have to read excerpts from Lin et al.’s paper to understand the present paper.
Response: Thank you for your suggestion. We clearly state the differences between our work and [1]. Additionally, we summarize the key results from [1] in Lemma 1 to ensure the paper is self-contained.
Q4. In the Introduction, it is mentioned that the challenges for the present study consist of “the coupling of the data matrix and label vector”. What exactly does this mean?
Response: As stated in the response to Q1, satisfies the equation . However, our training dataset (Section 3.3) includes both the original dataset and the memory buffer, meaning the expanded feature matrix contains samples from multiple tasks. Consequently, it is not possible to directly define a single for and . To address this issue, we decouple the matrix into a sum of matrices, where each decoupled matrix includes only the samples from a single task, as described in Eq. (1) and Eq. (2).
Q5. What exactly was the setup used to produce the plots in Figure 1? Were Gaussians sampled multiple times to create different “tasks” and then the results averaged for the plots?
Response: In Figure 1, we select specific values for , , , , and to calculate Term and Term . Based on the results, When increases, there may be some extreme situations, including: (1) Enlarging the memory size is not always advantageous for mitigating forgetting (as noted in Remark 4). (2) Enlarging the memory buffer can reduce the generalization error in some specific cases, even for highly dissimilar tasks (as noted in Remark 6).
Q6. The paper could benefit from more consistent mathematical notation. For example, in Section 3.1, is defined as the set of linear ground truth vectors of all T tasks, but then the vectors of are used to produce the output . However, y appears to be a prediction, not a ground truth, so this is unclear. Moreover, in Section 3.2.1, it is stated that “the memory buffer stores samples” and also that we have tasks. The inconsistency of variables makes the math extremely difficult to follow. These are just a few examples.
Response: Firstly, in our setting, we refer to the vector in as the ground truth, not , as described in the response to Q1. Secondly, in Section 3.2.1, the number of samples stored in the memory buffer for each previous task is nearly equal. We define the maximum buffer size as . Therefore, for task , the memory buffer stores approximately samples from each of the previous tasks. However, is not always an integer as changes. To address this, we set and , such that . Thus, the buffer stores samples for each of the previous tasks, with a remainder of samples. We first select tasks and store one additional sample from each in the buffer. This ensures that the difference in the number of samples stored per task does not exceed 1.
I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors.