Shape analysis for time series
This paper introduces an unsupervised representation learning algorithm for time series tailored to biomedical inter-individual studies using tools from shape analysis.
摘要
评审与讨论
This paper proposes a LDDMM method for time series data (TS-LDDMM), by representing time series as deformations of a reference time series. The TS-LDDMM can handle irregularly sampled multivariate time series of variable lengths, and provide shape-based representations of temporal data. They further show the advantages of the model using simulation and real-world examples motivated by biomedical applications, according to robustness to missing samples and classification performance.
优点
- The proposed method can handle multivariate time series irregularly sampled and with variable sizes.
- The way adapting LDDMM to time series may be helpful for applications of other methods for time series.
缺点
- The comparisons of methods mainly focus on robustness of missing and classification. It would be better to compare the “shape” detected for different methods, to show the interpretability advantage of TS-LDDMM.
- typos: line 58 (??) & line 106 (author?)
问题
- The base graph and diffeomorphism learned should be highly dependent on the model parametrization (otherwise they can be anything)? Maybe I miss some parts, but would you mind justifying the choice of your model?
- The results looks similar to naïve factor analysis models for time series (e.g. GP factor analysis model, or https://arxiv.org/abs/2307.02781), which can also handle irregular observations and downstream tasks easily. Would you mind providing some intuitions on benefits on your more complicated model?
局限性
The method may not be feasible for (super) high dimensional time series data. May consider do modeling in latent space for future research.
Thank you for your time and valuable comments. Here are the responses to the weaknesses and questions you raised:
-
Typos. You have pointed out our typos accurately. We acknowledge these errors and will make the necessary corrections.
-
Comparisons to other methods for the interpretability. In the attached PDF of the common author rebuttal, you will find new figures comparing TS-LDDMM with Shape-FPCA [1] on the mouse dataset, as you requested, to assess the interpretability of these methods.
The take-home message is the following: while Shape-FPCA managed to represent the main phenomena in the data, it missed the subtle behavior of the respiratory cycle after drug injection, which is captured by TS-LDDMM.
-
impact of the model. The goal is to encode time series by using the parametrization of the diffeomorphisms. In the LDDMM framework, the parametrization of the diffeomorphisms is highly dependent on the chosen RKHS's kernel. The kernel choice we give with TS-LDDMM offers a better representation compared to the classic LDDMM kernel, as depicted in Figures 2 and 3 in the paper. Remark that in both cases, TS-LDDMM and LDDMM, we can recover any time series graph (as you said) because of the great flexibility of LDDMM for learning diffeomorphisms, but the final representation is not the same depending on the chosen kernel. Although kernel hyperparameters inject prior information into the reference graph, the main source of dependency comes from the dataset.
-
Factor Analysis. Our method distinguishes of Factory analysis on two points. First, TS-LDDMM's goal is to find a vector representation encoding an irregularly sampled time series. The dimension is not dependent on the size of the represented time series but depends on the size of the time series graph of reference . It is particularly convenient to apply statistical methods to vectors of fixed dimension, such as PCA. In your suggested paper [2], the latent representation is related to each observation in such a way that if the dimension of the representation in and the size of the time series at hand is , the dimension of the representation of the time series is . In the suggested paper, their goal is to perform a factor analysis on the gene observations through time, while in our case, we carried out a PCA/factor analysis on the vectors representing the whole time series (not the latent variable related to each observation). Therefore, the two problems are different: We apply a factor analysis (PCA) to the sequence of observations, and they apply a factor analysis to the observations during a sequence.
We greatly appreciate your interest in interpretability and modeling, which will help others be more receptive to our paper. Thank you once again.
[1] Yuexuan Wu, Chao Huang, Anuj Srivastava, Shape-based functional data analysis, 2023
[2] Cai, Jiachen, et al. "Dynamic Factor Analysis with Dependent Gaussian Processes for High-Dimensional Gene Expression Trajectories." arXiv preprint arXiv:2307.02781 (2023).
Thanks for the authors for detailed explanation, and most my questions are resolved. I would keep my initial rating.
The paper presents and extension of the LDDMM shape analysis framework to time series data. The concepts of using deformations (diffeomorphisms) to transform data is extended to the graph of time series data. The authors define a kernel that extends to the augmented space of time+data by treating the time and data components individually, thus decoupling the regularization and ensuring the graph stays a graph. Techniques from the LDDMM world, sparse representations and varifolds, are used in the setup. The method is tested on synthetic and real-world data.
优点
- well-written and clearly presented paper
- interesting idea of deforming time and space together with the LDDMM framework
- the non-linear flows generated by LDDMM may have real benefits compared to the mostly linear time-warping models used in e.g. functional data analysis
- the method is evaluated on different datasets
缺点
- when reading the paper, I found the method to be a somewhat complicated setup for solving the problem, i.e. why apply the full LDDMM setup to graphs when in most applications the time and space warps could be treated separately at each step (each time-point of the flow). I believe the point is that this is handled in section 4 by the kernel. Perhaps the presentation would benefit if the method was presented with the new kernel from the start so that it was clear that the time-space separation happens
- still, the use of LDDMM, sparse representations and varifolds is a quite complicated setup for solving a time-series analysis problem
- there is extensive literature in the functional data analysis community on time-space warping and separation. I am not aware of methods using LDDMM for both time and space as the current paper does, but I think the background section should include relevant FDA papers
问题
- how does the method compare to methods like the metamorphosis framework where there is also to variable (space and image) similarly to the presented case (time and space) With the specific kernel, I could imagine that the underlying structure is very similar
局限性
yes
Thank you for your time and valuable comments. Here are the responses to the weaknesses and questions you raised:
-
Regarding the Functional Data Analysis (FDA) literature. We have compared TS-LDMMM to Shape-FPCA [1] in Figure 1 of the attached PDF to the common rebuttal and on the classification task in Appendix J.1 (Table 4). In both cases, TS-LDDMM compared favorably. Shape-FPCA employs the Square Root Velocity Field (SRVF) representation to separate space and time. However, this method is designed for continuous objects and only applies to time series of the same length. This issue can be addressed through interpolation, but this approach is not always reliable in sparse and irregular sampling scenarios. Most FDA papers [2,3,4] that we are aware of cope with the same issue using interpolation or basis function expansion. In a nutshell, FDA methods have to deal with continuous objects, while LDDMM algorithms can keep a discrete-to-discrete analysis. Thanks to your comments, we plan to extend the discussion on FDA approaches in our background section.
-
Time and space transformation analysis. This topic is indeed addressed in Remark 3 of Section 4.1. Due to the TS-LDDMM's kernel parametrization, the time transformation is independent of the space transformation, but the reverse is not true. Our representation encodes both space and time by design which is a competitive advantage compared to method separating the time and space representation. Indeed, post-hoc analysis of separated time and space representation are not straightforward. The separated space and time representations correlate, and understanding this correlation is crucial to interpreting the data. Consequently, you must concatenate the space and time representations, but there is no single way to do this because the two representations are not commensurable. This fact might explain why TS-LDDMM compared favorably to Shape-FPCA, as depicted in Figure 1 of the PDF in attachment to the common rebuttal.
-
Difference with the metamorphosis framework introduced by 3DMM [5]. The primary difference lies in the pre-processing requirements. The 3DMM framework requires that each mesh be re-parametrized into a consistent form where the number of vertices, triangulation, and the anatomical meaning of each vertex are consistent across all meshes (as stated in the introduction of [6]). In our context, we do not need such pre-processing; the time series graph can have different sizes. However, applying TS-LDDMM to videos to analyze the inter-variability of facial movement is a promising idea.
-
Varifold loss complexity. We acknowledge that the Varifold loss might seem quite complicated for solving a time-series analysis problem. However, in our framework, the loss must take as input two sets of possibly different sizes, which is not a common property for losses. During the design of TS-LDDMM, we also tried using the Maximum Mean Discrepancy (MMD), which is a more straightforward loss, but the performance was lower.
We greatly appreciate your interest in connecting our paper to the broader literature. We believe these connections will help others situate our work within ongoing research. Thank you once again.
[1] Yuexuan Wu, Chao Huang, Anuj Srivastava, Shape-based functional data analysis, 2023
[2] John Warmenhoven , Norma Bargary , Dominik Liebl , Andrew Harrison , Mark A. Robinson , Edward Gunning , Giles Hooker PCA of waveforms and functional PCA: A primer for biomechanics, 2020
[3] Han Lin Shang, A survey of functional principal component analysis, 2013
[4] Yu, Q., Lu, X., Marron, J. S. Principal nested spheres for time-warped functional data analysis, 2017
[5] Volker Blanz and Thomas Vetter, Face Recognition Based on Fitting a 3D Morphable Model, 1999
[6] A 3D Morphable Model learnt from 10,000 face, 2016
Thank you for the careful response. I appreciate the response, though I don't quite agree with the arguments in the first three bullets. Nevertheless, I am still overall positive towards the paper and keep my score.
This paper extends the large deformation diffeomorphic metric mapping (LDDMM) framework to the case of univariate and multivariate functional data (time series). In particular, the focus is on understanding sample variation in the shape of irregularly sampled time series. The proposed framework leverages a graph representation, which is subsequently deformed via a space-time diffeomorphism. The main methodological contribution of the paper appears to be an appropriate parameterization for the tangent space of the diffeomorphism group, which results in time series deformations that preserve the graph structure. The main computational task is to learn a reference graph and a set of shooting vectors (vector fields), which subsequently define the deformation of the reference graph to each time series in the sample via a flow on the diffeomorphism group. For this purpose, the oriented varifold representation is used. The main experiment presented in the manuscript is aimed at showcasing the interpretability of the proposed framework. The dataset under consideration is composed of respiratory cycles for mice (7 controls and 7 deficient in a particular enzyme). The appendices describe proofs a theorem and a lemma, various settings (including hyperparameter and experimental/computational) for the presented method. Additionally, the authors provide two sets of classification experiments where they compare their approach to competitors in the literature.
优点
- The paper appears to be technically sound. The authors encode the shape of a time series as its graph. Given a sample of such graphs, the main task is to learn a reference graph as well as a set of space-time diffeomorphisms that deform the reference graph to each graph in the sample. The entire framework leverages the LDDMM paradigm, which has proven extremely useful in the context of shape analysis of point clouds, landmarks, curves and surfaces.
- The authors carefully consider an appropriate structure on the (tangent space of the) group of diffeomorphisms to ensure that learned deformations preserve the time series graph structure. The framework is supported by a representation theorem that ensures existence of a space-time transformation that is able to warp one time series to another.
- The proposed method is able to handle time series that are irregularly sampled and with different numbers of observations of in time.
- The appendices support the authors' claims that the proposed framework performs better than competitors on two classification tasks, one that involves irregularly sampled time series and one that involves regularly sampled time series.
- The presented application to understanding variation in respiration cycles of mice is interesting and sufficient details to understand the main motivation behind the analysis are presented.
缺点
- In view of existing literature on shape analysis and functional data analysis, the presented framework essentially alters the popular LDDMM framework for the purposes of analyzing variation in univariate and multivariate time series. As such, the novelty of the proposed methods is not very high. The main contribution appears to be an appropriate parameterization of the diffeomorphism group tangent space, which allows the graph structure of time series to be preserved during deformation. This is done through a restriction of general diffeomorphisms of via a kernel that splits the time and space deformations (nonetheless, the space deformation does depend on the time deformation).
- I'm not sure I would consider sensitivity analyses and experiments as contributions. I appreciate the work the authors did to compare their approach to some of the existing approaches in the literature via classification tasks. I additionally like the real data example presented in the main manuscript. However, I'm not quite sure this example shows the benefits of the proposed approach or its interpretability. Figure 3 does perhaps show better principal directions of variation for the proposed TS-LDDMM vs standard LDDMM. I don't quite understand the following sentence though: "Compared to wt mice, the distribution of colq mice feature along the PC1 axis has a heavy left tail, and the associated deformation (-2 σPC) shows an inspiration with two peaks." Why is the left tail in the density for PC1 score significant here? Also, I don't quite see inspiration with two peaks at -2 σPC or why this is related to the two types of mice since I assume PCA was carried out using the pooled data. If I understand correctly, based on the densities of the PC scores, it appears that wt and colq mice differ in terms of the time warping of their respiration cycles since the two PC1 densities are quite different. At the same time, it appears that PC2 scores are fairly similar across the two mouse groups. Figure 4 is also not explained sufficiently. Are the presented respiration cycles examples from the data or reference cycles computed using the proposed method? In the PC1 vs. PC2 plot of the PC scores for the 14 learned reference time series graphs, each corresponding to one mouse, three of the seven colq mice appear more similar to wt mice. Why is this the case? This is not addressed in the description of the figure. Also, it is unclear to me how PCA was carried out in this case, i.e., was PCA carried out with respect to the overall reference graph? In Figure 5, the change in PC1 after exposure to the irritant is interesting and the exposure does appear to affect wt mice more than colq mice.
- Related to the real data analysis, I'm not sure this example fully showcases the usefulness of the presented methodology. Based on my understanding of the dataset, it appears that the respiration cycles are densely (and prehaps regularly) sampled. Each cycle starts at t=0 and the cycles progress for different amounts of time. As such one could alternatively perform the following analysis using existing methods: (1) resample all cycles to the same number of time points (using some form of interpolation), (2) standardize the domain of each cycle to some interval (perhaps [0,1]) and store the time dilation factor for each cycle, (3) apply the SRV framework, for example, to separate phase (x-axis or time) and amplitude (y-axis) variation, (4) apply separate PCA to time variation and amplitude variation to understand overall variation in the data. Could authors elaborate on the benefits of their approach over the one described in points (1)-(4)? I understand that redoing the analysis in the way I described is not realistic during the rebuttal period.
- Overall, the paper is not written well with many typos and grammatical error throughout. Some typos make the presented material difficult to understand. Notation throughout the paper is very dense and I'm not sure that some of the described concepts are necessary. In particular, I fee that the general description of LDDMM could be condensed quite a bit and most concepts related to oriented varifolds could be relegated to one of the appendices that already exists. These concepts are not described very clearly in the main manuscript anyway with the introduced additional notation.
- I appreciate the authors' efforts in assessing sensitivity of their method to different hyperparameter values. How does noise in the observed time series affect the learned reference graph and observation specific deformations? Is the approach fairly robust to noise or does one have to be quite careful about selecting the necessary hyperparameters? The presented sensitivity analysis only consider a sine function without noise.
问题
I've included a number of questions that require clarification in the Weaknesses above. The authors should carefully justify the novel contributions of the presented method. As written, the approach appears to be fairly straightforward extension of the LDDMM framework. Elaboration of various claims made in the real data analysis section is needed. Some of the claims do not appear well-supported by the presented results. Finally, the paper needs substantial editing to correct existing typos and to streamline notation throughout.
局限性
The authors have adequately addressed limitations of their work.
Thank you for your detailed comments and your time, which will help us improve the quality and clarity of this paper. We appreciate your accurate understanding of the paper. We are pleased to say that we have been able to carry out the experiment you suggested during the rebuttal period.
Here are our answers to your concerns:
-
First, we sincerely apologize for the typos and are thankful for your advice. The article has been reviewed and corrected.
-
Regarding the novelty. Representing multivariate and irregularly sampled time series of different lengths for analyzing inter-individual variability is a complex problem.
Our proposed solution might seem simple for someone knowing LDDMM, but the problem we tackled was never appropriately addressed with LDDMM. For instance, in Figure 1, we show that some deformations in the set of deformations considered by a classical LDDMM do not preserve the time series structure. Additionally, Figure 3 shows that the deformations learned on time series with LDDMM do not carry physiological meaning in the Mice experiment.
The given representation of the kernel has been carefully designed to integrate space and time while keeping time independent of space. Initially, we considered separating space and time, as suggested in Weakness comment 3. However, post-hoc analysis of this representation is not straightforward. The separated space and time representations correlate, and understanding this correlation is crucial to interpreting the data. Consequently, you must concatenate the space and time representations, but there is no single way to do this because the two representations are not commensurable. Therefore, we decided to have a representation that includes both space and time by design.
Moreover, we compare favorably to the state-of-the-art in both deep learning and Functional Data Analysis (FDA) literature addressing the same problem. In summary, we have capitalized on a gap in the LDDMM literature and shared our results with other communities working on similar topics.
-
The experiment you suggested. In the author rebuttal, you will find the experiment you suggested in Weakness comment 3. In brief, TS-LDDMM compares favorably to Shape-FPCA in this example, even though Shape-FPCA already provides impressive results. Moreover, Shape-FPCA requires good interpolation of time series, which is not feasible in scenarios with few or missing samples. Additionally, Shape-FPCA has a very low classification score compared to TS-LDDMM (0.38 for Shape-FPCA against 0.83 for TS-LDDMM). This low performance stems not from the package (which we verified with the owner) but from the method itself. This observation opens an avenue for future improvements to the Shape-FPCA method.
-
Mice breathing behavior experiment. To clarify the computation of the PCA, we performed a Kernel-PCA in the RKHS encoding the velocity fields. Each respiratory cycle is represented by its initial velocity field, which is the velocity field that encodes the geodesic from the reference graph to the respiratory cycle at time t=0.
When describing Figure 3, we stated: "Compared to wt mice, the distribution of colq mice feature along the PC1 axis has a heavy left tail, and the associated deformation (-2) shows an inspiration with two peaks". Indeed, inspirations in two peaks are representative of colq mice as they suffer from a motor control impairment [1]. Figure 4.a shows an extreme real example of colq respiratory cycle; it is the furthest on the left of PC1. The principal component preserves physiological meaning, and it can be used to differentiate colq mice from wt mice.
Figure 4.b shows the coordinates of individual learned reference graphs in the PCA coordinates associated with the overall learned reference graph. Figure 4.c shows an example of an individual learned reference graph. As you mentioned, some colq mice are close to wt mice. Indeed, colq mice suffer from a genetic mutation that affects them throughout their growth. The impact on motor control is variable, and some colq mice may appear closer to WT mice.
As you mentioned: "In Figure 5, the change in PC1 after exposure to the irritant is interesting, and the exposure does appear to affect wt mice more than colq mice". This remark is highly relevant. The irritant molecule prohibits the action of a neurotransmitter involved in motor control. Due to their genetic mutation, colq mice have a deficiency in this neurotransmitter, which leads to motor control impairments. When exposed to the irritant molecule, colq mice are already partially accustomed to the neurotransmitter deficiency, whereas WT mice, the control group, highly suffer from the prohibition.
In light of your remarks, we will improve the soundness of this part.
-
Noise sensitivity. TS-LDDMM focuses on fine-shape registration to encompass small sources of variability. It, therefore, has some sensitivity to noise with pros and cons. As depicted in Appendix G, TS-LDDMM performs good registrations for numerous kernel settings, even for noisy data, meaning that we register the exact time-series shape. On the downside, like other shape registration methods, the learning of the reference graph is affected by noise, as illustrated in the second experiment in the attached pdf. However, the regularity of the reference graph can be controlled by penalizing the velocity fields' norm in the loss. Further work on penalization will be conducted to leverage noisy data.
We thank you for your investment in reviewing our paper and the suggested experiments that we conducted, allowing us to demonstrate our method's significance further.
[1] Aurélie Nervo, André-Guilhem Calas, Florian Nachon, and Eric Krejci. Respiratory failure triggered by cholinesterase inhibitors may involve activation of a reflex sensory pathway by acetylcholine spillover. Toxicology, 424:152232, 2019.
I appreciate the authors' efforts in providing comparisons to the square-root velocity framework as part of the rebuttal. I am also satisfied with the authors' clarifications related to (i) novelty of the proposed approach, and (2) the real data example presented in the manuscript. I feel that, if accepted, the authors should improve the presentation significantly so that the methodology and significance of real data results are clearly understandable. I plan to increase my rating from Reject to Borderline Accept.
Thank you for your update and for encouraging us to enhance our presentation. Your insightful comments will help us significantly improve the clarity and comprehensibility of our work. We assure you that the time you've invested will lead to meaningful improvements.
This paper introduces an unsupervised method based on LDDMM to highlight inter-sample shape variability in time series, with the model being able to work on irregular multivariate time series. Extensive studies are conducted to theoretically and experimentally justify the authors' choices. The interpretability and usefulness of the proposed method are demonstrated through experiments on clinical datasets.
优点
- Originality : The proposed model innovates by bringing LDDMM to the time series analysis field, taking advantage of the research done in the representation learning domain (with, for example, their distance between oriented manifolds loss function).
- Quality : The claims made in the paper are all supported with thorough proofs or evidence (with varied datasets). Efforts were also focused on interpretability, which is paramount in clinical settings.
- Clarity : Thanks to the appendix containing the bulk of the technical details, the main body of the paper is easy to follow and well-explained.
- Significance : Inter-individual variability is a very useful tool for clinicians, as referring to standard practice and examples is common in medical settings - hence, this model could greatly help practitioners.
缺点
- It seems that the TS-LDDMM model is computationally expensive, which may be a tall hurdle to clinical usability, but this point is merely mentioned/addressed by the authors.
- For a work intended for adoption by medical professionals, this paper is technical and relies on mathematical concepts and equations that might hinder its reach as is.
- As far as I am aware, code was not made available for this review.
问题
- How long does a full training process take on the mentioned hardware?
局限性
The authors address both the technical limitations (specific libraries, trouble handling high-dimensional time series...) and the societal impacts (positive and negative) : energy consumption and clinical misuse.
Thank you for your time and valuable comments. Here are the responses to the weaknesses and questions you raised:
-
As depicted in Table 1 of the PDF and thanks to its minimal architecture, the training time of TS-LDDMM is below that of large neural networks and beyond that of classic statistical tools. Note that the TS-LDDMM computation time and memory usage can be further optimized with the dedicated package named KeOps.
-
We deeply apologize for the codes; we thought the deadline was one day after the paper. We send the link to an anonymized Github to the Area Chair in the hope that it will be made available to you soon.
-
Though the method is technical, its outcomes are easy for medical professionals to interpret. One possibility consists of visually representing a PCA's principal components as shape deformation (see Figures 3 & 5), enabling the physicians to interpret the principal sources of variability within a population. For example, in the mice experiment and from the deformations shown in Figure 5 we can see that the primary source of variability corresponds to the pause duration between inspiration and expiration. We are also working on creating an online demo with a user-friendly interface to enhance the impact of this paper on medical professionals.
We greatly appreciate your interest in the application and usability of the method. We hope to make this work available soon. Thank you again.
I acknowledge your response and appreciate the additional results provided.
Taking into consideration the new material and other reviews and rebuttals, my review remains positive, but I will follow closely the discussion with reviewer HVwJ.
We appreciate the attention you've given to our paper. We are pleased to note that Reviewer HVwJ has raised their rating from 3 to 5, acknowledging that our rebuttal addressed their concerns regarding novelty and real data experiments. Their remaining concern is the clarity of the presentation. We are committed to making all necessary efforts to improve the clarity of this work. Thank you once again for your time and valuable feedback, which help us remain humble and focused on delivering a high-quality paper.
Thank you for taking the time to review our manuscript. We are grateful for your constructive criticism and the effort you put into evaluating our work. Your careful analysis and suggestions have significantly enhanced the quality of our research.
We carried out all the experiments that you requested. Notably, most of you asked for additional experiments to compare the interpretability of TS-LDDMM with other methods. You will find our new experiments in the attached PDF: We have compared TS-LDDMM to Shape-FPCA [1] on the mouse dataset (Figure 1), we have investigated the sensitivity to noise of TS-LDDMM compared to Shape-FPCA (Figure 2), and we have performed a training time analysis (Table 1).
In what follows, we summarize the protocol and results of the additional experiments.
-
We have compared TS-LDDMM with Shape-FPCA, a state-of-the-art method in Functional Data Analysis, on the mouse dataset.
Protocol. As suggested, we carried out the following experiment: (1) we interpolate breathing cycles to an even number of samples and store the cycles' duration. (2) We aligned all cycles using the SRVF framework [1]. (3) We scale cycles' parametrization to their duration and perform a joint PCA of the parametrization and the aligned signals in their SRVF representation. We perform the analysis on instances after exposure to irritant molecules. We use the package fdasrsf, which only supports univariate signals. The results are presented in Figure 1 of the attached PDF.
Results. The main components look similar. However, a subtle difference, yet important, can be noticed. With Shape-FPCA, the deformation tends to be a uniform time scaling, whereas, with TS-LDDMM, the time dilatation mainly occurs during the pause between inspiration and expiration. Qualitatively, this last deformation fits the physiological phenomenon: Mice's muscles cannot relax after exposure to the irritant molecule, leading to pauses between inspiration and expiration [2].
The take-home message is the following: Qualitatively, contrary to Shape-FPCA, which manages to represent the main phenomena in the data, the deformations of TS-LDDMM capture subtle physiological behaviors essential to the understanding of the phenomenon at hand.
-
As suggested, we have analyzed the sensitivity of TS-LDDMM and Shape-FPCA to noise.
Protocol. Learning the reference graph with TS-LDDMM and SRVF Kacher-mean, a subroutine of Shape-FPCA, with different noise levels. The dataset includes 100 sine waves with randomly generated time parametrization. Four noise scenarios have standard deviations: 0, 0.05, 0.1, and 0.2. The results are presented in Figure 2 of the attached PDF.
Results. Although the overall sine wave shape is preserved, the noise level affects the learned reference graph in both cases. However, the regularity of the reference graph can be controlled by penalizing the norm of the velocity fields in the loss function. Further work on penalization will be conducted to better handle noisy data.
-
We performed a training time comparison.
Protocol. We compare the training time of TS-LDDMM with Shape-FPCA and Neural LNSDE [3] on 4 datasets: 2 univariate and 2 multivariate. We reported the time in Table 1 of the PDF.
Results. The training time of TS-LDDMM is below that of Neural LNSDE (2 to 3 times faster)and beyond that of Shape-FPCA (4 to 5 times slower). Given their respective classification performance, (f1score, Shape-FPCA: 0.38, Neural-LNSDE: 0.70, TS-LDDMM: 0.83), TS-LDDMM seems more relevant for shape analysis.
We hope that our personal responses will meet your expectations, as your feedback has been a great source of improvement for us.
[1] Yuexuan Wu, Chao Huang, Anuj Srivastava, Shape-based functional data analysis, 2023
[2] Aurélie Nervo, André-Guilhem Calas, Florian Nachon, and Eric Krejci. Respiratory failure triggered by cholinesterase inhibitors may involve activation of a reflex sensory pathway by acetylcholine spillover. Toxicology, 424:152232, 2019.
[3] YongKyung Oh, Dongyoung Lim, and Sungil Kim. Stable neural stochastic differential equations in analyzing irregular time series data. In The Twelfth International Conference on Learning Representations, 2024.
The reviewers recognized the interest of this work extending the LDDMM paradigm to account for inter-sample shape variability in time series. The proposed Riemaniann approach to graph representation of time series was found theoretically founded and motivated, and demonstrated through a relevant application. Questions varied across reviewers and concerned several points, in particular related to the need for improving clarity and better acknowledging the state-of-the-art in functional data analysis, and the perceived incremental nature with respect to the LDDMM setting and relationship with respect to the metamorphosis framework. Questions about interpretability were also raised by several reviewers.
The rebuttal focused on testing the proposed method with respect to alternatives from the state-of-the-art (Shape-FPCA), showcasing favorable results in terms of interpretability, sensitivity and computational complexity. These aspects were further discussed with the reviewers, along with the positioning with respect to the state-of-the-art, and the general motivation for the work. The reviewers found the discussion positive, and reached a consensus towards acceptance for this work.
Overall, the representation framework for analyzing time-series variability proposed in this paper is interesting and novel to the field. The results are convincing and potentially relevant to a broad community beyond the clinical/biostatistical domain. For these reasons, the paper is suitable for presentation to NeurIPS 2024.