PaperHub
5.3
/10
Rejected6 位审稿人
最低3最高8标准差1.5
6
5
3
8
5
5
3.5
置信度
正确性2.5
贡献度2.5
表达3.0
ICLR 2025

Node-Level Topological Representation Learning on Point Clouds

OpenReviewPDF
提交: 2024-09-27更新: 2025-02-05
TL;DR

We use Persistent Homology and ideas from Differential Geometry to extract point-level topological features on point clouds.

摘要

关键词
Topological Data AnalysisHodge LaplacianHodge TheoryGeometry ProcessingDifferential GeometryAlgebraic TopologyPoint CloudsRepresentation Learning on Point Clouds

评审与讨论

审稿意见
6

This paper proposes a method to extract node-wise topological information from point clouds. To extract the topological structure of a point cloud and to associate it with each point, they compute persistent homology and utilize information from the generators of these homology classes. Additionally, to make the topological information of each node more meaningful, they connect global and local information using the Hodge Laplacian.

优点

  1. The preliminaries and detailed explanations of the method in Sections 2 and 3 are appropriate and clear.
  2. The authors employ an interesting idea to obtain node-wise features using the information in the generators of persistent homology.
  3. They made an effort to stabilize the features by leveraging insights from algebraic topology.
  4. Theoretical results intuitively explain the advantages of the method.

缺点

  1. Regarding the hyper-parameters, such as γ\gamma on line 310 and δ\delta on line 342, there are some concerns about robustness and the way to determine them. Although the authors test the robustness of the hyper-parameters in Appendix D, these results are limited to the topological clustering dataset. It remains uncertain if they are effective on more practical data.
  2. The validity of the heuristics used in the paper should be further investigated. For example, the way to choose the points from persistence diagram in line 294, the strategy to give the topological information to each node in 342, and the process to aggregate the multiple information in line 350. These approaches should be validated with realistic datasets.
  3. Since various techniques are employed in the paper, it would be desirable to conduct an ablation study. Specifically, it should be tested on simple data how the effectiveness of features changes with or without the Hodge Laplacian.
  4. The intent of the experiments using VAE is unclear. It is difficult to understand what is meant by “topological structure inherent in the sample space,” as described in Figure 11. Additionally, it is unclear there are any implication for any specific applications.

问题

  1. Why did you use VR filtration for n>3? In higher dimension, doesn’t the computation cost of the persistent homology increase since the number of simplices in the VR complex increase?
  2. Is it possible to effectively use TOPF for machine learning tasks other than clustering, such as point cloud classification? Additionally, is quantitative evaluation feasible for such tasks?
  3. How did you choose the dimension of the persistent homology in the experiments?
评论

Thank your very much for your review and your suggestions! Based on the reviews, we have uploaded a revised version of our paper. We will reply to your comments below.

  1. Regarding the hyper-parameters, such as γ\gamma on line 310 and δ\delta on line 342, there are some concerns about robustness and the way to determine them. Although the authors test the robustness of the hyper-parameters in Appendix D, these results are limited to the topological clustering dataset. It remains uncertain if they are effective on more practical data.

The real-world datasets of the paper showed that TOPF with the default parameters performed well. In addition, we have added an experiment on points sampled from the high-dimensional conformation space of cyclooctane, see Figure 11, where TOPF again performed well. Thus we believe that we have reasonable hope for the design and the hyperparameter selection of TOPF to perform well on a wide variety of unseen datasets. (This is partly because hyperparameters like δ\delta and γ\gamma only fine-tune the implemented heuristic/formula for choosing thresholds.) Furthermore, we also give intuition and guidance on how to select the optimal hyperparameters in Appendix D.1. While we acknowledge the concern of the reviewer, we believe we have done all possible reasonable steps to ensure that the hyperparameter selection is sufficiently robust.

  1. The validity of the heuristics used in the paper should be further investigated. For example, the way to choose the points from persistence diagram in line 294, the strategy to give the topological information to each node in 342, and the process to aggregate the multiple information in line 350. These approaches should be validated with realistic datasets.

We thank the reviewer for this comment. In addition to the 11 datasets of the original paper, we have introduced the high-dimensional cyclooctane experiment (Figure 11), analysed the performance of TOPF with decreasing sampling density in Figure 15, and investigated the performance of TOPF on high-dimensional latent spaces of VAEs (Figure 13). Although more datasets are always good and we understand the concern of the reviewer, we have already validated TOPF on a wide range of possible topologically structured datasets.

  1. Since various techniques are employed in the paper, it would be desirable to conduct an ablation study. Specifically, it should be tested on simple data how the effectiveness of features changes with or without the Hodge Laplacian.

Thank you for this suggestion, we have run an ablation study on using the Hodge Laplacian to project the homology generators into the harmonic space in contrast to directly working with the homology generators:

datasetTOPFTOPF without HL
4spheres0.810.08
Ellipses0.950.53
spheresAndGrid0.700.11
HalvedCircle0.710.15
2Spheres2Circles0.940.95
SphereInCircle0.971.00
Spaceship0.920.49
mean0.860.47

This shows that in general, skipping the step with the Hodge Laplacian decreases the performance of TOPF by a wide margin. The exception to this are SphereInCircle or 2Spheres2Circles, where the two methods have similar performance. In these two last examples, points are directly sampled from a manifold, and thus the loops and spheres are very "thin". Thus, the homology generator is already a good approximation of all the simplices responsible for a topological feature. In general however, this is not the case, necessating the use of Differential Geometry and the Hodge Laplacian. We will add a discussion of this to the paper.

  1. The intent of the experiments using VAE is unclear. [..] Additionally, it is unclear there are any implication for any specific applications.

The goal of the VAE experiment was to show that it is possible for TOPF to detect topological structure in the latent space of variational autoencoders. This of course does not imply that the latent space of every VAE has a topological structure. However, we still believe that the experiment serves as a proof-of-concept for an interesting possible application of TOPF. As discussed above, we have now added a discussion of a high-dimensional latent space of a VAE as well. We will try to make this more clear in the paper.

评论

It is difficult to understand what is meant by “topological structure inherent in the sample space,” as described in Figure 11.

Thank you for this remark! The image patches sampled from the image were sampled across the sides of two squares and a line connecting these squares. This structure induces a similar topology in the sample spaces, which then gets embedded into the latent space and detected by TOPF. We have added a more detailed description on this in the caption of Figure 11.

  1. Why did you use VR filtration for n>3? In higher dimension, doesn’t the computation cost of the persistent homology increase since the number of simplices in the VR complex increase?

As is customary in TDA, we compute the persistent homology in dimensions 0, 1, and optionally 2 in our experiments. While computation of higher-dimensional homology is straightforward from a theoretical perspective, there are two separate reasons not to do this in practice:

  1. The manifold hypothesis [1] suggests that even in high-dimensional ambient spaces, the data lies on a low-dimensional submanifold, in practice often of dimension 1 or 2. A well known fact from topology is that k-dimensional manifolds will have homology in degrees only up to their intrinsic dimension independent of the ambient dimension. A wealth of literature in TDA seems to support this assumption.
  2. Computing persistent homology in degrees above 2 is computationally very prohibitive.

Thus, we keep the maximum homology degree fixed even for high-dimensional spaces. The computational complexity of computing the homology of a VR filtration is only depending on the number of points and the homology dimension and is independent of the ambient dimension. (Because the VR filtration operates on the distance matrix.) On the other hand, constructing the alpha-filtration depends on performing a Delaunay triangulation, which is computationally infeasible for high-dimensional spaces. Hence, the VR filtration is computationally optimal for high-dimensional spaces, even though it has more simplices. (But in order to compute k-dimensional homology, it suffices to consider the simplices of degree up to k+1, which can be far smaller than the ambient dimension.)

We will try to explain this reasoning more clear in the paper.

[1] Fefferman, Charles, Sanjoy Mitter, and Hariharan Narayanan. "Testing the manifold hypothesis." Journal of the American Mathematical Society 2016.

  1. Is it possible to effectively use TOPF for machine learning tasks other than clustering, such as point cloud classification? Additionally, is quantitative evaluation feasible for such tasks?

Informed by the effectiveness of topological point cloud level features in Machine Learning applications we believe TOPF has the potential to be useful for downstream machine learning tasks. However, because not all point clouds possess topological structure measurable by homology and topology is often not the only thing we care for, we suggest that TOPF will be most effective when used in conjunction with other more geometric point features or other features form the point level. However, because topological features are different from other existing features and thus integrating TOPF into existing ML pipelines would require a new or strongly modified architecture, we believe that this is an exciting topic for future work.

  1. How did you choose the dimension of the persistent homology in the experiments?

For the experiments in the TCBS, we set the persistent homology dimension to the maximum possible dimension permitted by the ambient dimension, which was 1 and 2 respectively. (It is a fact from algebraic topology that a closed submanifold of an nn-dimensional manifold can have homology only in degree up to n1n-1)

For the experiments with real data, we use prior knowledge on the topological structure we were looking for to set the homology dimensions: For the NALCN channelosome and GroEL proteins, we wanted to detect the loops/1-dimensional holes in the protein structure, and looked at dimension 1. For the Cys123 (with the added convex hull), we looked for protein pockets and thus only looked at dimension 2.

For the VAE latent space embedding experiment, we looked for loops informed by the sampling strategy, and thus set the dimension to 1. For the Lorentz attractor data set, we again looked for loops around the non-attractive fixed points and thus used a homology dimension of 1. For the newly introduced cyclooctane dataset, we did not have prior information and thus set d=2d=2.

We assume that in many use cases, we will already know for what kind of topological features we will be looking for. However, we can always run TOPF with a larger maximum homology dimension and TOPF auto-selects the relevant dimensions by comparing the life times of features across dimensions. (Which is of course more prone for error than using prior knowledge).

评论

Thank you again for your feedback! We hope to have addressed your concerns and will be happy to answer any follow-up questions!

评论

Thank you for your detailed response.

  • I appreciate your answers to the questions regarding hyperparameters and the additional details provided in Appendix D.1. When I first read the paper, I had large concerns about sensitivity to hyperparameters. There are still some minor concerns, but I think that the authors have provided sufficient explanation.
  • Thank you as well for sharing the results of the ablation study. I think it is nice to leverage the Hodge laplacian to refine the simple approach to use PH generators, and I am glad it has been validated experimentally.
  • I also appreciate the detailed explanation regarding aspects such as the dimension d of PH. I no longer have any concerns in this regard.

I have no further questions. Thank you again for your comprehensive responses.

评论

Thank you very much for your reply! We are happy to hear that we have addressed your concerns. We want to thank you once again for the suggestions and questions of your review, which have helped improve the paper!

评论

Your following statement is not correct:

Hence, the VR filtration is computationally optimal for high-dimensional spaces, even though it has more simplices.

VR filtration may contain O(nk+1)O(n^{k+1}) simplices while computing HkH_k, where kk = maximum homology degree. For a large number of points nn (e.g., 100K points), it becomes computationally infeasible to compute VR homology features in commodity machines (with, say < 32 RAM), even for k = 2.

评论

Thank you for your comment! We will reply to your other comments in more detail later, but will first address your concern above. First of all, thank you very much for reading our responses to the other reviewers! We appreciate this!

We wrote:

Hence, the VR filtration is computationally optimal for high-dimensional spaces, even though it has more simplices.

We could have worded this more carefully. In the paragraph we were talking about α\alpha-filtrations vs VR filtrations. What we meant with the above cited sentence is that the VR-filtration is computationally optimal in comparison to the α\alpha-filtration for high-dimensional spaces. The α\alpha-filtration needs to compute a d-dimensional Delaunay triangulation, which has an exponential runtime in the dimension of the ambient space. The VR filtration on the other hand does not depend on the ambient dimension. Thus, in the setting of high-dimensional spaces, the VR filtration is computationally optimal compared to the α\alpha-filtration. As you noted, this does not mean that the VR filtration is feasible for a very high number of points, it is not. In these cases we either have to

  1. Perform farthest-point/random downsampling as employed in TOPF,
  2. Pick a maximal ϵ\epsilon-radius for the VR-filtration. This can greatly decrease the number of simplices in the VR filtration, but it is not immediately clear how to choose ϵmax\epsilon_\text{max}. However, this would work well with the rest of the TOPF pipeline,
  3. Pick a different filtration type which might be more suited for the data (Like sparse VR). The only downside of this is that for other filtration types there are fewer existing implementations, im particular very few libraries compute the homology generators (instead of the cohomology generators), which work better for TOPF. This is why we stick to α\alpha-filtrations and VR filtrations in our proposition of TOPF. We note that the TOPF pipeline is independent of the underlying filtration, and is thus very flexible for generalisation to other types of filtrations.

We hope to have addressed your concern and provided enough context as to how to interpret the above claim. We apologise if we caused confusion. Thank you again for your remark!

审稿意见
5

This paper focuses on constructing node-level topological representation of point clouds. To achieve that, it proposes TOPF to do topological data analysis. By doing that, node-level topological features are extracted from complex point clouds through using discrete variants of concepts from algebraic topology and differential geometry. Experiments are conducted on both synthetic and real data for the verification.

优点

  1. The idea of topological representation on node-level for point clouds is somehow interesting;

  2. The English writing is acceptable

缺点

  1. Fig. 1 doesn't help understand the overall and core design of the TOPF. The authors are highly suggested re-drawing this part. From my personal perspective of view, this figure should consist the motivation and the overview of the proposed methods.

  2. For the experimental part, point cloud matching and registration are typical tasks that require good representation point clouds. I would suggest the authors add comparisons to those related methods, like the unsupervised ones (PPF-FoldNet, ECCV 2018 & WSDesc, TVCG 2022).

问题

See weaknesses

评论

Thank you very much for taking the time to read our paper and for your review! We have uploaded an updated version of the paper based on the feedback we received, and will reply to your points below:

  1. Fig. 1 doesn't help understand the overall and core design of the TOPF. The authors are highly suggested re-drawing this part. From my personal perspective of view, this figure should consist the motivation and the overview of the proposed methods.

Thank you for this suggestion! We have now added a brief explanation for the bars in step 1, added arrows to help readability, and improved the separation between the applications/motivation and the remainder of the method. We want to emphasise that the Figure 1 provides an overview over the method, providing illustration for input and output and the four main steps as described in section 3. Furthermore, we highlight potential applications which serve as the motivation for TOPF. We hope to have addressed your concerns. However, if you still have concrete suggestions for improving the figure, please let us know!

  1. For the experimental part, point cloud matching and registration are typical tasks that require good representation point clouds. I would suggest the authors add comparisons to those related methods, like the unsupervised ones (PPF-FoldNet, ECCV 2018 & WSDesc, TVCG 2022).

Thank you for the excellent suggestion! Benchmarking against unsupervised (or rather weakly supervised methods) seems indeed like a substantial strengthening of our claims. We have added benchmarks agains WSDesc (Pretrained on 3DMatch), as suggested by you!

While WSDesc has a better performance than PointNet on the 3D datasets, TOPF still outperforms it by a wide margin on the more challenging sets of the TCBS. In addition to the 32-dimensional features produced by WSDesc, we have added a feature channel for the normalised 3d coordinates, because this greatly improved the performance of WSDesc on TCBS.

The obtained results are very interesting: On the easiest 3D dataset, 2Spheres2Circles (with a large (>4000) number of points and uniform sampling), WSDesc matches the performance of TOPF with an ARI of 0.95. On Spaceship, WSDesc achieves an ARI 0.76 (TOPF: 0.92). Looking at the results directly, we interpret this as WSDesc struggling with the non-uniform sampling across the different features. Finally, on SphereinCircle, WSDesc achieves an ARI of 0.46. This is better than most other non-toplogical methods, but significantly worse than TOPF with 0.97. In our interpretation, this has to do with the sparseness of just 267 points.

On the 2D datasets (which we embedded into 3D) WSDesc performs very poorly, which might be expected because it was trained for 3D datasets. This shows that on the TCBS, TOPF is competitive even against weakly supervised methods which excel on current benchmarks.

In case you have any further questions, we will be happy to answer them! We are looking forward to the discussion and hearing whether you assessment of our paper has changed based on the rebuttal! Thank you again!

评论

Dear Reviewer, Thank you again for your review and feedback! Based on the feedback received, we have updated and substantially improved the paper. In particular, we have addressed your concerns above.

We would be very happy to hear your updated assessment on the paper and to answer any follow-up questions!

评论

Dear Authors,

I appreciate your efforts and response a lot, thanks again! After throughly reading your reply and discussion with other reviewers, my rating towards this paper is still slightly negative. Just as reviewer hKZz said, as for a journal submission, this paper should get one major revision, but for conference, I cannot recommend accpet for the current version. Mainly due to the clarity in paper writing and figure drawing, and also because of the experimental settings. One major flaw could be the lack of strong baselines, like unsupervised point cloud representation methods (PPFFoldNet or WSDesc) and general feature extractors (DGCNN or KPConv). As a result, I will keep my initial rating.

Best,

YUvn

评论

Thank you very much for your reply and reading our comments and the discussion! We will reply below.

as for a journal submission, this paper should get one major revision

We have uploaded a revised version of the paper, which addresses feedback by the reviewers. In particular, we have included experiments on sampling density and heterogeneity, experiments on high-dimensional data, included WSDesc as a strong benchmark, updated Figure 1, moved Theorem 1 to the appendix to have more time to speak about the experiments, and more. If interpreting this from a journal-submission perspective, this could be seen as the major revision. However, you are of course right to request as many major revisions as you deem fit.

Mainly due to the clarity in paper writing and figure drawing

We have already updated Figure 1 and tried to improve the clarity of writing. Without any detailed feedback, we sincerely do not know what you would want us to do. We believe Figure 1 in its current form is a good illustration of the method, as for example Reviewer 8pLm praised our illustrations. The same reviewer praised the presentation and the step-by-step description of section 3, while even reviewer smnw wrote that our paper is written clearly enough.

Despite this, you write that the clarity in paper writing and figure drawing is the main reason to reject our paper. We would thus ask you to either provide us with concrete feedback that we can act on or to reevaluate your assessment. We do not write this to invalidate your opinion, we would only like to understand whether there is anything concrete that we could change in our paper.

Thank you very much!

One major flaw could be the lack of strong baselines, like unsupervised point cloud representation methods (PPFFoldNet or WSDesc) and general feature extractors (DGCNN or KPConv)

Thanks to your feedback, we have included a benchmark against WSDesc in our revision (see the uploaded pdf), which we outperform. Does this address your concern?

We would furthermore like to add that one of the strengths of TOPF lies in the fact that it is not just a different neural architecture, but rather is founded on insights from algebraic topology, does not require a GPU or pretraining, and has interpretable output. Thank you again for your review!

审稿意见
3

The paper studies point-level features by using persistent homology.

优点

The paper is written clearly enough.

缺点

Lines 177-178: "the filtration value of a k-simplex in the α-filtration is the radius of its circum-k-sphere, which differs from its filtration value in the VR filtration".

The concept of a "filtration value" appears here for the first time and is never defined in the paper, which leads to the major misunderstanding of persistence.

Lines 368-372: "Theorem 4.1 (Topological Point Features of Spheres). Let X consist of at least (n + 2) points (denoted by S) sampled uniformly at random from a unit n-sphere in R^n+1 and an arbitrary number of points with distance of at least 2 to S. When we now consider the α-filtration on this point cloud, with probability 1 we have that (i) there exists an n-th persistent homology class generated by the 2-simplices on the convex hull hull of S".

Even a non-expert can notice that n-th dimensional (?) homology class cannot be generated by 2-dimensional (?) simplices for n>2. Also, Theorem 4.1(i) fails for n=1 and any 3 points on a unit circle that form a non-acute triangle.

There are infinitely many point clouds whose 1-dimensional persistence is empty. Look for a generic family of such clouds in J Applied and Comp Topology 2024.

Even Figure 3 in the paper demonstrates that only 3 points in the 1D persistence represent reasonable cycles, hundreds of others represent only noise. In fact, there is no sense to apply persistence to proteins whose atoms are ordered. In this ordered case, the complete isometry invariant of a point cloud is the classical distance matrix requiring only quadratic time, which was known at least since 1935, see I.Schoenberg, Annals of Mathematics, 724-732, 1935.

The persistence homology is more suitable for clouds of unordered points but is much weaker than the collection of all pairwise distances, which determine any generic point cloud in general position uniquely under isometry, which has been known in computational geometry for more than 20 years, see Boutin and Kemper, 2004.

The stronger invariant (a local distribution of distances) has been widely studied in shape analysis, e.g. see Memoli, "Gromov–Wasserstein distances and the metric approach to object matching" (FocM 2011), extended to a complete and polynomial-time invariant under rigid motion in any fixed Euclidean R^n (CVPR 2023).

The initial obstacle for the proposed research is the lack of a rigorous problem statement. The attempts to state problems in the final section seem a bit late:

Lines 514-520: "selection of the relevant features is a very hard problem" and "efficient computation of simplicial weights leading to the provably most faithful topological point features is an exciting open problem"

The words "relevant" and "most faithful" should be rigorously defined, else anyone can claim that their features are "relevant" and "most faithful".

问题

Can the authors consider 3 points (1,0), (0,1), (-1,0), draw three disks centered at the points with a growing radius alpha and check that the resulting union of disks always has trivial 1-dimensional homology without cycles?

伦理问题详情

This paper is a minor revision of the submission to NeurIPS 2024, where one of the reviews provided a counter-example to Theorem 4.1. Submitting almost the same paper without correcting this claim raises serious ethical concerns.

评论

We thank the reviewer for their review. We will reply to their comments below.

Lines 177-178: "the filtration value of a k-simplex in the α-filtration is the radius of its circum-k-sphere, which differs from its filtration value in the VR filtration". The concept of a "filtration value" appears here for the first time and is never defined in the paper, which leads to the major misunderstanding of persistence.

Thank you for raising this issue. We have added a sentence explaining that the value a simplex σ\sigma first appears in the filtration is called its filtration value σ\sigma. I.e., for a filtration FF, we have that F(σ)=inf(εR:σFε)F(\sigma)=\inf (\varepsilon \in \mathbb{R}: \sigma\in F_\varepsilon).

We apologise, but what is the major misunderstanding of persistence the reviewer is referring to?

Even a non-expert can notice that n-th dimensional (?) homology class cannot be generated by 2-dimensional (?) simplices for n>2.

Thank you for pointing out this typo, it should read nn-simplices instead of 22-simplices as used throughout the proof. We have fixed this.

Also, Theorem 4.1(i) fails for n=1 and any 3 points on a unit circle that form a non-acute triangle.

We realise now that this confusion stems from our definition of alpha-complexes (Section 2: Vietoris--Rips and α\alpha-complexes and Definition B.2 in our original submission) being slightly different from the definition of alpha complexes in the literature and current implementations (Including of the libraries TOPF uses). In order to avoid confusion, we have now renamed our original construction to "α\alpha^*-complexes" (Definition B.2, and Section 2: Vietoris--Rips and α\alpha-complexes), gave a definition of the literature version of α\alpha-complexes (Definition B.4)and discussed the differences. We now formulate Theorem 4.1 in terms of the α\alpha^*-filtration and give conditions for when the theorem holds for α\alpha-complexes as well. We want to reiterate that our theorem is true in its current form, and was true in the original submission when using the definitions given in the paper.

There are infinitely many point clouds whose 1-dimensional persistence is empty. Look for a generic family of such clouds in J Applied and Comp Topology 2024.

We appreciate the reference, but we fail to see the relation to our paper. There are infinitely many point clouds whose 11-dimensional persistence is non-empty, and a wealth of literature in TDA proves that these point clouds with topological structure appear in many applications, see our introduction for more references, or surveys like [1] or [2]. Clustering methods work only on data with a cluster structure, image classification methods only on images which are different to noise, and so on. Still all of these methods are valuable for the data they are designed for, and the same holds for TOPF.

[1] Elizabeth Munch. A user’s guide to topological data analysis. Journal of Learning Analytics, 2017

[2] Larry Wasserman. Topological data analysis. Annual Review of Statistics and Its Application, 2018.

Even Figure 3 in the paper demonstrates that only 3 points in the 1D persistence represent reasonable cycles, hundreds of others represent only noise.

We apologise, but we do not understand the concern of the reviewer here. In the cited example, our method is able to extract the three relevant persistent homology classes and build topological features using them while discarding the noise ones by considering the birth and death times of the persistence barcodes. This is exactly as persistent homology and our method was designed to perform. Figure 5b) shows that the output of TOPF represents meaningful topological structure.

评论

In fact, there is no sense to apply persistence to proteins whose atoms are ordered. In this ordered case, the complete isometry invariant of a point cloud is the classical distance matrix requiring only quadratic time, which was known at least since 1935, see I.Schoenberg, Annals of Mathematics, 724-732, 1935.

We apologise, but we are again very surprised by both the content as well as the style of this remark. There is a wealth of literature of the successful use of persistent homology on protein data, for example [1--6]. The goal of our paper, as stated in the abstract is that we "propose a novel method to extract node-level topological features from complex point clouds using discrete variants of concepts from algebraic topology and differential geometry." By node-level we mean assigning a feature vector to every single point, as discussed in the introduction, Figure 1 and the remaining sections. This is very different from trying to provide an isometry invariant of point clouds.

[1] Xia, Kelin, and Guo‐Wei Wei. "Persistent homology analysis of protein structure, flexibility, and folding." International journal for numerical methods in biomedical engineering 30.8 (2014): 814-844.

[2] Kovacev-Nikolic, Violeta, et al. "Using persistent homology and dynamical distances to analyze protein binding." Statistical applications in genetics and molecular biology 15.1 (2016): 19-38.

[3] Cang, Zixuan, and Guo-Wei Wei. "Analysis and prediction of protein folding energy changes upon mutation by element specific persistent homology." Bioinformatics 33.22 (2017): 3549-3557.

[4] Pun, Chi Seng, Si Xian Lee, and Kelin Xia. "Persistent-homology-based machine learning: a survey and a comparative study." Artificial Intelligence Review 55.7 (2022): 5169-5213.

[5] Ichinomiya, Takashi, Ippei Obayashi, and Yasuaki Hiraoka. "Protein-folding analysis using features obtained by persistent homology." Biophysical Journal 118.12 (2020): 2926-2937.

[6] Bou Dagher, Léa, et al. "Persistent homology reveals strong phylogenetic signal in 3D protein structures." PNAS nexus 3.4 (2024).

The persistence homology is more suitable for clouds of unordered points but is much weaker than the collection of all pairwise distances, which determine any generic point cloud in general position uniquely under isometry, which has been known in computational geometry for more than 20 years, see Boutin and Kemper, 2004.

As discussed above, we are not interested in providing a global isometry invariant. We are unsure how this misconception arose.

The stronger invariant (a local distribution of distances) has been widely studied in shape analysis, e.g. see Memoli, "Gromov–Wasserstein distances and the metric approach to object matching" (FocM 2011), extended to a complete and polynomial-time invariant under rigid motion in any fixed Euclidean R^n (CVPR 2023).

See above.

The initial obstacle for the proposed research is the lack of a rigorous problem statement.

As discussed above, we describe our goal (among other places) already in the abstract and in the contributions as to develop "a novel method to compute node-level topological features relating individual points to global topological structures of point clouds." We then spend section 2 explaining what we mean by the "global topological structure of points clouds". Because the notion of topological structure requires some background in algebraic topology, and the relation to the individual points some background in (discrete) differential geometry, we do not give a fully formal definition in the introduction.

The attempts to state problems in the final section seem a bit late:

Lines 514-520: "selection of the relevant features is a very hard problem" and "efficient computation of simplicial weights leading to the provably most faithful topological point features is an exciting open problem"

These quotes are from our limitations and future work sections at the end of the paper and thus do not represent our main contribution or the main problem our paper tries to solve. As discussed above, we formulate the goal of our paper for example in the abstract and in the introduction. However, we will make sure to use the word "problem" in our introduction when discussing the goal of the paper in order to make searching for it easier.

评论

Lines 514-520: "selection of the relevant features is a very hard problem" and "efficient computation of simplicial weights leading to the provably most faithful topological point features is an exciting open problem" The words "relevant" and "most faithful" should be rigorously defined, else anyone can claim that their features are "relevant" and "most faithful".

The reviewer seems to mistake our intention in the quotes above. The full quotes are

Limitations [...] Finally, selection of the relevant features is a very hard problem. While our proposed heuristics work well across a variety of domains and application scenarios, only domain- and problem-specific knowledge makes correct feature selection feasible.

Future Work [..] Furthermore, efficient computation of simplicial weights leading to the provably most faithful topological point features is an exciting open problem.

One of the reasons that selecting the relevant features is hard is that there is no unique definition of relevant features across different application domains. Thus, we conclude in the paper that while our heuristics works well, identifying the "correct features" requires domain- and problem-specific knowledge. We are trying to be very open about these limitations and challenges and at no point claim to select provably relevant features across all possible application domains. In particular, we have devoted section H of the appendix to discussing limitations of our method in detail. We apologise, but we do not understand the concern of the reviewer with our claims.

Can the authors to consider 3 points (1,0), (0,1), (-1,0), draw three disks centered at the points with a growing radius alpha and check that the resulting union of disks always has trivial 1-dimensional homology without cycles?

As discussed above, we have realised that we used a non-standard but a closely related definition of α\alpha-complexes. In particular, our definition assigned every simplex a filtration value of its circumradius, as the definition in the literature does. However, the literature definition then updates the filtration value of some select "Gabriel" simplices below the top dimension to the filtration value of one of their co-faces. In order to avoid confusion, we now call our previous definition α\alpha^*-filtrations, give a definition of α\alpha-filtrations and discuss the relation between the two. In particular, both filtrations operate over the same set of simplices and assign the same filtration values to all top-dimensional simplices and all non-Gabriel simplices. We now formulate Theorem 4.1 for alpha^*-filtrations and give conditions for when the theorem applies to alpha-filtrations as well. We thank the reviewer for helping us to find this issue.

We will now try to explain why your situation does not pose a counterexample to the theorem. According to the definitions used in our paper (now called the α\alpha^*-filtration), the filtration value of a 22-simplex is its circumradius r. A fact from elementary geometry tells us that r is at least the maximum of a/2a/2,b/2b/2 and c/2c/2, for aa, bb, cc the length of the edges of the simplex. a/2a/2, b/2b/2, and c/2c/2 are the filtration value of the edges. Equality holds if and only if the points form a right-angled triangle, which happens iff two points lie on diametrically opposed points of the circumsphere of the simplex. When sampling from this circle, this case has probability mass 00, which is why our theorem is formulated only with probability 11. In all other cases, the α\alpha^*-filtration value of the edges are smaller than the filtration value of the triangle, and a non-trivial persistent homology class forms in the α\alpha^*-filtration.

Now let us consider the α\alpha-filtration. For an acute triangle, there are no Gabriel edges and the claim from the α\alpha^*-filtrations holds for the α\alpha-filtration as well. For an obtuse triangle however, one of the edges is a Gabriel edge, meaning it gets assigned the (larger) filtration value of the 22-simplex. We have updated the theorem to reflect our new definitions and naming scheme.

We want to emphasise that both the α\alpha, as well as the α\alpha^*-filtration assigns filtration values based on the circumradius of some kk-simplex, i.e. the radius of the (k1)(k-1)-sphere defined by the k+1k+1 vertices of the simplex. This is a different sphere than the spheres obtained drawing circles around the vertices of the simplex.

We have tried to very carefully explain our reasoning and the relevant definitions. In case you have further questions, we will be happy to answer them. As discussed above, from our point of view a main part of your review does not appear to relate to the central points and ideas of our paper. We would be very interested in hearing your opinion on this, especially on the apparent "major misunderstanding of persistence". Thank you in advance for your reply!

评论

The original submission talked only about alpha-filtrations of alpha-complexes below, though without giving necessary references to the original paper https://ieeexplore.ieee.org/abstract/document/1056714 or later surveys of alpha-shapes, which are also missed in the revision.

Before Theorem 4.1 in the original submission, "alpha-filtration" appeared only in the following lines.

Lines 38-39: "the persistent homology is computed on an α/VR-filtration."

Lines 177-178: "We note that the filtration value of a k-simplex in the α-filtration is the radius of its circum-k-sphere"

Lines 233-265: "we first compute the persistent k-homology modules Pk including a set of homology representatives Rk of X using an α-filtration for n ≤ 3 and a VR filtration for n > 3."

Hence alpha-complexes were interpreted in the classical sense quoted below from the appendix.

Lines 767-770: "Definition B.2 (α-complex of a point cloud). Given a finite point cloud X in real space Rn, the α-complex α_ε(X) is the subset of the n-dimensional Delaunay triangulation DT(X) consisting of all k-simplices σ_k ∈ DT(X) with a radius r of its circumscribed (k − 1)-sphere with r ≤ ε for all k ≤ n."

This definition should be in the main paper because Theorem 4.1 essentially required alpha-complexes and "reviewers are not required to read the appendix", see https://iclr.cc/Conferences/2025/CallForPapers

So the counter-example to Theorem 4.1 on the points (1,0), (0,1), (-1,0) was for the classical case.

Is it correct that the revision only replaced alpha in Definition B.2 with alpha^*?

Lines 795-798 (revision): "Definition B.2 (α^-complex of a point cloud). Given a finite point cloud X in real space Rn, the α^-complex α^*_ε(X) is the subset of the n-dimensional Delaunay triangulation DT(X) consisting of all k-simplices σ_k ∈ DT(X) with a radius r of its circumscribed (k − 1)-sphere with r ≤ ε for all k ≤ n."

Because alpha-complexes were introduced more than 40 years ago, it seems a bit too late to propose a new name alpha^*-complex for the same object and give a new definition below for the 40-year-old concept.

Lines 805-807 (revision): "Definition B.4 (α-complex of a point cloud). Given a finite point cloud X in real space Rn, the α- complex α_ε(X) is the α^*-complex of V , except that the filtration value α(σ) of all Gabriel simplices σ with associated points X in the interior of the circumsphere of σ is given by min_x∈X α(σ ∪ {x})."

If you would like to introduce a new complex, feel free to choose another name, which sufficiently differs from the alpha-complex to avoid potential confusion and misunderstanding.

New Definition B.4 based on Gabriel simplices was not in the original submission, which can be checked by searching the word "Gabriel" (appears only as a name in the bibliography of the original submission).

By node-level we mean assigning a feature vector to every single point, as discussed in the introduction, Figure 1 and the remaining sections. This is very different from trying to provide an isometry invariant of point clouds.

Is it correct that all your simplicial complexes, persistent diagrams, and feature vectors (assigned to points) are preserved under isometry (any distance-preserving transformation) of a given point cloud? If yes, then your outputs are isometry invariants.

In this case, it is a matter of scientific integrity at least to mention other isometry invariants of point clouds, especially because distance-based invariants are easier, faster, and stronger than persistence, also known for more than 20 years and studied at the point level under the names of local distribution of distances and local shape descriptors.

wealth of literature of the successful use of persistent homology on protein data, for example [1-6]

Did you mean that this "wealth of literature" used complicated invariants (persistent homology) of unordered point clouds on atoms of protein chains that are naturally ordered along a protein backbone and hence have complete linear-size invariants?

wealth of literature in TDA proves that these point clouds with topological structure appear in many applications, see our introduction for more references, or surveys like [1] or [2]

Have [1,2] explained that persistence is an isometry invariant of a point cloud? If not, other sources can also be helpful.

The authors should be praised for their effort to correct original Theorem 4.1. Do you accept the counter-example on 3 points to the original statement for the classical alpha-filtration?

Because this substantial update took more than a week and the experimental results might be different for the newly proposed filtration,"reviewers reserve the right to ignore changes that are significant from the original scope of the paper", see https://iclr.cc/Conferences/2025/CallForPapers. The authors are kindly encouraged to more substantially revise the paper for a future submission by using appropriate names and including necessary references.

评论

Thank you for your reply. We apologise, but we believe that there has been some confusion and many of your points can already be answered from our previous rebuttal. We will nonetheless reply to your points.

The original submission talked only about alpha-filtrations of alpha-complexes below, though without giving necessary references to the original paper https://ieeexplore.ieee.org/abstract/document/1056714 or later surveys of alpha-shapes, which are also missed in the revision.

Thank you, we have now included referenes on the alpha-filtrations in the paper. What to you mean by "only"?

Before Theorem 4.1 in the original submission, "alpha-filtration" appeared only in the following lines.

Lines 38-39: "the persistent homology is computed on an α/VR-filtration."

Lines 177-178: "We note that the filtration value of a k-simplex in the α-filtration is the radius of its circum-k-sphere"

Lines 233-265: "we first compute the persistent k-homology modules Pk including a set of homology representatives Rk of X using an α-filtration for n ≤ 3 and a VR filtration for n > 3."

Hence alpha-complexes were interpreted in the classical sense quoted below from the appendix.

Lines 767-770: "Definition B.2 (α-complex of a point cloud). Given a finite point cloud X in real space Rn, the α-complex α_ε(X) is the subset of the n-dimensional Delaunay triangulation DT(X) consisting of all k-simplices σ_k ∈ DT(X) with a radius r of its circumscribed (k − 1)-sphere with r ≤ ε for all k ≤ n."

First, we would like to mention that the description of line 177/178 in the main text paraphrases the definition in line 767-770, except for a typo we have fixed.

As written in our rebuttal above, there is a technical difference between the above-quoted definition from our appendix and the "classical" definition of alpha-filtrations from the literature. Thus we do not know what you mean by the above comment. There are multiple equivalent ways of defining alpha-filtrations:

The most classical way defines alpha complexes as the geometric realisation of the restricted Voronoi tesselation of the point cloud in Rn\mathbb{R}^n. However, any implementations of alpha complexes use a computationally more accessible definition in terms of Gabriel simplices and the Delaunay triangulation, which we now give in Definition B.4. (See for example [1] for a proof of the equivalence.)

[1] Kerber, Michael, and Herbert Edelsbrunner. "3D kinetic alpha complexes and their implementation.", 2013

Is it correct that the revision only replaced alpha in Definition B.2 with alpha^*?

As written in the rebuttal, we have now renamed what we previously called α\alpha filtrations to alphaalpha^*-filtrations to reflect the technical difference in definition.

Because alpha-complexes were introduced more than 40 years ago, it seems a bit too late to propose a new name alpha^*-complex for the same object and give a new definition below for the 40-year-old concept.

There still seems to some confusion. We want to reiterate the statement from our rebuttal and our above comment: What we now call α\alpha-complex in the paper reflects what has been called α\alpha complexes in the literature. Definition B.2 is not a definition for what is known as α\alpha-complexes. (Although it looks like it is.)

Lines 805-807 (revision): "Definition B.4 (α-complex of a point cloud). Given a finite point cloud X in real space Rn, the α- complex α_ε(X) is the α^*-complex of V , except that the filtration value α(σ) of all Gabriel simplices σ with associated points X in the interior of the circumsphere of σ is given by min_x∈X α(σ ∪ {x})."

If you would like to introduce a new complex, feel free to choose another name, which sufficiently differs from the alpha-complex to avoid potential confusion and misunderstanding.

As we wrote in the rebuttal, definition B.4 is a valid definition of α\alpha-complexes. It is in fact the definition used in implementations, for example in gudhi [2]. (Note however, that the implementation in gudhi chooses to square the alpha values.) If you feel like we made a mistake in the definition, please let us know exactly what you would want us to change.

[2] https://gudhi.inria.fr/doc/latest/group__alpha__complex.html

New Definition B.4 based on Gabriel simplices was not in the original submission, which can be checked by searching the word "Gabriel" (appears only as a name in the bibliography of the original submission).

Yes, this is what we wrote in the rebuttal:

In order to avoid confusion, we have now renamed our original construction to "α\alpha^*-complexes" (Definition B.2, and Section 2: Vietoris--Rips and α\alpha-complexes), gave a definition of the literature version of α\alpha-complexes (Definition B.4) and discussed the differences.

As mentioned above, Gabriel simplices appear in the definition of α\alpha-filtrations most suited for direct practical applications.

评论

By node-level we mean assigning a feature vector to every single point, as discussed in the introduction, Figure 1 and the remaining sections. This is very different from trying to provide an isometry invariant of point clouds.

Is it correct that all your simplicial complexes, persistent diagrams, and feature vectors (assigned to points) are preserved under isometry (any distance-preserving transformation) of a given point cloud? If yes, then your outputs are isometry invariants.

Yes, these feature vectors are preserved under isometries, just as many other constructions like the clustering labels of any non-random clustering algorithm, some statistical descriptors, the number of points, or any function in the distances of the point cloud.

The class of "isometry invariants" of point clouds is thus a very diverse set of constructions with many different goals. When the literature talks about finding new isometry invariants of point clouds, they usually aim to find a single complex descriptor which completely distinguishes all non-isometric point clouds and is itself being preserved under isometries, ideally with low computational complexity. This is a very specific goal and very different from the goal of our paper. This is what we meant to express by the above quote, we apologise for not being specific enough.

In this case, it is a matter of scientific integrity at least to mention other isometry invariants of point clouds, especially because distance-based invariants are easier, faster, and stronger than persistence, also known for more than 20 years and studied at the point level under the names of local distribution of distances and local shape descriptors.

As we discussed above, there is a wide class of isometry invariants which try to solve very different poblems, often very different from the goals of our paper. We already benchmark against other "isometry-invariants" like diverse clustering algorithms and the local shape descriptors extracted by pgeof, and discuss various "isometry invariants" from TDA in the related work section. We will however add a reference to the suggested 2011 Memoli paper in the related work. Thank you for this reference!

Did you mean that this "wealth of literature" used complicated invariants (persistent homology) of unordered point clouds on atoms of protein chains that are naturally ordered along a protein backbone and hence have complete linear-size invariants?

To the best of our knowledge, yes. If the goal of the papers were simply to distinguish different proteins based on their associated point clouds, your objection would be valid. However for many more complex tasks, a better descriptor which is robust and organises the information in a good way is needed. It turns out that persistent homology is such a descriptor.

wealth of literature in TDA proves that these point clouds with topological structure appear in many applications, see our introduction for more references, or surveys like [1] or [2]

Have [1,2] explained that persistence is an isometry invariant of a point cloud? If not, other sources can also be helpful.

These surveys have not explicitly stated this. (However it is obvious from the fact that VR-persistence diagrams are functions in the distance matrix.) We suppose that this omission is due to the perceived relevance of this to the topic discussed in the surveys. What do you mean by stating "If not, other sources can also be helpful."?

Do you accept the counter-example on 3 points to the original statement for the classical alpha-filtration?

As discussed above, the theorem pre-revision was proven using the definitions given in the paper (and cited by you above), whereas the theorem in the revised version now holds for and is proved using the corrected definitions.

Because this substantial update took more than a week and the experimental results might be different for the newly proposed filtration,"reviewers reserve the right to ignore changes that are significant from the original scope of the paper", see https://iclr.cc/Conferences/2025/CallForPapers.

We do not know how to respond to this. We have never seen a ML conference rejecting a paper because the revision took a week. (In many ML conferences, the rebuttals will even only get published after a week, so should every paper be rejected then?) Furthermore, as we write in our rebuttal, we made a technical adaption to one of the definitions so that it reflects the definition used in implementations (including ours) and literature. This does not change any of the experiments. In particular, the scope of the paper remains virtually unchanged, and we have added only further evidence for our contributions based on feedback by the reviewers.

审稿意见
8

This paper develops a method to assign topological information to each point xx of a given point cloud XRdX \subset R^d a vector.

Roughly speaking, assuming that XX exhibits a set F\mathcal{F} of significant topological features (connected component, loops, or cavities, etc.), the idea is to quantify for each xXx \in X to which extend it contributes to each topological feature fFf \in \mathcal{F}. Eventually, each xXx \in X is thus turned into a vector in dimension F\mathcal{F} that describes the relation between the local point xx and the global (topological / geometric) structure of XX.

The construction is performed by relying on the Hodge decomposition theorem (in discrete form) and eventually boils down to a (actually two) least square problem(s) which can be solved efficiently.

The approach is then showcased in a broad range of experiments. In particular, a "topological clustering benchmark" is introduced.

优点

  • Good introduction, clear and nicely written
  • I really appreciate the step-by-step description of Section 3, making the flow of the exposition pretty nice despite the technicality of the construction.
  • Nice illustrations overall.
  • Elegant and fairly-well motivated construction.
  • Nice set of experiments. The introduction of a "topological clustering benchmark" is a valuable contribution of the work, we need more of these in the community.

缺点

  • The approach depends on several seemingly important hyper-parameters. This may be a necessary price to pay, but belongs to its weaknesses nonetheless.
  • Though acknowledged by the authors, the setting of Theorem 1 is very idealized. In particular, the nn-sphere has somewhat "maximal reach" and it is less clear to me that the theorem remain (with probability <1< 1 but still reasonable) valid when the underlying manifold is a "pinched" sphere with low reach. The theorem remains nice nonetheless and its validity in applications is not crucial to the work.

Minor (not real weaknesses)

  • Typo : "hull hull" in Theorem 1
  • In the first experiment (Topological Point Cloud Clustering Benchmark), I believe that including topological regularization term in point embedding methods (node2vec and pointnet) has been attempted in some works (e.g. in Topological Autoencoder of Moor et al., some experiments in Optimizing Persistent Homology based function of Carriere et al., Topological Node2vec of Meehan et al., etc.). I believe that this would not invalidate the claims made in that paragraph, namely that these methods requires specific training to work and may be slower than the proposed approach, but it may be worth to briefly mention these.

问题

  1. Points vv in XX are vectorized in dimension F|\mathcal{F}|, where F\mathcal{F} designates the set of "most significant topological features", typically detected with the proposed heuristic. However, I believe that this embedding dimension may be somewhat unstable to even small perturbation of XX (e.g. XX is made of a circle of radius 2 and a circle of radius 1; then the persistence of the big circle is twice the one of the smaller one, which is itself twice the one of small circles arising from the noise). Similarly, if one (theoretically) refrain from using an heuristic and pick all topological features for F\mathcal{F}, the dimension is (clearly, this time) unstable as well. Is there a way to somewhat "smooth" the choice of F\mathcal{F} to avoid relying on a hard threshold but rather considering (possibly infinitely) many dimensions but whose relative importance would be weighted by i\ell_i? I ask this question out of curiosity, I do not expect an extensive answer in the rebuttal.
评论

We thank the reviewer for their review and their feedback! We put a lot of effort into all of these points, and thus are very happy about the positive things the reviewer had to say about our introduction, the step-by-step description of section 3, our illustrations, the construction of TOPF and the set of experiments. We have uploaded a revised version of the paper in response to your and the other reviewers' feedback and will reply to your comments below.

Though acknowledged by the authors, the setting of Theorem 1 is very idealized. In particular, the -sphere has somewhat "maximal reach" and it is less clear to me that the theorem remain (with probability < 1 but still reasonable) valid when the underlying manifold is a "pinched" sphere with low reach. The theorem remains nice nonetheless and its validity in applications is not crucial to the work.

We agree with the reviewer that the setting of Theorem 1 is very idealised. Proving more general guarantees appears to be very hard due to the interplay of possible complex topological structures, noise, the non-linear nature of the underlying Delaunay triangulation of the alpha-complex, and the continuous nature of the harmonic representations.

Please note that another reviewer made us aware that our definition of α\alpha-complexes had some technical differences to the definition of alpha complexes used in implementations and the literature. In order to avoid confusion and be open about this, we have renamed our previous definition to α\alpha^*-complexes, and introduced the more commonly used definition as α\alpha-complexes. We now state Theorem 4.1 for α\alpha^*-complexes and give conditions on when it holds for α\alpha-complexes as well.

Typo : "hull hull" in Theorem 1

Thank you for catching this typo, we have fixed it!

In the first experiment (Topological Point Cloud Clustering Benchmark), I believe that including topological regularization term in point embedding methods (node2vec and pointnet) has been attempted in some works (e.g. in Topological Autoencoder of Moor et al., some experiments in Optimizing Persistent Homology based function of Carriere et al., Topological Node2vec of Meehan et al., etc.). I believe that this would not invalidate the claims made in that paragraph, namely that these methods requires specific training to work and may be slower than the proposed approach, but it may be worth to briefly mention these.

Thank you for these related references! We will include a discussion of them in a future work! In Topological autoencoders, the authors basically set up a topological loss functions that ensures that topological features of the original point cloud are preserved in the latent space of the autoencoder. This is different from the approach TOPF takes, as TOPF tries to detect and extract topological features. Similar things can be said about Topological Node2Vec and some experiments in "Optimizing Persistent Homology Based Function", or some experiments of the newer "Diffeomorphic interpolation for efficient persistence-based topological optimization" by Carriere, Theveneau, and Lacombe: They all try to come up with ways to preserve global topological structure while embedding a point cloud, while we on the other hand construct the embedding based on the relationship between the points and the global topological structure.

We will add a discussion of this to the paper and thank you for the suggestion!

  1. [..] Is there a way to somewhat "smooth" the choice of to avoid relying on a hard threshold but rather considering (possibly infinitely) many dimensions but whose relative importance would be weighted by i\ell_i? I ask this question out of curiosity, I do not expect an extensive answer in the rebuttal.

We thank the reviewer for this excellent question! You are correct that in certain circumstances the number of features picked by the heuristics is not stable. In general, it seems impossible to design a heuristic that is stable in this regard without any prior knowledge on the data. However, the current implementation of TOPF already permits to extract a fixed number NdN_d of topological features per dimension for reasons outlined in your question. When then weighing the extracted features by a suitable weight function in the length of their associated persistence bar (like exponential/polynomial functions of high enough d), this feature will converge for large NdN_d. It is very plausible that this is the desirable approach in some applications, especially those requiring fixed-dimensional features/positional encodings. Choosing the correct weight function seems like an interesting research question for future work. We have added a discussion of this to appendix F and consider adding a brief discussion to the main text.

In case you have further ideas on this, we would be very happy to hear them!

Thank you again for your feedback! We are happy to answer any follow-up questions.

评论

Dear reviewer, Thank you again for the detailed feedback and review! Based on the feedback received from you and the other reviewers, we believe to have further strengthened the paper. As the discussion phase closes soon, we would be happy to hear whether you have any follow-up questions or comments!

审稿意见
5

Paper proposes a novel idea for extraction of continuous topological features via projection of simplicial generators into harmonic subspaces. It works as follows. First, given a point cloud sample from a manifold, it computes persistent homology classes based on alpha or VR filtration of the point cloud. Second, it selects the most relevant classes based on the smallest quotient heuristic. Third, these relevant global classes are used to compute local topological features that indicate local ‘strength’ of chosen topological class. This is done via aggregation of simplices for chosen birth-depth timestep width and their projection into harmonic subspace of simplices. Finally, simplicial features are aggregated on point level to produce pointwise features. These features can be utilized to provide topology-aware visualizations and/or topology-aware clustering.

Proposed method evaluates the Topological Clustering Benchmark Suite (TCBS) that consists of 7 point clouds that represent complex topological arrangements.

POST DISCUSSION UPDATE

I strongly appreciate replies by the authors and all additional work and experiments they have done. However, I think that paper still cannot be accepted in the current state because it need substantial rework. I am especially concerned about fair comparison to learning-based baselines using point cloud normalizations that were used to train them.

If this would be a journal paper I would recommend 'resubmit a major revision' but for conference I have to recommend reject.

Below are final justifications/suggestions (up to authors' discretion):

  1. Main text of the paper needs to be adjusted to better balance theory and practical considerations of the method. I recommend checking the following paper as examples on how to balance heavy mathematical background and practical aspects Original diffusion models paper: https://arxiv.org/abs/1503.03585 Product Manifold learning: https://arxiv.org/pdf/2010.09908 SGLD: https://www.stats.ox.ac.uk/~teh/research/compstats/WelTeh2011a.pdf As bare minimum, I recommend removing Theorem 4.1 completely and moving Figures 13 and 15 to main paper.

  2. Experimental evaluation should use 'native' normalizations for all baselines. TOPF also ideally evaluated in 'standard' normalization space of baselines, otherwise there is a concern that hyperparameters of TOPF might require adjustment to 'standard' normalization.

  3. I strongly recommend including results on sampling regularity and sampling density (for baselines as well) in the main paper.

  4. I strongly recommend including stronger baselines like DGCNN.

  5. TCBS benchmark data might benefit from additional examples.

  6. I have also checked authors' discussion with reviewer smnw and I agree with the following reviewer's points: alpha*-complexes might be confusing both for readers who are familiar and who are unfamiliar with the theory; proposed work seem to be proposing isometric invariant and thus local shape descriptors probably should be mentioned in related work. That being said, I still think that main focus of the paper should be empirical validation of the paper because theoretical contribution seems to be very limited.

优点

— Novel idea for extraction of continuous topological features via projection of simplicial generators into harmonic subspaces and subsequent extraction of pointwise features; — Good visual presentation of the method; — Somewhat good introduction into fundamentals of algebraic topology; — Ablation with respect to some of the hyperparameters of the method; — Quantitative results on the proposed TCBS benchmark seem to be strong.

缺点

— Main body of the paper focuses too much on introduction into algebraic topology and too little on practical considerations of the method; — Algorithm 1 is too schematic and it is hard to reproduce the method based on it. It can heavily benefit with more detailed substeps; — Applicability of proposed method to high-dimensional spaces (D>3) seems to be limited; — Related work discussion (including supplement) is limited; I recommend including some of the work discussed in [Rev 1] — Theorem 4.1 has very limited practical applicability for 3D data analysis; — No ablation with respect to sampling density; — No ablation with respect to non-uniform sampling from the manifold; — Learning-based baseline (PointNet) is very weak; — Important details in some figures are missing (see suggestions/questions).

[Rev1] Wasserman, L., 2018. Topological data analysis. Annual Review of Statistics and Its Application, 5(1), pp.501-532.

问题

My main issue with the paper is that it focuses too much on the theory of algebraic topology that works with continuous spaces; and focuses too little on practical considerations that stem from the discretization aspect of topological data analysis. Overall, the proposed model is built upon discrete approximation of the manifold through N-simplices based on localized Delaunay triangulation based on epsilon radius neighbors. It means that two considerations are important for the method. First, number of points: we want enough points to approximate manifold topology but not too much because dense sampling can be expensive or unavailable. Second, sampling density: we want points to be sampled somewhat uniformly from the manifold (if all samples are the same point, it does not help us much). Good TDA method should ideally require as few points as possible to recover structure of the manifold while being robust to non-uniformity and noise in sampling. I think the current iteration of the paper largely ignores these considerations. Namely:

— How sampling density affects performance of the method? For example, what happens with predicted labels in Figure 8 if we reduce sample sizes of point clouds by 2x? 3x? 5x? My concern is that the model might not work well on sparse point clouds.

— Similarly, what happens with the method if we have non-uniform sampling from the manifold? For example, what happens when the target manifold has a high-curvature area that is not sampled densely enough?

— My understanding is that the epsilon radius of local Delaunay triangulation (Definition B.2.) is the hyperparameter of the method. It is not clear how this parameter should be chosen and it looks like it can strongly affect the results: if it is too large, local Delaunay triangulation will start closing holes; if it is too small, the manifold will be disconnected. Do authors have some practical suggestions how this radius should be chosen, especially if we don’t have good priors about the data?

— Theorem 4.1 has very limited practical applicability to actual 3D data analysis. It assumes two sets of points: one set is sampled from the unit sphere; another set is sampled with distance at least 2 to this unit sphere. For the majority of 3D point cloud applications, point clouds are normalized to unit cube. My understanding is that in this case assumptions of Theorem 4.1 do not hold. Is my understanding correct? Maybe the setting that I have described can be adapted for Theorem 4.1 somehow?

— On a related note, what is the normalization of the data for the experiments? Are point clouds normalized to unit cube? Is this normalization the same across all evaluated methods?

— PointNet baseline is very weak. It was introduced in 2016 and was superseded by a large number of architectures: PointNet++, several versions of PointTransformers and other attention-based architectures. PointNet performance on the majority of benchmarks is significantly worse than for all of the mentioned methods. More up-to-date method can give significantly better performance on proposed benchmark.

— Applicability of the method to high-dimensional spaces seems to be very limited. How well does the proposed method scale with the number of dimensions? My understanding is that local Delaunay triangulation will be significantly slower for higher dimensions. It is also not clear, how to choose the triangulation radius.

— On a related note, the experiment in Figure 11 has very limited applicability. To analyze image features with TOPF one needs to train additional variational autoencoder to map image features in 3D space. This scenario does not seem to be practically appealing What happens if relevant information cannot be contained in 3 dimensions? What if one wants to run TDA on original features?

— What dictated the sampling density of the TCBS dataset? Point clouds in Figure 7 have a number of samples that can differ by order of magnitude: 267 points for SphereInCircle and 4600 points for 2Spheres2Circles. Does method assume/TCBS data assume some upper bound on minimum distance to nearest neighbor for each point and that is what dictates sampling density? Or is it something else?

— Figure 5 is missing information about the average distance from each point to its nearest neighbor. Without this information, it is hard to assess how large is the relative value of the noise compared to manifold sampling density.

— Figure 10 is missing time units, so it is impossible to infer speed of the method from it. Also, how indicative are those running times (they are from two point clouds from the benchmark)?

— Figure 13 seems to be missing 9 of 12 subfigures.

评论

Thank you very much for your detailed review and the many good questions and suggestions! In particular, we are of course very happy that you commended liked quite a few things about our paper. :) We have uploaded a revised version of the paper based on the feedback, and will reply to your points below

Algorithm 1 is too schematic and it is hard to reproduce the method based on it. It can heavily benefit with more detailed substeps;

Thank you for this suggestion. We will add details and substeps for Algorithm 1 and thus try to make it easier to reproduce the algorithm. Because TOPF and the involved concepts are rather complicated, we are afraid that a complete reproduction of the algorithm is not possible without reference to the explanations of section 3 or the submitted code.

How sampling density affects performance of the method? For example, what happens with predicted labels in Figure 8 if we reduce sample sizes of point clouds by 2x? 3x? 5x? My concern is that the model might not work well on sparse point clouds.

Thank you very much for this valid and valuable remark! In the selection and construction of the point clouds of the TCBS, we have already tried to incorporate very different sample sizes, from very sparse point clouds with 158 points (Ellipses) to 4600 points (2Spheres2Circles). Based on your comment, we have now introduced an experiment (See Figure 15) where we decrease the sampling density for our point cloud with the highest original sampling density (2Sphers2Circles). It shows that TOPF still maintains an ARI over 0.9 even for 700 points, which is a more than six-fold decrease in sampling density, and still performs reasonably well for even stronger downsamplings. We believe that the robustness from TDA thus allows TOPF to perform well on sparse sets, especially when compared with modern Computer Vision techniques using neural architectures.

Similarly, what happens with the method if we have non-uniform sampling from the manifold? For example, what happens when the target manifold has a high-curvature area that is not sampled densely enough?

Thank you again for this comment! Already in the TCBS we have included datasets with samling irrgularities: In Ellipses, the small 2 ellipses are sampled much more densely than the outer ellipses, in Spheres+Grid, the middle grid is sampled much more sparsely than the remaining blobs, and finally in Spaceship, the two small ellipsoids which look like ears are sampled much more densely than the remaining ellipsoids.

The quantitative experiments seem to indicate that TOPF can adapt well to these cases, at least if every topological feature has a homogenous enough sampling density/noise level. The reason for this is that by making use of the birth and death time of the associated bars in the persistent homology diagram, TOPF can pick a scale/alpha radius for each of the features independently. (See our answer to the next point.)

My understanding is that the epsilon radius of local Delaunay triangulation (Definition B.2.) is the hyperparameter of the method. It is not clear how this parameter should be chosen and it looks like it can strongly affect the results: if it is too large, local Delaunay triangulation will start closing holes; if it is too small, the manifold will be disconnected. Do authors have some practical suggestions how this radius should be chosen, especially if we don’t have good priors about the data?

The epsilon radius of Delaunay triangulation is not a hyperparameter of the method and quite a nice trick (we think.) All hyperparameters are listed in appendix D "Hyperparameters". Every topological feature corresponds to a bar with a death time dd and a birth time bb in the persistent homology diagram. Thus, for each feature we have a range of possible alpha-radii where the feature is guaranteed to exist. Using the interpolation-parameter lambdalambda as hyperparameter with default λ=0.3\lambda = 0.3, we then choose as α\alpha-radius r=b1λdλr=b^{1-\lambda}d^\lambda, the weighted geometric mean. In dimension 00, all birth times are 00, hence we have to pick the weighted arithmetic mean. We do a sensitivity study on lambda in the appendix and discuss this procedure in section 3 step 3. **In conclusion, TOPF makes use of persistent homology to choose a provably valid α\alpha-radius for every feature.$$

评论

Theorem 4.1 has very limited practical applicability to actual 3D data analysis. It assumes two sets of points: one set is sampled from the unit sphere; another set is sampled with distance at least 2 to this unit sphere. For the majority of 3D point cloud applications, point clouds are normalized to unit cube. My understanding is that in this case assumptions of Theorem 4.1 do not hold. Is my understanding correct? Maybe the setting that I have described can be adapted for Theorem 4.1 somehow?

The goal of theorem 4.1 is to prove that TOPF works in a very specific and constructed setting. This is to show that the theory can work in principle. You are, however, right that Theorem 4.1 does not give widely applicable guarantees in real-world 3D data analysis. Proving such guarantees appears to be very hard due to the interplay of possible complex topological structures, noise, the non-linear nature of the underlying Delaunay triangulation of the alpha-complex, and the continuous nature of the harmonic representations. (Of course, we can adapt the argument to scaled versions of the hyper-sphere.)

Please note that another reviewer made us aware that our definition of alpha-complexes had some technical differences to the definition of alpha complexes used in implementations and the literature. In order to avoid confusion and be open about this, we have renamed our previous definition to α\alpha^*-complexes, and introduced the more commonly used definition as α\alpha-complexes. We now state Theorem 4.1 for α\alpha^*-complexes and give conditions on when it holds for α\alpha-complexes as well.

On a related note, what is the normalization of the data for the experiments? Are point clouds normalized to unit cube? Is this normalization the same across all evaluated methods?

We did not normalise the point cloud, as we assumed some underlying meaning in the scaling of the data.

PointNet baseline is very weak. It was introduced in 2016 and was superseded by a large number of architectures: PointNet++, several versions of PointTransformers and other attention-based architectures. PointNet performance on the majority of benchmarks is significantly worse than for all of the mentioned methods. More up-to-date method can give significantly better performance on proposed benchmark.

We agree with the reviewer that PointNet is not the current SOTA on ShapenetPart Segmentation. However, PointNet is still in the range of 6% IoU of the current SOTA, in comparison to the 42% difference on TCBS with TOPF. This suggests that the performance of PointNet is still roughly representative of other similar Deep Learning approaches.

Perhaps more interesting than benchmarking against PointNet is benchmarking TOPF against other modern unsupervised (or weakly supervised) learning methods for point cloud feature extraction. We have done this with WSDesc[1] (pretrained on 3DMatch) on suggestion of another reviewer. WSDesc is competitive on current benchmark tasks. While WSDesc has a better performance than PointNet on the 3D datasets, TOPF still outperforms it by a wide margin on the more challenging sets of the TCBS. In addition to the 32-dimensional features produced by WSDesc, we have added a feature channel for the normalised 3d coordinates, because this greatly improved the performance of WSDesc on TCBS.

The obtained results are very interesting: On the easiest 3D dataset, 2Spheres2Circles (with a large (>4000) number of points and uniform sampling), WSDesc matches the performance of TOPF with an ARI of 0.95. On Spaceship, WSDesc achieves an ARI 0.76 (TOPF: 0.92). Looking at the results directly, we interpret this as WSDesc struggling with the non-uniform sampling across the different features. Finally, on SphereinCircle, WSDesc achieves an ARI of 0.46. This is better than most other non-toplogical methods, but significantly worse than TOPF with 0.97. In our interpretation, this has to do with the sparseness of just 267 points.

On the 2D datasets (which we embedded into 3D) WSDesc performs very poorly, which might be expected because it was trained for 3D datasets. This shows that on the TCBS, TOPF is competitive even against weakly supervised methods which excel on current benchmarks.

[1] Li, Lei, Hongbo Fu, and Maks Ovsjanikov. "WSDesc: Weakly supervised 3D local descriptor learning for point cloud registration." IEEE Transactions on Visualization and Computer Graphics 29.7 (2022): 3368-3379.

评论

Applicability of the method to high-dimensional spaces seems to be very limited. How well does the proposed method scale with the number of dimensions? My understanding is that local Delaunay triangulation will be significantly slower for higher dimensions. It is also not clear, how to choose the triangulation radius.

Thank you for your question! There is a runtime increase when going from ambient dimension 3 to dimension 4, as Delaunay triangulations and thus computing the alpha complex becomes infeasible and TOPF uses VR filtrations instead. However, VR filtrations only depend on distances between points in the point cloud and thus the runtime is independent of the ambient dimension. (Except for the computation of the relevant distances, which can be neglected for reasonable numbers of dimensions.) TOPF automatically chooses the alpha-radius based on persistent homology computations, see above.

Based on your feedback, we have now introduced an experiment where we use TOPF on 6500 points sampled from the 24-dimensional conformation space of cyclooctane, see Figure 11. Furthermore, we have repeated the VAE experiment on a latent space of dimension 16, which essentially replicated our findings of the 3-dimensional case, see Figure 13. This shows that TOPF works on higher-dimensional point clouds as well. Thank you again for this suggestion!

On a related note, the experiment in Figure 11 has very limited applicability. To analyze image features with TOPF one needs to train additional variational autoencoder to map image features in 3D space. This scenario does not seem to be practically appealing. What happens if relevant information cannot be contained in 3 dimensions? What if one wants to run TDA on original features?

As discussed above, TOPF can be used on high-dimensional data and thus on higher-dimensional latent spaces as well. We have conducted the experiment now for a latent dimension of 16 and added this to the paper (Figure 13), where we basically repeated the results of the latent space dimension of 3.

We however agree with the reviewer that the presented experiment and finding is not immediately applicable in a practical setting. However, we believe that the experiment shows that TOPF can be used on the latent space of VAE to recover hidden topological point-level information, which we found cool.

What dictated the sampling density of the TCBS dataset? Point clouds in Figure 7 have a number of samples that can differ by order of magnitude: 267 points for SphereInCircle and 4600 points for 2Spheres2Circles. Does method assume/TCBS data assume some upper bound on minimum distance to nearest neighbor for each point and that is what dictates sampling density? Or is it something else?

We have tried to incorporate a wide range of sampling densities both across as well as within the point clouds of the TCBS benchmark. Thus we have opted to include both data sets with many points, as well as with very few points. In order to enable a fair comparison with algorithm Topological Point Cloud Clustering [1] from the literature and because topological algorithm simlpy do not need this amount of sampling density, we did not use point clouds with hundreds of thousands of points. (TPCC relies on costly eigenvector computations, which TOPF avoids.) We will add more explanation about the construction and challenges of the TCBS to Appendix C to make the points above more clear! Thank you for raising this!

[1] Grande and Schaub, Topological Point Cloud Clustering, ICML 2023

Figure 5 is missing information about the average distance from each point to its nearest neighbor. Without this information, it is hard to assess how large is the relative value of the noise compared to manifold sampling density.

Thank you for this remark, this is indeed relevant information which we added to the paper. The experiment has been conducted on the dataset introduced in [1], which has been included in the TCBS with the authors' agreement as SphereinCircle, (Figure 7f). We have added an explanation on this. The radius of the centre sphere is 1.0, whereas the mean distance to the closest neighbour is 0.22, which is now included in the caption of the figure.

Figure 10 is missing time units, so it is impossible to infer speed of the method from it.

Thank you for noticing this! The time units are seconds; we have updated the corresponding figure.

评论

Also, how indicative are those running times (they are from two point clouds from the benchmark)?

As mentioned above, the point clouds all possess a fixed number of points. However, we have sampled a different number of points from the same distributions we used to generate two point clouds of TCBS. An overview of TOPF runtimes on data sets with a different number of points can be found in table 1. It seems that the presented runtimes are roughly indicative of the behaviour of TOPF. A main part of the runtime is just due to a complete persistent homology with generators calculation (and construction of the alpha-complexes in dimensions smaller than 4.), and projections to the harmonic generators. The runtimes of computing persistent homology have been studied in the literature, whereas we now added Table 2, where we give an overview of the runtimes of the individual steps in TOPF. The runtime for higher-dimensional point clouds is larger due to the use of the VR filtration: it took 189.8 seconds (including landmark sampling) to run TOPF on the 24-dimensional point cloud in Figure 11 on the system described in the appendix.

Figure 13 seems to be missing 9 of 12 subfigures.

We have checked and re-downloaded the pdf from openreview. For us, all 12 subfigures of figure 13 show. Could you maybe check whether this was an error with your download? If this problems persists with certain pdf-readers, we will replace the pdf-figure with a png-version, which should hopefully solve these problems. Thank you for raising this!

Related work discussion (including supplement) is limited; I recommend including some of the work discussed in [Rev 1]

We thank the reviewer for this suggestion! We will expand the related work section in the appendix using Rev1 in the revision phase. Due to space constraints, we have only added a reference to Rev1 and the references therein in the main paper, but link the supplementary material. If you feel that an important paper is missing, we would be happy to take the suggestion!

Main body of the paper focuses too much on introduction into algebraic topology and too little on practical considerations of the method;

We hope to have eased at some of your concerns about practical considerations of the method in the above comments, which we have incorporated into the main text and the appendix. In general, we would love have more space in the main part for talking about our method and detailed technicalities. However, we are afraid that in order for this paper to be understable and relevant for a Machine Learning audience, the introduction and motivation are important. This is especially true, because we do not simply try to apply persistent homology vectorisations to a new domain, as done in many other ML TDA papers, but try to combine methods and ideas from Algebraic Topology and Differential Geometry (see Appendix B.2) in a novel way. We have tried to write the preliminaries in such a way that it simultaneously explain concepts from TDA and their importance to TOPF. However, if you feel that a certain part of section 2 is not needed for understanding and motivating TOPF, we would be happy to engage in a discussion and adapt the section.

Thank you again for your useful feedback and suggestions! We believe that, regardless of the outcome, the paper has already significantly improved thanks to the reviews. We are looking forward to the discussion and would be happy about any additional comments or questions.

评论

I appreciate comments and additional work done by the authors. Due to lack of time, I will reply only to most important comments.

In the selection and construction of the point clouds of the TCBS, we have already tried to incorporate very different sample sizes, from very sparse point clouds with 158 points (Ellipses) to 4600 points (2Spheres2Circles). <..>

I don't think this addresses my initial concern. My understanding is that the goal of TCBS dataset is to provide a set of diverse and challenging manifolds to benchmark TDA approaches. Those manifolds can be sampled depending on experimental setting and constraints of the method. I think that having a fixed point sample from each test manifold greatly diminishes potential usefulness of TCBS as TDA benchmark dataset.

Based on your comment, we have now introduced an experiment (See Figure 15) where we decrease the sampling density for our point cloud with the highest original sampling density (2Sphers2Circles).

I appreciate this experiment. It clearly shows (at least on this point cloud) that there are two practical settings of proposed method: ~100 points (okay performance) and ~1000 points (good performance). This type of empirical evidence serves as a good argument in favor of the method. However, it should be compared with performance of other baselines and performance of the method on other point clouds.

Already in the TCBS we have included datasets with samling irrgularities. In Ellipses, the small 2 ellipses are sampled much more densely than the outer ellipses, <...>

I don't think this addresses my concern. In the examples of irregularities you have provided, sampling is still regular inside mentioned submanifolds. What happens sampling irregularity happens inside of the manifold? My suspicion is that it will throw triangulation off and method's performance might degrade.

In conclusion, TOPF makes use of persistent homology to choose a provably valid alpha-radius for every feature.

This mostly addresses my concern. One leftover issue that I have with this is that this heuristic probably heavily implies somewhat uniform sampling of the manifold (but this can be tested empirically).

The goal of theorem 4.1 is to prove that TOPF works in a very specific and constructed setting. This is to show that the theory can work in principle.

This is my biggest problem with theorem 4.1 -- it does not work in almost any practical scenario as pointed out by me and other reviewers. The fact the theory can work in principle has been already shown by research in algebraic topology and TDA that authors cite. To me, this theorem does not seem to add any additional insight and takes space from meaningful discussion of the method (implementation details, additional experiments, etc).

We did not normalise the point cloud, as we assumed some underlying meaning in the scaling of the data.

This is very strong concern for me. Incorrect normalization can significantly degrade performance of learning based method that was trained with specific normalization (e.g. [-0.5,0.5]^3 cube). Can authors provide details on TCBS point clouds normalization: max/min along each axis across all test point clouds?

However, PointNet is still in the range of 6% IoU of the current SOTA, in comparison to the 42% difference on TCBS with TOPF. This suggests that the performance of PointNet is still roughly representative of other similar Deep Learning approaches.

This assumes that performance in segmentation is indicative of how well point features capture topological properties of the data. I think this is a bit of a stretch. A lot of subsequent methods showed to have good topology awareness (see DGCNN, Fig. 4, https://arxiv.org/pdf/1801.07829) and they might be much stronger baseline than PointNet.

While WSDesc has a better performance than PointNet on the 3D datasets, TOPF still outperforms it by a wide margin on the more challenging sets of the TCBS.

I think your results should be interpreted with extreme caution. WSDesc was trained with very sparse point clouds: 512 points for 3DMatch dataset and 128 points for ModelNet40. If other sampling was used for comparison to this baseline (and I assume it was), performance of the method might not be indicative of performance in 'native' sample size. What was the sampling size that was used for WSDesc baseline comparison? Concerns about normalization of the data apply here as well.

评论

Furthermore, we have repeated the VAE experiment on a latent space of dimension 16, which essentially replicated our findings of the 3-dimensional case, see Figure 13. <..> TOPF can be used on high-dimensional data and thus on higher-dimensional latent spaces as well.

From practical point of view, additional usage of VAE seems to very suboptimal. Any compression method (e.g. linear projection) can change topological properties of the data we aim to study (like projection of 1-sphere can make it a closed interval). Ideally, we want to apply proposed TDA method to unmodified features to preserve all topological information. I also want to point out that 16 is not high dimension space from practical point of view. Modern feature extraction methods trained on large scale data an order of magnitude larger dimensionality: 768 for PointBERT, 768+ for CLIP variants, 1024+ for DINO.

In order to enable a fair comparison with algorithm Topological Point Cloud Clustering [1] from the literature and because topological algorithm simlpy do not need this amount of sampling density, we did not use point clouds with hundreds of thousands of points.

My concern are not computation concerns related to dense point clouds. My concern is the performance of TDA methods on sparse point clouds (much more common case). As I mentioned above, TCBS should probably provide manifolds (e.g. meshes) instead of point clouds that can be sampled to desired density depending on experimental setting.

评论

Thank you very much for your great and relevant comments! We will respond to your points below:

[..] I think that having a fixed point sample from each test manifold greatly diminishes potential usefulness of TCBS as TDA benchmark dataset.

This sounds like a convincing point to us. We will update the TCBS to not only include the reference point clouds already contained in a paper, but rather a the method to generate these point clouds with desired sampling densities/properties. Thank you very much for this suggestion!

In the examples of irregularities you have provided, sampling is still regular inside mentioned submanifolds. What happens sampling irregularity happens inside of the manifold? My suspicion is that it will throw triangulation off and method's performance might degrade.

Thank you for clarifying this! This is a valid concern. We have now performed experiments on four of the point clouds with inhomogenous downsampling. We have thus reduced the sampling density of one half of the point cloud, while keeping the sampling density of the other half. In particular, we have done so in a manner that the sampling density changes within the topological features/individual manifolds. (See figure 15 of the revised paper for an exemplary point cloud.)

The results show that TOPF performs well even in the presence of inhomogeneous sampling, and the degrading of performance can largely be attributed to a decrease in minimum sampling rate, rather than the inhomogeneous nature.

From a theoretical point of view, this should not be surprising: The birth of the topological feature will now depend on the minimum sampling density, while the death time still depends on the size of the feature. The heuristic employed by TOPF will then choose a maximal triangulation radius which lies between these two radii. In the dense part, the triangulation will be denser then in the sparse parts. However, due to the the character of the alpha filtration, the weighting of the simplices discussed in appendix G, and the thresholding this does not cause an issue in the features generated by TOPF.

Thank you for suggesting this kind of experiments!

This mostly addresses my concern. One leftover issue that I have with this is that this heuristic probably heavily implies somewhat uniform sampling of the manifold (but this can be tested empirically).

We hope to have addressed this with our above experiments.

The fact the theory can work in principle has been already shown by research in algebraic topology and TDA that authors cite. To me, this theorem does not seem to add any additional insight and takes space from meaningful discussion of the method (implementation details, additional experiments, etc).

In principle, we agree with the reviewer. In particular, we are happy that the reviewer sees that it is not surprising that TOPF works (in principle). However, to us it seems this point of view already requires some maturity in the background material, and especially people in ML put rather high value on these sorts of guarantees. Thus, we would ask the other reviewers how they stand on the matter and revise our paper accordingly. We could definitely use the space in the main text for details now in the appendix. Thank you for the suggestion!

We did not normalise the point cloud, as we assumed some underlying meaning in the scaling of the data. This is very strong concern for me. Incorrect normalization can significantly degrade performance of learning based method that was trained with specific normalization (e.g. [-0.5,0.5]^3 cube). Can authors provide details on TCBS point clouds normalization: max/min along each axis across all test point clouds?

We have re-run WSDesc on a normalised version of the point clouds: This did not seem to affect performance. Below we provide a summary of the normalisation:

filenamemin_xmax_xmin_ymax_ymin_zmax_z
0spheresAndGridTRUE.csv-0.960.21-0.670.600.000.00
14spheresTRUE.csv-0.970.09-0.930.310.000.00
2HalvedCircleTRUE.csv-0.580.61-0.85-0.200.000.00
3EllipsesInEllipsesTRUE.csv-3.691.82-4.00-2.000.000.00
4Two_Spheres_2_CirclesTRUE.csv-3.005.00-1.001.00-1.001.00
5spaceship_v2TRUE.csv-4.974.95-7.8323.81-3.934.00
6SphereinCircleTRUE.csv-3.003.00-3.003.00-1.001.00
评论

I appreciate this experiment. It clearly shows (at least on this point cloud) that there are two practical settings of proposed method: ~100 points (okay performance) and ~1000 points (good performance).

We have ran more experiments and updated Figure 15 so that it now incorporates the performance of TOPF on all datasets while downsampling. We interpret the results as basically repeating the good performance of TOPF under downsampling, except for the datasets EllipsesinEllipses and spaceship, where the performance drops rapidly (at least in a logarithmic plot). Our interpretation of this is that both datasets consist of submanifolds with different sampling density, which intersect/are contained in one another. Thus, the point density is a key to distinguishing the classes, and a random change in these densities due to downsampling can make the encoded structure hard to detect or even causes them to vanish.

However, PointNet is still in the range of 6% IoU of the current SOTA, in comparison to the 42% difference on TCBS with TOPF. This suggests that the performance of PointNet is still roughly representative of other similar Deep Learning approaches.

This assumes that performance in segmentation is indicative of how well point features capture topological properties of the data. I think this is a bit of a stretch. A lot of subsequent methods showed to have good topology awareness (see DGCNN, Fig. 4, https://arxiv.org/pdf/1801.07829) and they might be much stronger baseline than PointNet.

This is a very good and convincing perspective. We will drop the cited claim from our paper and instead include the above discussion.

I think your results should be interpreted with extreme caution. WSDesc was trained with very sparse point clouds: 512 points for 3DMatch dataset and 128 points for ModelNet40. If other sampling was used for comparison to this baseline (and I assume it was), performance of the method might not be indicative of performance in 'native' sample size. What was the sampling size that was used for WSDesc baseline comparison? Concerns about normalization of the data apply here as well.

Again, thank you very much for this comment! It is true that WSDesc was trained with 512 key points per point cloud. However, the authors have evaluated the method on the test set, where:

each point cloud has 5K randomly sampled keypoints for sparse descriptor extraction and matching.

This suggests (at least to us) that WSDesc was built for a flexible number of keypoints and it is not unreasonable to use a different number. In the above experiments with WSDesc, we gave the model the same number of points as the other methods, which differed from point cloud to point cloud.

However, based on your command we have now

  1. Tested the performance of WSDesc pretrained with modelnet on the TCBS. The performance was significantly worse than of the version pretrained on 3DMatch.
  2. Normalised the input data. This did not seem to significantly change the performance.
  3. Experimented with a different number of sampled keypoints (using farthest point sampling). The results of WSDesc are listed below:
ARI
('4spheresTRUE.csv', 128)0.0134642
('4spheresTRUE.csv', 256)0.0756748
('4spheresTRUE.csv', 512)0.107041
('EllipsesInEllipsesTRUE.csv', 128)0.664618
('HalvedCircleTRUE.csv', 128)0.0207727
('SphereinCircleTRUE.csv', 128)0.441502
('SphereinCircleTRUE.csv', 256)0.437732
('Two_Spheres_2_CirclesTRUE.csv', 128)0.946663
('Two_Spheres_2_CirclesTRUE.csv', 256)0.963316
('Two_Spheres_2_CirclesTRUE.csv', 512)0.963498
('Two_Spheres_2_CirclesTRUE.csv', 1024)0.965893
('spaceship_v2TRUE.csv', 128)0.921852
('spaceship_v2TRUE.csv', 256)0.893698
('spaceship_v2TRUE.csv', 512)0.885185
('spheresAndGridTRUE.csv', 128)0.466177
('spheresAndGridTRUE.csv', 256)0.200311
('spheresAndGridTRUE.csv', 512)0.0936102

This shows that farthest landmark sampling prior to input improves the performance of WSDesc, even when sampling to fewer keypoints than WSDesc was trained for. As of right now, we are not sure why this is the case. One reason might be that we evaluated the ARI only on the keypoints obtained through farthest-point sampling, potentially making the problem easier. However, TOPF still outperforms WSDesc on most of the datasets, and requires no expensive pretraining nor a GPU. We will include some form of these results in a revised version of the paper.

评论

From practical point of view, additional usage of VAE seems to very suboptimal. Any compression method (e.g. linear projection) can change topological properties of the data we aim to study (like projection of 1-sphere can make it a closed interval). Ideally, we want to apply proposed TDA method to unmodified features to preserve all topological information. I also want to point out that 16 is not high dimension space from practical point of view. Modern feature extraction methods trained on large scale data an order of magnitude larger dimensionality: 768 for PointBERT, 768+ for CLIP variants, 1024+ for DINO.

Thank you for your comment! You are right in saying that if we are only interested in the topology of the input data, we should not use a VAE. We included the experiment only to show that if we are in the case that we already computed a latent space embedding, we can then use TOPF to better understand the embedding.

However, based on your comment we have now included an experiment where we run TOPF on the raw data, i.e. the 8748-dimensional image space. The result is basically the same as the one obtained for the lower-dimensional cases, although the separation between the desired features and the noise becomes larger. (See Figure 13.) This shows that the VR-version of TOPF can work even for very high-dimensional point clouds. Thank you again for this suggestion!

My concern are not computation concerns related to dense point clouds. My concern is the performance of TDA methods on sparse point clouds (much more common case). As I mentioned above, TCBS should probably provide manifolds (e.g. meshes) instead of point clouds that can be sampled to desired density depending on experimental setting.

We now understand your concerns better. We want to add that some datasets of TCBS contain things that cannot be approximated by meshes, at least when interpreted in the sense of surfaces. For example in the spaceship dataset, the elipsoid are "thick" and only have a small hole in the middle, while being intersected with smaller ellipsoids with larger sampling density, and beingn below a 1-dimensional object. However as written above, we will provide improved sampling capabilities in an updated version of the TCBS software.

Thank you again very much for your replies! We hope to have addressed or at least acknowledged your comments, and we would be happy for any follow-up questions or replies.

评论

However, based on your comment we have now included an experiment where we run TOPF on the raw data, i.e. the 8748-dimensional image space. This shows that the VR-version of TOPF can work even for very high-dimensional point clouds.

This is strong result and I recommend including it in revised or resubmitted paper.

We want to add that some datasets of TCBS contain things that cannot be approximated by meshes,

This does not seem to be a big problem since all of your manifolds are defined analytically and thus can be sampled with required degree of precision.

评论

I strongly appreciate replies by the authors and all additional work and experiments they have done. However, I think that paper still cannot be accepted in the current state because it need substantial rework. If this would be a journal paper I would recommend 'resubmit a major revision' but for conference I have to recommend reject.

Here are final justifications/suggestions:

  1. Main text of the paper needs to be adjusted to better balance theory and practical considerations of the method. I recommend checking the following paper as examples on how to balance heavy mathematical background and practical aspects

Original diffusion models paper: https://arxiv.org/abs/1503.03585 Product Manifold learning: https://arxiv.org/pdf/2010.09908 SGLD: https://www.stats.ox.ac.uk/~teh/research/compstats/WelTeh2011a.pdf

As bare minimum, I recommend removing Theorem 4.1 completely and moving Figures 13 and 15 to main paper.

  1. Experimental evaluation should use 'native' normalizations for all baselines. TOPF also ideally evaluated in 'standard' normalization space of baselines, otherwise there is a concern that hyperparameters of TOPF might require adjustment to 'standard' normalization.

  2. I strongly recommend including results on sampling regularity and sampling density (for baselines as well).

  3. I strongly recommend including stronger baselines like DGCNN

  4. TCBS benchmark data might benefit from additional examples.

  5. I have also checked authors' discussion with reviewer smnw and I agree with the following reviewer's points: alpha*-complexes might be confusing both for readers who are familiar and who are unfamiliar with the theory; proposed work seem to be proposing isometric invariant and thus local shape descriptors probably should be mentioned in related work. That being said, I still think that main focus of the paper should be empirical validation of the paper because theoretical contribution seems to be very limited.

评论

We interpret the results as basically repeating the good performance of TOPF under downsampling, except for the datasets EllipsesinEllipses and spaceship, where the performance drops rapidly (at least in a logarithmic plot).

Those are useful insights that should be centerpiece of this paper, in my opinion. Those results also suggest that maybe TCBS benchmark should be expanded to include additional cases. Again, I appreciate the effort that went into this.

Experimented with a different number of sampled keypoints (using farthest point sampling). The results of WSDesc are listed below

Again, I really appreciate additional effort that went into these results. I recommend including table that compares all baselines and TOPF with respect to sampling density in revised version of the paper.

评论

I appreciate all additional work that authors did.

The results show that TOPF performs well even in the presence of inhomogeneous sampling,

This is interesting result and strong empirical argument in favor of TOPF. I recommend including it in revised version of the paper (if this iteration is not accepted). Again, I really appreciate additional effort that went into it.

Below we provide a summary of the normalisation:

This confirms my concerns: normalization of your data is very different from normalization that majority of learning based methods uses (PointNet, DGCNN, WSDesc): usually they use point clouds centered at 0 in [-1,1]^3 or [-0.5,0.5]^3 cubes. It makes it really hard to trust experimental comparisons with learnable methods like PointNet.

We have re-run WSDesc on a normalised version of the point clouds: This did not seem to affect performance.

I find it hard to believe. WSDesc uses local point patches (see Fig. 3, https://arxiv.org/pdf/2108.02740) for voxelization which are dependent on fixed voxelization radius r^LRF (0.3 in their experiments). Improper normalization should affect size of patches and therefore resulting voxel resolution and therefore results. I would recommend double checking those experiments.

Also, PointNet experiments almost certainly require adjusted normalization too.

评论

Dear reviewer,

thank you for your reply and your remarks. We apologise for the late reply!

We have already incorporated a large number of your points into the revised pdf which is uploaded to openreview. In particular, we have done the following:

I recommend removing Theorem 4.1 completely and moving Figures 13 and 15 to main paper.

We have done this and briefly discuss these experiments in the main paper. Thank you for your suggestion!

Experimental evaluation should use 'native' normalizations for all baselines.

We have re-run WSDesc with and without normalisation to the unit cube on all of the datasets of TCBS with 50 tries and looked at the results in detail. As reported, this does not change the mean ARI achieved by much, and in particular it worsens it on most of the datasets.

This shows that even WSDesc can utilise the natural scaling in our input point clouds. We will post exact numbers on this before the end of the discussion phase, our server is not in a good mood at the moment and we are currently trying to talk to it, we apologise.

UPDATE: We apologise, but our server is still experiencing issues and we cannot post the results here. We will, however, include them in the final version of the paper.

I strongly recommend including results on sampling regularity and sampling density (for baselines as well).

Thank you for the feedback, we now report the desired results in the uploaded pdf in Fgure 15 and 16.

I have also checked authors' discussion with reviewer smnw and I agree with the following reviewer's points:

Thank you very much for reading the discussion with other reviewers! We value your time and a second opinion highly!

alpha*-complexes might be confusing both for readers who are familiar and who are unfamiliar with the theory;

Thank you for this point! We have moved Theorem 1 to the appendix, and now only talk about α\alpha^*-complexes in the theory section of the appendix, where we have added a detailed explanation on the role both of these concepts play. We believe that this reduces the confusion, in particular because we have now fixed the previously misleading definitions, as remarked by reviewer smnw.

proposed work seem to be proposing isometric invariant and thus local shape descriptors probably should be mentioned in related work.

This is a point where we agree with the review, we have already added a discussion of the 2011 Memoli paper to the related work section in the appendix. Thank you for pointing this out again: we are open to more suggestions.

We will include a detailed pseudocode version and an even further expanded related work section in the camera-ready or resubmitted version of the paper.

We want to sincerely thank you for all the time and effort you invested in reviewing this paper and for all the constructive feedback we have received! Despite all the negative publicity regarding the state of the peer review process at large ML conferences, this was great!

We believe we have significantly improved our paper since your first review and were able to address the majority of your concerns. We would thus be interested to hear how your assessment of the paper has changed.

Thank you again! The authors

审稿意见
5

Many machine learning applications like classification needs point-level information and features. This paper proposes a method to extract node-level topological features from 2D and 3D topologically structured point clouds.

优点

The authors proposed topological point cloud clustering benchmarks that could benefit future works in TDA. The proposed method outperforms other methods on topological structured point cloud data and enjoys the `Robustness to noise' characteristic of persistence homology.

缺点

  1. Datasets:

    a) The Topological Clustering Benchmark Suite => TPCC is introduced by the authors in this paper to evaluate the effectiveness of the proposed clustering method. The construction of these datasets in this benchmark has not been sufficiently discussed in the paper/appendix. For instance, it is unclear how the ground truth labels have been obtained since all the datasets are synthetic in nature.

    b) 3D datasets => What are the ground truth labels for these datasets, and how have these labels been obtained? Why were auxiliary points added in Cys123?

  2. Experiments:
    a) There are a total of 12 datasets presented in the paper: seven 2D datasets in the TPCC benchmark and five 3D datasets, as shown in Figure 5. Then why are results on only four 2D and three 3D datasets presented in Table 1?

    b) Please provide a quantitative comparison with the baselines on the 3D datasets.

    c) Why are some of the run-times in Table 1 0.0 seconds? What is the breakdown of TOPF run-times in terms of Steps 1-4 of Algorithm 1?

  3. Computational Complexity: What is the complexity of Algorithm 1, especially Steps 3 and 4?

  4. Presentation: The algorithm presentation needs significant improvement. Steps 3 and 4 of the TOPF algorithm are hard to read. Please explain these steps using a concrete, small-scale toy point cloud dataset. The terms “simplex-valued harmonic representatives” (line 337) and “simplex-valued vector” (line 341) are used without any explanation. What is the role of interpolation parameter γ\gamma?

  5. Novelty: Which component of the algorithm is novel? Is it step 4, since Steps 1-3 are standard tools of TDA?

  6. Limited Applicability: TOPF only obtains meaningful results on point cloud with topological structure. However, many datasets do not have such a structure. Furthermore, TOPF is computationally expensive on datasets with a large number of points (e.g., those obtained from CAD designs or 3D Scanners). While the authors mentioned using landmarks to improve performance, no experiments in the paper did so. It seems the method is only practical in a very limited setting, which is points with only topological shapes and a small number of points.

问题

  1. What is the maximum homology dimension dd in the experiments?
  2. What is the relation between the output of step 3 (e^ki\hat {e}_k^i) of Algorithm 1 with the Circular coordinates of De Silva & Vejdemo-Johansson (2009)?
评论

We thank the reviewer for their review and their detailed feedback! We have uploaded a revised version of the paper based on the feedback, and will reply to your points below

a) The Topological Clustering Benchmark Suite => TPCC is introduced by the authors in this paper to evaluate the effectiveness of the proposed clustering method. The construction of these datasets in this benchmark has not been sufficiently discussed in the paper/appendix. For instance, it is unclear how the ground truth labels have been obtained since all the datasets are synthetic in nature.

We constructed the benchmark by sampling from topologically different shapes with varying sampling density in different ambient dimensions. For example for the point cloud 2Spheres2Circles, we combined points sampled from two spheres and two circles. We then divided the point cloud into four ground truth clusters, one for each of the two spheres and two circles and assigned every point the cluster corresponding to the object it was sampled from. Thus, the ground truth label of a point corresponds to the topological structure/object it was sampled from.

We have now added an explanation for this to appendix C. Thank you very much for your comment!

b) 3D datasets => What are the ground truth labels for these datasets, and how have these labels been obtained?

We only have ground truth labels for the 3D datasets of the TCBS, which we obtained as described above. In a reply below, we will explain the different datasets of the paper.

a) There are a total of 12 datasets presented in the paper: seven 2D datasets in the TPCC benchmark and five 3D datasets, as shown in Figure 5. Then why are results on only four 2D and three 3D datasets presented in Table 1?

There seems to be some confusion over the data sets used in the paper. We are grateful for the reviewer to raise these issues. We have subsequently made the passages in the paper more clear and will try to explain our data sets in the rebuttal: There are in total 11 (now 12 in the revised version) data sets used in the paper. 7 of these are synthetic and are contained in the topological clustering benchmark suite (TCBS) and thus have ground truth clustering labels, obtained as described above. Out of these 7, there are four 2d and three 3d data sets, as described in the caption of Table 1, in section C in the appendix and in Figures 7&8. This is how we obtained the ground truth labels for the 3D data sets for the experiments reported in Table 1.

The remaining point clouds in figure 4 (a), b), c) and e)) are protein coordinates, or the state space of a simulation of a Lorentz attractor, or the latent space of a VAE. Furthermore, we have now added a high-dimensional space of points sampled from the conformation space of cyclooctane (Fig. 11). We have not used ground truth data for these four point clouds, as it does not seem to be clear what would constitute a ground truth. We could have of course came up with a notion of ground truth aligning with the results of TOPF, but we felt this to be disingenuous. Thus we have not included benchmark results on these point clouds in Table 1. We note however, that qualitative evaluation in Figure seems to indicate that TOPF is able to recover meaningful topological features.

We have now explained the construction in more detail in the appendix and thank the reviewer for their feedback!

Why were auxiliary points added in Cys123?

In the experiment with Cys123, we wanted to show that TOPF can detect so-called protein pockets. Because pockets do not form holes detectable by TDA straightforward, the literature suggests adding (and later removing) these points on the convex hull, which turns the pockets into holes, see for example "Novel definition and quantitative analysis of branch structure with topological data analysis." Haruhisa Oda et al., 2024. This demonstrates just another possible use-case of TOPF. We have added a more detailed explanation of this to the paper to appendix E and thank the reviewer for their feedback.

Please provide a quantitative comparison with the baselines on the 3D datasets.

As we discussed above and mentioned in the caption of table 1, three of the datasets, namely 2Spheres2Circles, SphereinCircle and Spaceship, are 3D datasets. There does not seem to be a major performance difference of TOPF on 2D and 3D datasets.

Why are some of the run-times in Table 1 0.0 seconds?

This is due to rounding. Times less than 0.05 seconds were rounded to 0.0. We have their entries to <0.1 to avoid confusion.

What is the breakdown of TOPF run-times in terms of Steps 1-4 of Algorithm 1?

We have now added a detailed breakdown of the individual steps of TOPF on the point clouds of TCBS to Table 2. In summary, computing the persistent homology and the harmonic projections constitute the majority of the runtime. Thank you for the suggestion!

评论
  1. Presentation: The algorithm presentation needs significant improvement. Steps 3 and 4 of the TOPF algorithm are hard to read.

We are sorry to hear that steps 3 and 4 of the TOPF algorithm in section 3 are hard to read. We have tried very hard to make the entire paper as accessible and easily understandable as possible. Section 3 is without doubt the most technical section, where an understanding relies on a good grasp of concepts of algebraic topology and discrete Hodge Theory. We have thus tried to write the paper in such a form that the remaining sections are accessible with only a very rough understanding of the technicalities of section 3. Nonetheless, we believe they are important to include in the paper to enable exact reproduction of our algorithm. However, we have re-read the section and will try to improve readability by adding explanations and being more clear about the concepts used in the revision period.

Please explain these steps using a concrete, small-scale toy point cloud dataset.

As referenced in the introduction of section 3, we provide an overview with a toy point cloud in Figure 1. Please note that the steps from Figure correspond to the steps in Section 3. Furthermore, we have illustrated the algorithm in the three subfigures of Figure 3. We believe that these two figures are good illustrations, but are open to feedback by the reviewer.

The terms “simplex-valued harmonic representatives” (line 337) and “simplex-valued vector” (line 341) are used without any explanation.

We have rewritten the parts and added an explanation. Thank you for raising this issue!

What is the role of interpolation parameter ?

The mathematical role of the interpolation parameter γ\gamma is explained right after its introduction in Step 3. Given a topological feature with birth time bb and death time tt, γ\gamma controls the time step tt where we construct the associated simplicial complex with t=b1γdγt=b^{1-\gamma}d^\gamma. (For homology dimensions above 00. In dimension 00, the birt time of the feature is 00 and thus we use linear instead of exponential interpolation.) A more intuitive explanation of the interpolation parameter can be found in section D.1 'Hyperparameters' of the appendix. In a nutshell, larger interpolation parameter will result in smoother topological features, whereas a smaller interpolation will lead to 'sharp' features. We just noticed that the subsections D.1 'Hyperparameters' and D.2 'Runtime' were partially swapped, we have corrected this issue.

  1. Novelty: Which component of the algorithm is novel? Is it step 4, since Steps 1-3 are standard tools of TDA?

First we would like to remark that novelty is often not only a sum of the individual components, but much more the entire idea to combine previous techniques in a new way to obtain a novel result/goal, see for example Michael Black's Guide to novelty.

The entire idea to compute node-level features relating the global topology in terms of persistent homology back to the local points is very novel, the only somewhat similar papers being [1] and [2], although we highlight the differences and the advantages of our approach in the related work.

Step 1, computing persistent homology is indeed a standard tool of TDA. Step 2, picking significant generators, was done in [3], but we have introduced significant refinements to the heuristic, see appendix F "How to pick the most relevant topological features", and furthermore realised that for our applications, considering homology instead of cohomology is the right approach. Step 3, Projecting homology generators to the harmonic space was recently considered in [1] and [4], however, there the projections where performed at the birth of the persistence bar. In comparison, we have introduced the interpolation parameter, leading to smoother and more stable topological features. Furthermore, we have introduced a simplex weighting improving the harmonic representatives, and present an easy-to-solve least squares problem to compute them using the Hodge decomposition, which we are unaware of previous applications. (Edit: There are prior similar implementations, phrased differently) Step 4 is novel in this context in this form.

We hope to have answered the question of the reviewer!

[1] Gurnari, Davide, et al. "Probing omics data via harmonic persistent homology." arXiv preprint arXiv:2311.06357 (2023) [2] Grande, Vincent P., and Michael T. Schaub. "Topological point cloud clustering." Proceedings of the 40th International Conference on Machine Learning. 2023. [3] De Silva, Vin, and Mikael Vejdemo-Johansson. "Persistent cohomology and circular coordinates." Proceedings of the twenty-fifth annual symposium on Computational geometry. 2009. [4] Basu, Saugata, and Nathanael Cox. "Harmonic persistent homology." SIAM Journal on Applied Algebra and Geometry 8.1 (2024): 189-224.

评论
  1. Limited Applicability: TOPF only obtains meaningful results on point cloud with topological structure. However, many datasets do not have such a structure.

It is true that many point clouds do not have a topological structure. However, the research on Topological Data Analysis over the last decade reveals that there are a wide range of applications where point clouds possess topological structure, see for example the wide range of applications cited in our introduction from biology, physics and computer science. We thus believe that these topologically structured point clouds are important enough to warrant research. In particular, as our experiments show, TOPF clearly outperforms any other method in this niche of topologically structured data.

Furthermore, TOPF is computationally expensive on datasets with a large number of points (e.g., those obtained from CAD designs or 3D Scanners). While the authors mentioned using landmarks to improve performance, no experiments in the paper did so. It seems the method is only practical in a very limited setting, which is points with only topological shapes and a small number of points.

In Figure 10 in the appendix, we provide experiments comparing the runtime of TOPF on point clouds with and without sampling enabled with close to 50000 points. We believe that most point clouds can be sampled down to 50.000 points. For larger point clouds, the Ripserer.jl library used in our implementation sometimes crashes, which is the only thing prohibiting experiments without downsampling on larger point clouds. In Figure 15, we have added an experiment on how downsampling affects performance of TOPF.

While we agree that TOPF does not have the performance of some Computer Vision models running on multiple GPUs, we have shown that using the base implementation in conjunction with landmark and random sampling can make TOPF scale to rather large point clouds.

  1. What is the maximum homology dimension in the experiments?

For the experiments in the TCBS, we set the persistent homology dimension to the maximum possible dimension permitted by the ambient dimension, which was 1 and 2 respectively. For the experiments with real data, we use prior knowledge on the topological structure we were looking for to set the homology dimensions: For the NALCN channelosome and GroEL proteins, we wanted to detect the loops/1-dimensional holes in the protein structure, and looked at dimension 1. For the Cys123 (with the added convex hull), we looked for protein pockets and thus only looked at dimension 2. For the VAE latent space embedding experiment, we looked for loops informed by the sampling strategy, and thus set the dimension to 1. Finally, for the Lorentz attractor data set, we again looked for loops around the non-attractive fixed points and thus used a homology dimension of 1. We assume that in many use cases, we will already** know for what kind of topological features** we will be looking for. However, we can always run TOPF with a larger maximum homology dimension and TOPF auto-selects the relevant dimensions by comparing the life times of features across dimensions. Which is of course more prone for error than using prior knowledge.

What is the relation between the output of step 3 of Algorithm 1 with the Circular coordinates of De Silva & Vejdemo-Johansson (2009)?

We discuss differences in our step 3 with previous literature including the circular coordinates in an above comment. The authors of the cited paper integrate the harmonic cohomology representative to obtain, for every node, a circle-valued coordinate on a circle. This assigns the coordinate to the point independent of whether the point lies close to the circle the coordinate is representing. (This is because by Brown's representability theorem, every cohomology class is representable by a map of the entire space into an Eilenberg-MacLane Space, which is for dimension 1 a circle.) These coordinates then don't measure how much a point belongs to a topological features (as we do) but rather at which point of the circle (even if it does not lie on the circle) the point lies. This then only works for homology of dimension 1. This is a very interesting and somewhat orthogonal approach, but for example would not work for clustering on the TCBS. (We apologise for the mathematical explanation but hope that this helped to at least gain some intuition.)

We want to again thank you for your feedback, remarks, and questions. We believe that these already resulted in a better paper. We would be happy to answer any follow-up questions, clear out any remaining misunderstandings, or hear whether your assessment of the paper has changed. In particular, we are looking forward to the discussion!

评论

Thank you for your detailed response and for clarifying my questions. My concerns about the dataset, novelty, and experiments have been resolved, so I have adjusted my score.

My question about computational complexity remains unanswered, and my concerns about Presentation and limited applicability remain. I have the following questions and suggestions:

Referring to Table 2,

Julia complex is the time Ripserer needs to build an α\alpha-complex, …, Harmonic projection denotes the time needed to build the α-complexes by gudhi and to compute the weighted harmonic representatives

  • Why α\alpha-complex is built in both Ripserer and gudhi, twice?
  • What is “Julia output”?

We notice that for smaller and lower-dimensional point clouds, a major part of the runtime is due to steps performed in julia and ripserer calculations, over which we have no influence over.

  • Can these bottlenecks be overcome if you use Sparse-Rips filtration or Witness filtration? Can the “Julia complex” and “Julia Homology” computation times (Table 2) be further reduced by considering these methods? For instance, the runtime on the 6500 point-cloud in Figure 11 is more than 3 minutes, which shows that there is a need to improve the efficiency of TOPF in such cases.

While we agree that TOPF does not have the performance of some Computer Vision models running on multiple GPUs, we have shown that using the base implementation in conjunction with landmark and random sampling can make TOPF scale to rather large point clouds.

  • Are you referring to Figure 11? What do you mean by “mixture of landmark and random sampling” in the line: “To counteract the increase in computational complexity, TOPF automatically downsamples the point cloud using a mixture of landmark and random sampling.”?

In Figure 10 in the appendix, we provide experiments comparing the runtime of TOPF on point clouds with and without sampling enabled with close to 50000 points. We believe that most point clouds can be sampled down to 50.000 points. For larger point clouds, the Ripserer.jl library used in our implementation sometimes crashes, which is the only thing prohibiting experiments without downsampling on larger point clouds. In Figure 15, we have added an experiment on how downsampling affects performance of TOPF.

  • What does sparsification = ’off’ and ‘auto’ mean in Figure 10?
  • Such crash is typically due to the memory overhead of simplices (in the filtration) generated by large number of points. This highlights my emphasis on considering sparse filtrations such as sparse Rips.
  • “.. we have shown that using the base implementation in conjunction with landmark and random sampling can make TOPF scale to rather large point clouds.” => By larger point clouds, did you mean up to 50000 points?

Please explain these steps using a concrete, small-scale toy point cloud dataset.

As referenced in the introduction of section 3, we provide an overview with a toy point cloud in Figure 1... We believe that these two figures are good illustrations, but are open to feedback by the reviewer.

  • Can you write the steps of how each homology representatives are projected into harmonic space as an algorithm? The computational steps appear to be lost under the discussion.

The entire idea to compute node-level features relating the global topology in terms of persistent homology back to the local points is very novel, the only somewhat similar papers being [1] and [2], although we highlight the differences and the advantages of our approach in the related work.

  • Computing node-level topological features in an unsupervised manner relating the global topology in terms of persistent homology back to the local points is novel for point clouds however other works exist that did the same on graphs in supervised setting. See for example, the local topology encoding in section 3.1 in [1]. Please clarify this distinction with [1] in your related works where such works are discussed.

[1] When Witnesses Defend: A Witness Graph Topological Layer for Adversarial Graph Learning (https://openreview.net/forum?id=cnAeyjtMFM )

For the experiments with real data, we use prior knowledge on the topological structure we were looking for to set the homology dimensions: For the NALCN channelosome and GroEL proteins, we wanted to detect the loops/1-dimensional holes in the protein structure, and looked at dimension 1..…. We assume that in many use cases, we will already** know for what kind of topological features** we will be looking for...

  • Using knowledge about ambient dimension to set homology dimension is a strong assumption about the extent of prior knowledge. In higher dimensions, we may not know what kind of topological features we should be looking for.
评论

Thank you very much for your reply! We are happy that we were able to resolve your concerns regarding the dataset, novelty, and experiments! We will reply to your comments below:

My question about computational complexity remains unanswered

Thank you for your reminder! We apologise, we forget to address this point in our previous reply. We will discuss the complexity of the involved steps:

1a). Constructing the complex: The Vietoris--Rips complex will have O(nk+1)O(n^{k+1}) simplices, where nn is the number of points and kk the maximum homology dimension. Apart from this, there is no significant computational effort needed. Note that we can decrease this if we set a maximum VR-radius, as done in many works from the literature. Computing the alpha-complex in ambient dimension dd is equivalent to computing the convex hull in ambient dimension d+1d+1, which has a complexity (edit:) and number of simplices of at most O(n(d+1)/2)\sim O(n^{(d+1)/2}).

1b). Computing persistent homology and generators: As we need the homology generators, we use the recently introduced involuted persistent homology [1]. While the authors discuss the runtime performance, they do not give a concrete complexity bound in [1]. However, they deduce that it has a similar run-time to usual persistent cohomology computations, whose performance is for example discussed in [2].

  1. Picking the significant generators has negligible impact.

  2. Computing harmonic projections: Given a homology representative rr in dimension kk, computing the harmonic representative amounts to computing a sparse matrix vector product of BkxB_{k}x and a sparse least squares problem lsmr(B_{k},r), i.e. solving minxRSk+1rBkx2\min_{x\in \mathbb{R}^{S_{k+1}}} \lVert r-B_{k}x\rVert_2 where Sk+1S_{k+1} is the number of (k+1)(k+1)-simplices in the simplicial complex and BkB_{k} has (k+2)Sk+1(k+2)S_{k+1} non-zero entries. This is a sparse least-squares problem, which we solve using the iterative sparse solver lsmr [3]. Because this is an iterative solver, the authors do not give runtime complexity, but discuss the computational requirements in [3].

  3. Pooling and averaging over the simplicial neighbours has negligible runtime constraints.

We have added the above discussion to the section on runtime in the appendix (which we will upload after finishing these comments) and thank the reviewer for suggestion!

[1] Čufar, Matija, and Žiga Virk. "Fast computation of persistent homology representatives with involuted persistent homology." Foundations of Data Science 5.4 (2023): 466-479.

[2] de Silva, Vin, Dmitriy Morozov, and Mikael Vejdemo-Johansson. "Dualities in persistent (co) homology." Inverse Problems 27.12 (2011).

[3] Fong, David Chin-Lung, and Michael Saunders. "LSMR: An iterative algorithm for sparse least-squares problems." SIAM Journal on Scientific Computing 33.5 (2011): 2950-2971.

Why α\alpha-complex is built in both Ripserer and gudhi, twice?

This is just an implementational detail that will be optimised in a future release of TOPF. The implementation of above cited method [1] is in Julia, whereas we wanted to provide a python package while minimising the needed communication between julia and python.

What is “Julia output”?

This is the communication time between Julia and python. We did not want to rely on the sometime complicated setup of PyJulia. Again, this is an implementational detail that does not have an impact on the theoretical contributions of the paper.

While we agree that TOPF does not have the performance of some Computer Vision models running on multiple GPUs, we have shown that using the base implementation in conjunction with landmark and random sampling can make TOPF scale to rather large point clouds.

Are you referring to Figure 11? What do you mean by “mixture of landmark and random sampling” in the line: “To counteract the increase in computational complexity, TOPF automatically downsamples the point cloud using a mixture of landmark and random sampling.”?

We believe we are referring to Figure 10 "Runtime on two examples from Topological Clustering Benchmark Suite while increasing point density.", but the ordering could have changed.

By a mixture of landmark and random sampling we mean that in the first step, we sample down the point cloud randomly, and afterwards select points by iteratively adding the point with the largest distance to the already selected points.

评论

What does sparsification = ’off’ and ‘auto’ mean in Figure 10? This means that TOPF does not performance input sparsification to improve runtime.

Such crash is typically due to the memory overhead of simplices (in the filtration) generated by large number of points. This highlights my emphasis on considering sparse filtrations such as sparse Rips.

This is an excellent suggestion. From a theoretical point of view, TOPF is agnostic in the used underlying filtrations. For simplicity of the presentation, we have opted to consider the two most prominently used and best implemented filtration types, VR and α\alpha filtrations. However, TOPF can be easily generalised to work with arbitrary filtrations.

By larger point clouds, did you mean up to 50000 points?

Yes, you are right. For point clouds above this size this, it would probably be best to first sample the point cloud and then apply TOPF to some landmarks representing the underlying topology.

Can you write the steps of how each homology representatives are projected into harmonic space as an algorithm? The computational steps appear to be lost under the discussion.

Yes, we can do this! Are you talking about adding them to the caption of Figure 1, or to section 3? In section 3, we already already give explicit equations (~lines 337--340). Let eke_k be the homology representative, then the harmonic representative can be written down as

e^k=ekBkargminxekiBkx22\hat{e}_k = e_k-B_k\text{argmin}_x\lVert e_k^i-B_kx\rVert_2^2

We have made the point of the harmonic representative now clearer in step 3 of section 3, if this was what you meant. Thank you for raising the issue that this was not clear enough.

Computing node-level topological features in an unsupervised manner relating the global topology in terms of persistent homology back to the local points is novel for point clouds however other works exist that did the same on graphs in supervised setting. See for example, the local topology encoding in section 3.1 in [1]. Please clarify this distinction with [1] in your related works where such works are discussed.

Thank you for the relevant reference! We have added a discussion to the related work, which we will repeat it here here:

The authors of [1] construct topological representations of graphs to study adversarial graph learning. They construct a local topological features for nodes and global topological features for the entire graph. In contrast to our approach, they do not relate the global topological features back to the local level, but rather consider the local topology on the point level and the single global topological summary separately. Furthermore when constructing their features, they use vision transformers on the persistence images of their witness filtrations, which does not allow for the same amount of interpretability, as TOPF does.

(This is not meant as a critique of their approach, but rather as an attempt to point out the differences to our work.)

Using knowledge about ambient dimension to set homology dimension is a strong assumption about the extent of prior knowledge. In higher dimensions, we may not know what kind of topological features we should be looking for.

This of course is true. We would like to point out that -- quite surprisingly -- relevant topological features in pratice often occur in dimension 1. (And even dimension 2 is a lot rarer then dimension 1.) So while this might be limiting factor in theory, for the most applications even 1-dimensional homology will suffice. (There is a wealth of applied TDA literature in dimension 1, dimension 2 is a lot rarer, and finally we are not aware of a significant application of 3-dimensional persistent homology.)

Thank you again for your comments! We hope to have addressed your concerns. We would be happy about any follow-up questions or comments!

评论

Dear Reviewers, thank you very much for reviews and suggestions! We have incorporated your feedback into a revised version of the paper and believe that your feedback has already significantly improved our paper. (We are still working on some things.)

We will now give a brief summary of the most important revisions and give more details in the individual rebuttals and in the revised paper:

  1. We have added the weakly supervised modern point cloud learning method WSDesc as a further baseline on TCBS, validating the claim that TOPF outperforms competitors on the TCBS. (mean ARI 0.86 vs 0.39) (Thanks to reviewer YUvn)
  2. We added an experiment on the 24-dimensional conformation space of cyclooctane, showing that TOPF works on higher-dimensional spaces. (Fig. 11) (Thanks to reviewer hKZz)
  3. We repeat the VAE experiment for a latent space of dimension 16, mirroring previous findings. (Fig. 13) (Thanks to reviewers hKZz and 3Zxm)
  4. We discovered that our definition of α\alpha-complexes slightly differed from the literature definition used in implementations. We now call the previous construction α\alpha^*-complexes, give both definitions, discuss the similarities and differences, formulate theorem 4.1 for α\alpha^*-filtrations and give conditions for when it holds for α\alpha-complexes as well. (Thanks to reviewer smnw)
  5. We analyse the performance of TOPF experimentally while decreasing sampling density, showing promising results. (ARI>0.9 even for 6-fold downsampling, Fig. 15) (Thanks to reviewer hKZz)
  6. We give a breakdown of the runtime of TOPF on the separate steps. (Table 2.) (Thanks to reviewer MZq5)
  7. We expanded our related work discussion and will expand it even further. (Thanks to reviewers hKZz and 8pLm)
  8. We conducted an ablation study for not using the Hodge Laplacian (I.e. skipping step 3 of the pipeline) in the reply to reviewer 3Zxm, which we will include in the main paper. The results indicate that not using the Hodge Laplacian significantly affects TOPF performance, validating the theory behind our approach. (Thanks to reviewer 3Zxm)
  9. We added a discussion of a weighted fixed-number-of-features version of TOPF which does not depend on a feature selection heuristic. (Appendix F) (Thanks to reviewer 8pLm)
  10. We improved the clarity of writing, explanations, and figures at various points in the paper based on the feedback received. (Thanks to all of you!)

Thank you again for your detailed reviews and valuable feedback! We look forward to the discussion phase and will be happy to answer any follow-up questions.

评论

Dear Reviewers,

Under normal cicumstances, we believe that the submission history of a paper in a review process is irrelevant: The reviewers are competent to form their own opinions, which they can then discuss with the authors and their fellow reviewers. However, in one of the reviews, reviewer smnw has publicly flagged our paper for an ethic's review, writing:

This paper is a minor revision of the submission to NeurIPS 2024, where one of the reviews provided a counter-example to Theorem 4.1. Submitting almost the same paper without correcting this claim raises serious ethical concerns.

From our point of view, this claim is misleading. We would thus like to provide additional context. If you are not interested in this, you are free to skip this comment, as it does not deal with the paper on a scientific level.

It is true that an earlier form of this paper was already submitted to NeurIPS. After a very productive rebuttal phase, all of the four reviewers recommended to accept the paper for NeurIPS.

The area chair participated in the discussion with a brief comment, asking whether a certain example was a counter example to our Theorem 4.1. We replied to this (5 days before the end of the discussion phase) with an explanation of the definitions used in our paper (see point 4 of our summary of revisions), and why these definitions do not lead to the discussed instance being a counter-example. We did not receive another reply or a message acknowledging our comment.

As reviewer smnw wrote, we then received a rejection notification, claiming, for instance:

The paper tried to develop a "single complex description of the global structure of the point cloud" (lines 3-4 quoted from the abstract) [..].

For reference, the full quote of the abstract is:

Tools like Persistent Homology or the Euler Transform give a single complex description of the global structure of the point cloud.

All of the above claims are verifiable, as we opted to make our openreview discussion public. (Which it is since November 6.) We ask you however not to search for this on openreview as this would violate the anonymity policy of ICLR.

We have not written the above comment in order to influence the reviews or because we believe that NeurIPS reviews have any relevance to the ICLR decision process. However, we felt that the above cited claim by reviewer smnw was misleading and wanted to offer you a more complete picture. In particular, we want to emphasise that we have gained a lot of valuable feedback from the ICLR reviews and want to once again thank you for this.

We are looking forward to the discussion phase!

评论

Dear reviewers,

Thank you again for the productive discussion and rebuttal phase! In addition to the changes summarised in the previous comment, we have incorporated feedback from the discussion phase to further strengthen our paper. In particular, we have:

  1. Added experiments on sampling density and sampling heterogeneity, showing that TOPF performs well even on sparse point clouds or point clouds with heterogeneous sampling on all/4 point clouds including a comparison with baselines. (See Figure 15 and Figure 16 and Section 4: Robustness Against Noise, Downsampling, and Sampling Heterogeneity, and Appendix F.

  2. Added experiments showing that TOPF works even on very high-dimensional latent spaces of VAE and on the original 8748-dimensional base space of the image patches, Figure 6 and Appendix F.

  3. Added a discussion on the theoretical runtime complexity, appendix E.2, and the runtime of individual TOPF steps, table 2.

  4. Changed the organisation of the latter sections of the paper based on suggestions from reviewer hkKz, including figures on high-dimensional point clouds and a discussion of the performance of TOPF under sampling heterogeneity and sparseness in the main text of the paper.

  5. Added a reference to local shape descriptors in the related work section in the appendix.

  6. Added explanation on α\alpha-complexes and α\alpha^*-complexes, hopefully making this distinction clearer.

Including references and appendices, the paper now contains 32 pages.

Thank you again for your reviews and the constructive discussion phase! We hope to have addressed your concerns and would be interested in how this affects your assessment of the paper.

The authors

AC 元评审

The authors propose algorithmic approach to extract node-level topological features from complex point cloud data points. The proposed topological point features TOPF is evaluated on both synthetic and real world datasets. The Reviewers raised concerns on the definition and theoretical result (e.g., alpha*-complexes definition, Theorem 4.1), it is better to give more care to polish the new theoretical findings and place it in the literature. Additionally, the Reviewers raised concerns on the baselines (including recent advanced methods for point clouds), ablation study on hyperparameters, normalization. Overall, we think a careful updated version is necessary given the Reviewers' comments. The authors may consider them to improve the submission.

审稿人讨论附加意见

The Reviewers raised several concerns on the experiments (e.g., baselines, hyperparameters, normalization) and new theoretical findings (e.g., Theorem 4.1) together with the definition on alpha*-complexes which is required more care to prevent confusion with the standard alpha-complexes in the literature. Therefore, we think a major revision is required.

最终决定

Reject