PaperHub
6.1
/10
Poster4 位审稿人
最低2最高4标准差0.8
4
4
3
2
ICML 2025

Point-Level Topological Representation Learning on Point Clouds

OpenReviewPDF
提交: 2025-01-23更新: 2025-07-24
TL;DR

We use topological data analysis to relate global topology to local point features on point clouds.

摘要

关键词
Topological Data AnalysisDifferential GeometryRepresentation LearningHodge LaplacianSimplicial ComplexPersistent HomologyPoint Clouds

评审与讨论

审稿意见
4

The paper proposes to extract point-level features given the global structure of the point cloud, using concepts from algebraic topology and differential geometry.

给作者的问题

No

论据与证据

The proposed method can compute point-level topological features conditioned on the global topological structures of the point cloud. The proposed method outperforms other methods in down streaming tasks. It also achieves provably meaningful representation, and is robust to noise.

The design of the module is grounded with strong theoretical foundation and explained clearly. The qualitative results and experiments verify the effectiveness of the proposed method.

方法与评估标准

The proposed method is novel and grounded with theoretical foundation. What I found interesting and important is, the proposed module does not require training.

理论论述

The theoretical claims make sense to me. But the detailed math is not carefully checked.

实验设计与分析

The experiments and the visualization are good. I wonder whether it is possible to evaluate on more diverse tasks, like ModelNet40 classification, ShapeNet segmentation, and S3DIS segmentation, like many point cloud papers evaluate.

补充材料

Supplementary material is briefly skimmed. The codes in the supplementary material look very well organized and documented.

与现有文献的关系

This work seems to be very useful in point cloud segmentation tasks, because it can extract pointwise feature conditioned on the global information. I will be very interested in seeing the performance of the proposed method on a point cloud segmentation task.

遗漏的重要参考文献

No

其他优缺点

No

其他意见或建议

As mentioned in the previous sections, I found the proposed method interesting and strongly grounded, while showing an experiments on point cloud segmentation will make the paper much stronger.

作者回复

Thank you for your review and your valuable feedback!

The experiments and the visualization are good. I wonder whether it is possible to evaluate on more diverse tasks, like ModelNet40 classification, ShapeNet segmentation, and S3DIS segmentation, like many point cloud papers evaluate.

Thank you for your comment! The point clouds in datasets like ModelNet40 or ShapeNet have none to very few topological features, while the neural architectures learn to extract some different form of features. However, TOPF was specifically design to extract the topological features from homology. It might be interesting to extend the ideas from TOPF to geometrical information as well in future work, but this will require many new ideas and is out of scope for the current paper. The above is the reason why we have introduced the novel Topological Clustering Benchmark Suite to benchmark TOPF.

审稿意见
4

The paper introduces TOPF (Topological Point Features), a method for extracting point-level topological features from point clouds using tools from algebraic topology and differential geometry. The authors propose leveraging persistent homology and harmonic representatives from the Hodge Laplacian to relate global topological structures to local point features. Experiments on synthetic and real-world datasets demonstrate that TOPF outperforms existing methods in clustering tasks, exhibits robustness to noise and heterogeneous sampling, and scales to high-dimensional data. A new topological clustering benchmark suite is introduced to evaluate performance.

给作者的问题

See the comment above

论据与证据

Yes

方法与评估标准

Yes

理论论述

I'm not an expert in this field, which cost me a lot of time to understand the meaning of each theoretical claim. I didn't recognize any obvious proof of error.

实验设计与分析

The experimental design is sound.

补充材料

I have checked the appendix and the provided example code.

与现有文献的关系

The work builds on persistent homology [1] and harmonic representatives [1], extending TPCC (Grande & Schaub, 2023a) by reducing computational cost and improving robustness.

[1] Edelsbrunner H, Harer J. Persistent homology-a survey[J]. Contemporary mathematics, 2008, 453(26): 257-282. [2] De Silva V, Vejdemo-Johansson M. Persistent cohomology and circular coordinates[C]//Proceedings of the twenty-fifth annual symposium on Computational geometry. 2009: 227-236. [3] Grande V P, Schaub M T. Topological point cloud clustering[C]//Proceedings of the 40th International Conference on Machine Learning. 2023: 11683-11697.

遗漏的重要参考文献

No.

其他优缺点

Strength: This paper introduces a novel benchmark suite for topological clustering. The paper is well-written, with comprehensive theoretical analysis and visualizations to support the hypothesis.

Weakness: It could be better to analyze the runtime of the computational complexity of persistent homology regarding different point cloud sizes.

其他意见或建议

See the comment above

作者回复

Thank you very much for your careful review and feedback!

Weakness: It could be better to analyze the runtime of the computational complexity of persistent homology regarding different point cloud sizes.

Thank you for this suggestion! We analysed the computational complexity in appendix E.2 in detail, broken up into the different steps of the TOPF pipeline. Furthermore, we provided numerical experiments for runtime while increasing the size of the point cloud up to 40,000 points in Figure 11. Finally, we provide the number of points in the point clouds of the Topological Clustering Benchmark Set in Figure 8 in the appendix and provide the mean runtime of TOPF in Table 1.

We hope to have addressed your concerns and want to thank you once again for your review!

审稿意见
3

The paper presents a method (TOPF) to extract point-level topological features of point clouds, i.e., to assign to each point in the cloud a feature vector that encodes to which generators of homology it contributes. Topological features are thereby computed across all scales using persistent homology on a (Vietoris-Rips or alpha) filtered simplicial complex built on the point cloud, and then, the 'correct' scale of interest is selected by a heuristic. To identify, which simplices of the complex generate which topological features Hodge theory is used which ensures the existence of unique harmonic representatives of (co)homology classes. In the last step, the topological features of the simplices in the complex are converted to topological features of the points in the underlying point cloud by an averaging procedure.
Empirically, the generated features are evaluated on a clustering task and their robustness against noise, downsampling and the addition of outliers is studied.

Update after rebuttal

I thank the authors for their rebuttal and the additional explanations. Unfortunately, the main issue I identified, i.e., a limited empirical evaluation, see Claims and Evidence (i), was not addressed. I will therefore keep my score.

给作者的问题

  • When you describe the main ideas in section 2, you relate the kernel of the Laplacian to the homology groups. However, the Laplacian typically operates on cochains and thus, Hodge isomorphy typically holds between cohomology and the space of harmonic forms. Would you elaborate on how you identify homology and the kernel of the Laplacian and how this relates to Step 3 in section 3? Without this information, I found the latter (which is a key component of the method) a bit confusing. (I am aware of your comparison with de Rham homology in Appendix B.2 but did not find the neccessary information there).

  • Would you please compare your method to TPCC by Grande & Schaub. I am specifically curious why TOPF is able to outperform TPCC on some datasets (HalfCircles, Ellipses, ...) by a large extent but achieves similar performance on SphereInCircle .

  • Minor: Do you have an explanation or conjecture why TOPF performs rather weak on the 4Circles+Grid data?

论据与证据

The paper makes the following claims (in the 'contributions' paragraph)

  1. "TOPF (i) outperforms other methods and embeddings for clustering downstream tasks on topologically structured data."
  2. "it 'returns provably meaningful representations"
  3. "it is 'robust to noise and heterogeneous sampling"

Regarding 1. This is true for the clustering tasks studied in the empirical evaluation. However, the underlying data is introduced with this work and specifically designed such that the method can shine. That is, the only signal in the data is the topological information, see Figure 8. Therefore, the claim (1) is not wrong, but the bar is set rather low.

Regarding 2. This likely refers to Theorem C.1 in the appendix which states that the method returns correct results when points are sampled uniformly from a sphere in Rd\mathbb R^d. While it seems plausible that the method is correct under more general settings, provably meaningful representations are showed only in this very simple setting.

Regarding 3. This is true for the addition of outliers and noise in Figure 6. The downsampling results in Figure 5 are mixed, however. The method outperforms the baselines to a large extent under moderate downsampling (factor 1--10) but fails (much more than the baselines) for larger factors.

However, while the author exaggerate when listing the contributions, in the remainder of the paper they are actually very nuanced and careful about the scope, limitations and underlying assumptions of their method. This gives a very positive and honest impression.

方法与评估标准

The chosen evaluation criteria do make sense. That being said, the empirical evaluation is of too limited scope to verify the claims from above, as (aside from the qualitative evaluation in Figure 3) the generated features are only evaluated with respect to clustering on one dataset which was specifically designed so that the method can shine (see above).
It is surprising that the generated features are not evaluated for different tasks, particularly as a central motivation to this work is that "common machine learning applications like classification require point-level information and features". When reading this sentence in the abstract I was already expecting experiments which verify that the topological features computed by TOPF provide complementary information that is beneficial for these common machine learning tasks the authors had in mind.

Regarding the clustering results in table 1. I was wondering what performance classic clustering algorithms based on manifold distances would achieve on the datasets, e.g., PAM using distance estimates from Isomap (without the MDS step) or something similar.

理论论述

Theoretical results (even the statements) are only presented in the supplementary material, which according to the reviewing instructions "encouraged (but not required) to read". Due to the high reviewing load I decided not to thoroughly check the these results.

实验设计与分析

There are no issues apart from the ones already listed in §Methods And Evaluation Criteria.

补充材料

I read sections A -- D and I (but did not thoroughly check the proof in section C), but only skimmed Sections E -- H.

与现有文献的关系

The paper is situated in the domain of topological data analysis and introduces a method to generate point-level features that encode topological information. The connection to machine learning consists in that these features can be used by machine learning models and that this is done in the experiments in this paper. From my perspective, ICML might not be the best fit for the paper, but it still fits.

Closely related is work by Grande & Schaub (Topological point cloud clustering, ICML 2023) who also cluster points based on topological features computed from the Hodge laplacian and proceed in similar steps, i.e., building a complex, computing homology representatives using the laplacian, aggregating information from the complex to the point cloud. In fact, the authors list sevaral limitations of the former (TPCC) and state that they revamp their pipeline in a way that resolves them, but they do so only in the supplementary material.
Given the close relation between these two works, a short description of TPCC should be added to the main part together with a list of changes done for TOPF and their effects.

遗漏的重要参考文献

I am unaware of missing related work.

其他优缺点

A major strength of this work (which regrettably did not fit into the previous points) is the proposed method itself. The way the point features are computed leads to interpretable features, appears to be easy to implement, computationally efficient and overall, a good idea. In addition to the usefulness of the features for downstream learning tasks (which as discussed above should have been analyzed more extensively) they can serve also for visualization purposes, cf. Figures 2 & 3. Moreover, source code is provided.

其他意见或建议

Typos: In the caption of figure 5, there is a missing ':' after 'Left'
There appears to be a typo in Step 3 regarding the multiple usage of ϵ\epsilon around line 261 and one has to guess what is actually meant there.

作者回复

Thank you very much for your detailed and thorough review. We will reply to the raised issues in as much detail as the character limit permits.

Claims And Evidence: Thank you for your detailed feedback! While we still believe that the listed contributions are technically true, we now realise that parts might be misleading. We omit the “provably” from (ii). Regarding point (iii), we interpret this differently: When downscaling and using heterogeneous downsampling, from a certain degree onwards, there is no topological signal left in the data. The other algorithms detect simpler structural features and thus are not affected. However, without topological features, the performance of TOPF degrades, which is some form of sanity check. In Figure 16, Bottom left, we see that even a downsampling factor of 0.2 significantly deteriorates the topological signal.

However, we will write "[...] robust to moderate noise [...]".

I was wondering what performance classic clustering algorithms based on manifold distances would achieve on the datasets,

This is a good suggestion: PAM using Isomap distances has a mean ARI of 0.58 with 0.39/0.56/0.78/0.13/0.63/0.58/0.94 on the individual datasets. This puts PAM+ISOMAP between the performance of clustering algorithms directly on the points (kMeans mean ARI: 0.44) and TOPF (mean ARI: 0.86). We will add this to table 1.

a short description of TPCC should be added to the main part together with a list of changes done for TOPF and their effects.

Thank you for this good suggestion, we will do this!

Comments and Suggestions: Thank you very much, we have addressed your comments.

Questions for authors:

[...] Would you elaborate on how you identify homology and the kernel of the Laplacian and how this relates to Step 3 in section 3? [...]

This is a good question! Formally, the cochain space is the dual of the chain space. For finite-dimensional vector spaces with the standard inner product, there is a canonical isomorphism between the two spaces induced by sending a vector vv to the linear extension of the map sending vv to 11. Thus, we can identify the chains and cochains. We have the boundary maps BiB_i and the coboundary maps BiTB_i^T. By the univ. coefficient theorem real-valued homology and cohomology are isomorphic, with Hi=kerBi/ImBi+1Hi=kerBi+1T/ImBiTH_i=\ker B_i/Im B_{i+1}\cong H^i=\ker B^T_{i+1}/Im B_i^T. Thus, using the canonical isom. to the dual vector space, kerBiker B_i are homology and kerBi+1Tker B^T_{i+1} are cohomology representatives. kerLi\ker L_i is the intersection of kerBi\ker B_i and kerBi+1T\ker B^T_{i+1}. Hence every element in kerLi\ker L_i automatically is a homology representative. This gives us an explicit isomorphism between kerLi\ker L_i and the (co-)homology.

In step 3, we start with some arbitrary homology representative r in kerBi\ker B_i. However, we want a harmonic representative h in kerLi=kerBikerBi+1T\ker L_i = ker B_i\cap ker B^T_{i+1} for the same homology class, i.e. with h =r - c for c in imBi+1im B_{i+1}, the curl space quotiened out in homology. By the orthogonality of the Hodge decomposition, this c is given by the projection of r to the curl space imBi+1im B_{i+1}. We will discuss this in step 3 of section 3 and in more detail in appendix B.2!

I am specifically curious why TOPF is able to outperform TPCC on some datasets (HalfCircles, Ellipses, ...) by a large extent but achieves similar performance on SphereInCircle .

As you mentioned, we already give a brief comparison. The performance differences of TPCC and TOPF on different comes mainly down to how “difficult” and “robust” the topological structure encoded in the datasets is. In particular, 2Spheres2Circles and SphereInCircle are directly sampled from unions of manifolds without any noise and sufficient sampling density. In comparison, in Ellipses and Spaceships, the topological features have shorter life times and live at different scales. There is no single scale containing all features, and for most holes/voids, there is not even a scale whithout noisy holes with small persistence present in the filtration. TOPF can deal with these situations, whereas TPCC cannot. For the difference on HalvedCircle, we posit that this is due to the better feature aggregation of TOPF. TPCC requires to perform subspace clustering on an harmonic edge embedding, which is both unstable and sensitive to parameters, which we think TPCC has no information to choose correctly in this setting.

Do you have an explanation or conjecture why TOPF performs rather weak on the 4Circles+Grid data?

Figure 8&9 show the ground truth and TOPF clustering on 4Circles+Grid. In short, TOPF chooses suboptimal scales which are not well-enough connected. Figure 10 supports this interpretation, increasing the interpolation hyperparameter lambda improves TOPF performance on this dataset. We have used a fixed set of hyperparameters for all experiments on the TCBS for transparency. We will add an extended discussion of this to the paper.

审稿意见
2

Inspired by TDA, the authors proposed a point-level topological representation learning method for point cloud data analysis. Specifically, they introduced topological point features (TOPF) to extract point-level features from point clouds through discrete algebraic topology and differential geometry. The TOPF allows local feature extraction while retaining high-order geometric information. Experimental results show that TOPF outperforms existing methods in clustering and robustness tests, and is verified on synthetic data and real datasets.

Update after rebuttal

I maintain my rating.

给作者的问题

See weakness.

论据与证据

Yes.

方法与评估标准

Yes.

理论论述

Yes. The author introduces some theoretical concepts of topology and persistent homology in detail.

实验设计与分析

Yes. In Table 1, the authors compared TOPF clustering with other feature/clustering algorithms, showing some effectiveness.

补充材料

Yes. The author provides the concept of simplicial complexes and the continuous homology process based on different simplicial complexes in the supplementary materials. Finally, the topological features on the point cloud are shown.

与现有文献的关系

The authors use TDA and geometric learning for point cloud analysis. They include current literature on persistent homology, Hodge Laplace operator, and topological clustering methods.

遗漏的重要参考文献

No.

其他优缺点

Strengths:

  1. The authors propose a new approach based on persistent homology to link global topological features with point-level information.
  2. This method is robust to noise and non-uniform sampling.
  3. TOPF does not require a large amount of training data and is more interpretable and widely applicable. Weakness: 1.Although the author analyzed TOPF from a theoretical perspective, I still think that the author's TDA cannot be actually applied in the current field of point cloud analysis, and it is currently only in the theoretical analysis stage. Its high complexity and time consumption make it difficult to apply in practice.
  4. At the same time, the authors use simple point cloud datasets. Can you provide the results and time complexity on ModelNet40 and ScanObjectNN, or even experimental results in large-scale scenarios?
  5. This method may not be applicable to very high dimensional point clouds, and the authors should explicitly discuss its limitations.
  6. Is it possible to provide a detailed computational complexity analysis of TOPF?
  7. Compared with existing neural network-based methods, can TOPF outperform them? Please provide relevant experiments and analysis.

其他意见或建议

  1. The legend of Figure 1 is too long, and it is recommended to split it into multiple sentences.
  2. The resolution of Figure 3 needs to be improved to enhance readability.
作者回复

Thank you very much for your feedback! We are happy about the many strengths of TOPF identified by you. We will now address your comments:

Weakness 1: Although the author analyzed TOPF from a theoretical perspective, I still think that the author's TDA cannot be actually applied in the current field of point cloud analysis, and it is currently only in the theoretical analysis stage. Its high complexity and time consumption make it difficult to apply in practice.

Thank you for your comment! We applied TOPF on synthetic and a variety of real-world datasets and conducted experiments on large-scale point clouds in appendix E. Furthermore, persistent homology, which has similar runtimes, has been used successfully in the literature and many applications. We believe that this is convincing evidence that TOPF is a valuable scientific contribution. Because TOPF has the single goal of extracting topological information from homology, we do not claim it to be relevant in all point cloud tasks.

(Indexing continues at weakness 4 in your review.)

  1. Please refer to our rebuttal to reviewer yYZ5. Furthermore, we evaluate large-scale scenarios of point clouds with up to 40,000 points in E.2, showing that TOPF, particularly with the proposed landmark downsampling heuristic, is feasible even on large point clouds.

  2. This method may not be applicable to very high dimensional point clouds,

Thank you for your comment! TOPF works well on high-dimensional data sets. In the paragraph “Embedding Space of Variational Autoencoders and High-dimensional spaces” of sec. 4, we apply TOPF successfully to high-dimensional latent spaces and directly to an 8478(!)-dimensional image space. In Figure 4, we show a projection of the results of TOPF on a 24-dimensional point cloud.

and the authors should explicitly discuss its limitations.

We agree with your comment: We already give a summary of the limitations of TOPF in sec. 5. We discuss the limitations in great detail in section I: Limitations in the appendix.

  1. Is it possible to provide a detailed computational complexity analysis of TOPF?

Thank you for your comment! We discuss the computational complexity in detail in section E.2 in the appendix, split into the different steps of the algorithm and conduct experiments regarding the scaling behaviour.

  1. Compared with existing neural network-based methods, can TOPF outperform them? Please provide relevant experiments and analysis.

Thank you for your comment! TOPF outperforms neural network architectures on the task TOPF was designed for, namely extracting topological features from point clouds. We benchmark TOPF’s performance on the TCBS against the neural network-based methods PointNet and WSDesc, see Table 1. Furthermore, we have now added a benchmark against DGCNN (Wang et al.). TOPF outperforms these approaches (pretrained on large-scale part segmentation datasets) with a mean ARI of TOPF of 0.86 to 0.44 (PointNet) 0.39 (WSDesc) 0.64 (DGCNN). We discuss this in more detail in Section 4. We have pretrained the neural architectures on large-scale shape segmentation data sets. For a detailed experimental setup, see Appendix F: More Details on the experiments.

The legend of Figure 1 is too long, and it is recommended to split it into multiple sentences.

Thank you for this suggestion! We hope the caption of Figure 1 to be self-contained and an explanation for the figure, thus we do not know how to further shorten it. Do you have any suggestions? We already broke down the caption into 10 short to medium-length sentences. Does this suffice for you?

The resolution of Figure 3 needs to be improved to enhance readability.

We are apologise, but we do not understand this comment. On our end, the figure has sufficient resolution and an associated file size of 1.5 MB. Could you further elaborate on your suggestion?

Thank you again for your review! Based on your feedback, we believe the paper is in an even stronger state than before! We believe to have addressed a majority of your concerns and would be interested in hearing back from you!

最终决定

The draft proposes the association of a vector of topologically derived features to points in a cloud as a means to enrich the representation for downstream tasks such as segmentation. I'm somewhat divided in my appreciation of the novelty of the main idea/rationale as some other authors (including myself) have used similar tricks of associating vectors with some topological information to our observations in order to achieve some desired boost to whatever the task at hand. But at the same time, the methodology itself is clearly novel. Further, in line with this year's guidelines for meta reviews, the paper is certainly sound and correct. I suppose this dichotomic feeling is somewhat backed by the original split scores (two 2 and two 4; although one of the 2 was later increased after the rebuttal). Among the things appreciated by the reviewers are the sensible theoretical claims, the well-explained approach (to which I add it is supportive of the replicability), and the lots of good stuff in the appendices. On the weaknesses, unfortunately, the results shown in Table 1 are disparate and there are no inferential statistics to support empirical claims. When it comes to the rebuttal it is clear that the authors played a strategy of providing strong answers to those comments they appear to like, but attempted to discredit those they did not like instead of affording a credible answer or simply disregarding them e.g. "we have addressed your comments" yet I can't see any explanation. This results in a mixed rebuttal with lights and shadows.