PaperHub
6.0
/10
Poster4 位审稿人
最低6最高6标准差0.0
6
6
6
6
3.5
置信度
ICLR 2024

LiDAR-PTQ: Post-Training Quantization for Point Cloud 3D Object Detection

OpenReviewPDF
提交: 2023-09-20更新: 2024-04-21

摘要

关键词
Quantization3D Object DetectionAutonomous Driving

评审与讨论

审稿意见
6

This paper proposes a post-training quantization method for LiDAR-based 3D object detection. It uses min-max quantization and proposes task-guided global positive loss. The experiments are conducted in the Waymo Open Dataset with CenterPoint detector, achieving competitive performance among other quantization methods.

优点

  • This paper achieves competitive results compared with other quantization methods.
  • The writing is clear and easy to follow.
  • Although the proposed method is quite simple, it is fast and pratical.

缺点

This paper lacks novelty and originality.

  • Although the authors propose the so-called sparsity-based calibration and conduct the spatial sparsity analysis, the implementation simply adopts the max-min calibration without any unique designs.
  • The proposed TGPL is also a trivial technique, which basically selects some high-confidence prediction for supervision and this is widely-adopted in many fields such as distillation.

Some claims are inaccurate or wrong.

"Point cloud coordinates have explicitly encoded the shapes and geometrics of objects, so the larger arithmetic range of input point cloud preserves essential geometric information." By this sentence, the authors indicate the large range of point cloud coordinates makes the detectors sensitive to quantization range. However, in fact, LiDAR-based object detectors do not take the absolute point cloud coordinates as input. Voxel encoders usually calculate the relative coordinates within each voxel as input.

The application is a little bit narrow.

There are emerging methods of "fully sparse detectors[1][2][3]", which show impressive efficacy and efficiency and do not lead to zero feature maps. Thus the proposed method (sparsity-based calibration) has no advantages in these methods.

[1] Fully Sparse 3D Object Detection, NeurIPS 2022.
[2] Super Sparse 3D Object Detection, TPAMI 2023.
[3] VoxelNeXt: Fully Sparse VoxelNet for 3D Object Detection and Tracking, CVPR 2023.

问题

No more additional questions. Looking forward to the authors' response to my concerns in the weaknesses box.

评论

Q1: Defend our novelty.

A1: Our research is rooted in a deep understanding of the sparsity characteristics in the LiDAR scenario. We discover two fundamental phenomena (i and ii on Page 5) when it comes to quantization over point cloud data that intrinsically bears sparsity. The choice of the max-min calibration strategy is non-trivial. By selecting maximum/minimum values, we are free from the influence of sparse data distribution. It retains the most effective information and is not affected by the large number of zeros. On the contrary, an entropy-based calibration seeks to preserve data distribution which suffers from abundant zero values and sparse information, rendering much inferior performance (Table 1 and 2). For instance, when the distance range gets beyond 50m, its mAPH suddenly drops from 38.09 to 5.9. Besides, Our proposed max-min doesn't come alone, we apply a simple grid search strategy for better quantization scales. As for TGPL, we pursue a global supervised mechanism to bridge the performance gap caused by the previous per-layer quantization. These three components altogether bring a state-of-the-art performance shown in Table 4. Each component alone may seem simple, but an orchestration of them requires deeper thoughts.

Q2: Claims about point cloud coordinates.

A2: In fact, absolute coordinates are also used in practice. In BEV-based point cloud detectors, such as voxel feature encoding (VFE) and pillar feature encoding (PFE), the Euclidean distance calculated based on the absolute coordinate of the point cloud or the absolute coordinates are also utilized as input features, along with the relative coordinates of voxels or pillars, which is a common operation. Please refer to open-source code https://github.com/tianweiy/CenterPoint/blob/d3a248fa56db2601860d576d5934d00fee9916eb/det3d/models/readers/pillar_encoder.py#L147 or https://github.com/open-mmlab/mmdetection3d/blob/5c0613be29bd2e51771ec5e046d89ba3089887c7/mmdet3d/models/voxel_encoders/voxel_encoder.py#L432.

Q3: Application on other LiDAR detectors.

A3: Thanks for the advice about more LiDAR detectors. We have conducted experiments on FSD, an excellent fully sparse 3D detector, as shown in the table below. Adopting entropy calibration still leads to a significant accuracy drop of -61.50 mAPH/L2. We discover that quantized FSD readily delivers the desired performance while employing a vanilla max-min calibration. Nonetheless, using LiDAR-PTQ can further achieve comparable accuracy to its float counterpart. This demonstrates the effectiveness of LiDAR-PTQ on fully-sparse detector as well. To our best knowledge, the mainstream approaches in industrial applications are still non-fully sparse detectors like VoxelNet[1], PointPillars[2], and CenterPoint[3] because of friendly deployment using off-the-shelf frameworks like TensorRT. We emphasize that our method can be seamlessly integrated to enjoy a significant speed-up, which is vital for the community. In summary, our LiDAR-PTQ is not only limited to specific BEV-based point cloud detectors. Given such generalization ability and practical value for promoting the application of 3D point cloud detectors on edge devices, we believe our method deserves to be shared with our community.

ModelsMethodsBits(W/A)Mean mAP/L2Mean mAPH/L2Veh mAP/L2Veh mAPH/L2Ped mAP/L2Ped mAPH/L2
FSDFull Prec.32/3273.0170.9470.3469.9873.9569.16
FSDEntropy8/810.54 (-62.47)9.44 (-61.50)0.060.0621.8818.86
FSDMax-min8/871.24 (-1.77)69.37 (-1.57)68.4268.1872.0967.60
FSDLiDAR-PTQ8/872.8470.7369.9569.6273.8568.95

References:

  • [1] VoxelNet: End-to-end learning for point cloud based 3D object detection. CVPR 2018.
  • [2] Pointpillars: Fast encoders for object detection from point clouds. CVPR 2019.
  • [3] CenterPoint: Center-based {3d} object detection and tracking. CVPR 2021.
评论

Thanks for the additional experiments, which suggest the proposed method is also suitable for the fully sparse detectors. It seems to be a good method in industrial, so I suggest authors release the source code, making the work more valuable. There is another minor issue. Although the authors point out that many methods can adopt the absolute coordinates, but, in fact, this functionality is usually not enabled. See https://github.com/open-mmlab/mmdetection3d/blob/5c0613be29bd2e51771ec5e046d89ba3089887c7/configs/_base_/models/pointpillars_hv_secfpn_waymo.py#L21

The additional experiments should be incorporated into the revised paper.

评论

Dear reviewer:

Many thanks for your valuable remarks.

Q1: Code and additional experiments.

A1: Thanks for your advice. Regarding the code, we will release it to the community soon. About the additional experiments, we have incorporated them in the supplementary materials and have added more details in the revised paper.

Q2: Make claims more accurate.

A2: We have updated the statement of point cloud encoding in Sec 3.2. in the revised paper. (changes are marked with blue).

We hope to resolve your doubts with our best efforts. Please share your thoughts on viewing our reply.

Regards,

The authors

审稿意见
6

Running neural nets for object detection on an edge device requires quantization of the network so that it can fit on the limited memory of the device and perform inference at an adequate (real-time) speed. Traditionally, post-training quantization (PTQ) for vision tasks utilizes a small set of unlabeled samples as a calibration set. These samples are used to quantize the weights and activations of the network using max-min and information theory to minimize KL-divergence between the quantized and non-quantized tensor. The information-theory based method which performs well for 2D object detection does quite poorly on 3D LIDAR object detection because of sparsity in the point cloud. This paper describes a LIDAR-PTQ method that quantizes existing LIDAR object detectors like the Centre-point-pillar. It uses a sparsity based calibrator, a Task-Guided Loss (which calibrates the quantization parameters using performance on the final object detection task) and an adaptive rounding-to-nearest loss which determines whether the weight in a particular neuron should be rounded up or down.

Experiments demonstrate similar performance to the floating point model on the NuScenes and Waymo datasets, but with a 3x inference speedup on a Jetson AGX Orin edge device.

优点

The paper is on the whole well written and does a good job of introducing the reader to the subject of neural network quantization. It describes an algorithm for the important task of adapting large neural nets for operation on the edge, while preserving model accuracy. Since there is relatively little work on LIDAR network quantization, the paper is timely and important.

缺点

There is not enough detail about the optimization steps: particularly which parameters are optimized when and what percentage of training data is required for this post-training quantization. The task-guided loss minimizes the distance between the output of the quantized network vs the non-quantized network and so presumably requires the labelled training data. This seems to deviate from the requirements of PTQ, which only seems to need a small number of unlabeled samples as the calibration set. Even though the combined system with max-min quantization + grid search + TGPL + Adaptive Rounding together approach the mAP of the FP model, the contribution in narrowing the gap between max-min and the final model seems to be mainly by the grid search. The TGPL and Adaptive Rounding contribute relatively small gains to the mAP. This could be a quirk of the dataset / optimization parameters / training regime and I would ideally like to see more experiments to be convinced of the efficacy of TGPL and Adaptive Rounding, even though from a conceptual point of view, they make sense. I would like the code to be released so that independent validation of the authors’ experiments can be performed.

问题

-What is LEVEL_2 mAPH in Table 1? Please clarify this closer to the table -Page 5: Huge ‘sparisity’ lead to inappropriate quantization range. ‘Sparisity’ should be sparsity -Page 7: ‘… are the soft-quantized weights follows Eq 11’. I think you mean equation 10? -Page 7: ‘NuSecens’ should be NuScenes

评论

Q1: Training details

A1: In LiDAR-PTQ, for every quantized layer, the optimized parameter is weight scale sws_w, weight zero-point zwz_w, activation scale sas_a, activation zero-point zaz_a and adaptive rounding value for weight θ\theta. We generally summarize the overall optimization step as follows: in every quantized layer

  1. using sparsity-based calibration to get initial sws_w, zwz_w, sas_a, zaz_a and θ\theta.

  2. using Ltotal{L}_{total} to fine-tune the sws_w, zwz_w, sas_a, zaz_a and θ\theta.

The above steps are calculated for every quantized layer in quantized model. More details refer to Algorithm 1 in Sec 3.2.

Q2: Calibrated data percentage.

A2: In WOD dataset, we randomly sample 256 frames point cloud data from the training set as the calibration set. The calibration set proportions is 0.16% (256/158,081) for WOD. Notably, we do not use the ground truth labels for these point cloud data due to we are PTQ method.

Q3: Presumably requires the labelled training data.

A3: We do not require any labeled data to train the quantized model. We use the prediction of FP model to supervise the prediction of quantized model. Employing a method similar to distillation to enhance the quantized model using its float counterpart is effective in PTQ scenarios.

Q4: Gains of TGPL and Adaptive round.

A4: The contribution of TGPL and adaptive rounding is to mitigate the final performance gap and achieve almost the same performance as FP model. There are many advanced 2D PTQ methods (like BRECQ[1], QDROP[2]) that can be used in point cloud detection models quantization. However, they still suffer from performance drops (see Tab 3). In this paper, our motivation is to propose a PTQ method designed for point cloud detection tasks and achieve comparable accuracy with FP model. Although the proposed sparsity-based calibration has shown huge performance gain, there still exists a significant gap compared with the FP model. Therefore, we propose TGPL and adaptive rounding to narrow the final performance gap.

Q5: Open-sourced code.

A5: We are arranging the code and will release it to the community later.

Q6: LEVEL2 in Waymo dataset:

A6: In Waymo Open Dataset (WOD), each ground truth label is categorized into different difficulty levels:

  • LEVEL1 (L1): if number of LiDAR points nn \textgreater\textgreater 5 in ground truth boxes.
  • LEVEL2 (L2): if number of LiDAR points 1 n\leq n \leq 5 in ground truth boxes. When evaluating, L2 metrics are computed by considering both L1 and L2 ground truth. The Waymo official leaderboard ranks based on Mean Average Precision with Heading / LEVEL2 (mAPH/L2) metrics, which is also widely adopted metrics by the community. We have added more details in the manuscript to make it clearer.

Q7: Typos:

A7: We apologize for our error. We have corrected all the typos carefully in the manuscript.

Q8: Page 7, Eq 11:

A8: Yes. In Sec 3.4, the soft-quantized weights is calculated by Eq 10 and Eq 2, and then we calculate Ltotal{L}_{total} using Eq 11.

References:

  • [1] Brecq: Pushing the limit of post-training quantization by block reconstruction. ICLR 2021.
  • [2] QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization. ICLR 2022.
评论

Dear Reviewer 3wrT,

Thanks again for the reviews. We expect any further feedback to enhance our paper and address your rest concerns before the deadline. We are desirous to have a productive conversation.

Sincerely,

The Authors

审稿意见
6

This paper proposes a Post-Training Quantization (PTQ) compression approach, LiDAR-PTQ, for the real applications of 3D LiDAR point cloud detection that are deployed on edge GPU systems. The key ideas include density-aware calibration, task-guided global positive loss and layerwise-reconstruction-based adaptive rounding. The proposed method is evaluated on the Waymo dataset to validate the low performance-drop with INT8 compared with FP32.

优点

  • The paper is easy to understand. The analysis of why 2D-PTQ can’t be directly applied to 3D is reasonable and the corresponding solution is practical.
  • The proposed LiDAR-PTQ achieves competitive detection performance as INT8 quantization for bothe CP-Pillar and Voxel.

缺点

  • The method is only evalutated on the PointCenter model, how about the effect for other types of detectors (Transform-based, etc.).
  • It is said that the method is evaluated on various datasets (Waymo and NuScenes). However, results on NuScenes seem not available.

问题

  • Comparisons of inference acceleration performance with other baselines are not available in Tab.3. Accuracy together with speed performance can better demonstrate the advantages of the proposed method.
  • How is γ set for TGPL?
  • It is wrotten that “there is no extra access to labeled training data”. Dose this work like the knowledge distillaatio from the full FP32 model for an INT8 quantizated model?
  • For results on Waymo, can authors give some explaination about why select LEVEL2?
评论

Q1: Add inference performance in Tab.3

A1: Thanks, we have updated it in Tab. 3.

Q2: γ setting in TGPL

A2: In TGPL loss, we set γ as 0.1 to filter predictions from the FP model, and then we select the top K boxes, where we set K as 500. We have updated in manuscript.

Q3: This work like the knowledge distillation from FP model for quantized model?

A3: Yes, employing a method similar to distillation to enhance the quantized model using its float counterpart is effective in PTQ scenarios. To further improve our approach, we have introduced a global distillation scheme called TGPL. Different from the commonly used knowledge distillation paradigm, we did not use any labels of the calibration set in the whole quantization process in LiDAR-PTQ.

Q4: LEVEL_2 (L2) in Waymo dataset

A4: In Waymo Open Dataset (WOD), each ground truth label is categorized into different difficulty levels.:

  • LEVEL1 (L1): if number of LiDAR points nn \textgreater\textgreater 5 in ground truth boxes.
  • LEVEL2 (L2): if number of LiDAR points 1 n\leq n \leq 5 in ground truth boxes.

When evaluating, L2 metrics are computed by considering both L1 and L2 ground truth. The Waymo official leaderboard ranks based on Mean Average Precision with Heading / LEVEL2 (mAPH/L2) metrics, which is also widely adopted metrics by the community.

Q5: Effect for other types of detectors (such as Transform-based).

A5: 1. Thank for your suggestions. Currently, this work primarily focuses on CNN-based Lidar detectors, which remain the mainstream approach in current industrial applications. Besides, we endorse the potential and promising prospects of Transformer-based methods. Combination of Transformer networks and Lidar data will bring further intriguing challenges. We are eager to explore this topic in our future works.

  1. Not only limited to the CenterPoint (-Pillar and -Voxel) model, we also evaluated the effectiveness of our method on fully sparse detectors, such as FSD[1], and achieved comparable performance with the FP model. Additionally, we provided the differences between different model architectures, which show LiDAR-PTQ is applicable to differnt cnn-based detectors. | Models | Methods | Bits(W/A) | Mean mAP/L1 | Mean mAPH/L2 | Veh mAP/L2 | Veh mAPH/L2 | Ped mAP/L2 | Ped mAPH/L2 | |-----------|------------|-----------|---------------|----------------|--------------|---------------|--------------|---------------| | FSD[1] | Full Prec. | 32/32 | 73.01 | 70.94 | 70.34 | 69.98 | 73.95 | 69.16 | | FSD[1] | LiDAR-PTQ | 8/8 | 72.84 | 70.73 | 69.95 | 69.62 | 73.85 | 68.95 |
MethodRepresentationBackboneNeckHead
CP-PillarPillarDenseDenseDense
CP-VoxelVoxelSparseDenseDense
FSDPoint+VoxelSparseSparseSparse

References:

  • [1]. Fully Sparse 3D Object Detection, NeurIPS 2022.
评论

Q1: nuScenes dataset experiments

A1: Thank you for your reminder. In the revision, we have updated the performance on the nuScenes [1] dataset. In the nuScenes dataset, we randomly sample 256 frames point cloud data from the training set as the calibration set. The calibration set proportions are 0.91% (256/28,130). Our performance evaluation involves two metrics, average precision (mAP) and nuScenes detection score (NDS). NDS is a weighted average of mAP and other attributes metrics, including translation, scale, orientation, velocity, and other box attributes. As shown in the table below, LiDAR-PTQ sitll shows state-of-the-art performance, consistent with its accuracy on Waymo dataset. Our LiDAR-PTQ achieves almost the same performance as the full precision model on the nuScenes dataset. All experiments have been updated in the supplementary materials.

ModelsMethodsBits(W/A)NDSmAPCarTruckBusTrailerCVPedMotorBCTCBarrier
CP-PillarFull Prec.32/3260.350.083.850.661.831.29.279.444.120.257.761.3
CP-PillarBRECQ8/856.943.675.941.454.321.63.878.137.415.755.053.3
CP-PillarQDrop8/857.845.978.844.257.023.85.278.440.117.656.756.8
CP-PillarPD-Quant8/859.648.381.847.659.428.27.878.441.619.857.661.0
CP-PillarLiDAR-PTQ8/860.249.883.750.861.830.69.079.043.620.457.861.0
CP-VoxelFull Prec.32/3264.856.684.654.566.736.416.983.156.139.664.064.3
CP-VoxelBRECQ8/862.051.276.546.860.528.912.580.453.734.858.859.1
CP-VoxelQDrop8/863.254.082.148.564.932.915.181.155.136.960.963.7
CP-VoxelPD-Quant8/863.755.283.751.166.634.116.682.855.136.462.663.2
CP-VoxelLiDAR-PTQ8/864.756.584.654.266.736.416.683.356.039.463.664.4

We hope to resolve your doubts with our best efforts. Please share your thoughts on viewing our reply.

References: [1] Nuscenes: A multimodal dataset for autonomous driving. CVPR, 2020.

评论

Dear Reviewer Dy1B,

Many thanks for your valuable remarks.

To make the best use of the discussion period and to improve our work, we'd like to hear your response to know whether we have addressed your concerns or there are any new questions arise.

We hope to resolve your doubts with our best efforts. Please share your thoughts on viewing our reply.

Regards,

The authors

审稿意见
6

This paper describes an approach for model compression for applications using 3D Lidar data, namely 3D object detection. The approach is based on post-training quantization. The paper shows that direct application of post-training quantization methods used with 2D images leads to a decrease in performance. The method proposed in the paper is based on a sparsity-based calibration followed by the application of a task -guided global positive loss and finally by the application of an adaptive rounding-to-nearest operation. Experimental results show that the proposed approach achieves state-of-the-art performance, namely that the accuracy of the model with PTQ INT8 is at the level of the accuracy of the model using FP32. The inference speed is increased 3 times.

优点

The paper shows that the simple and direct application of approaches commonly used for model compression in applications with 2D images leads to a decrease in performance (when applied to object detection with 3D lidar data) and proposes a new method for model compression (for 3D Lidar data and for object detection). This new model achieves similar accuracy to a full precision model and achieves an increase in inference speed (3x). The authors also analyze and discuss the reasons that may be the cause of such behavior. Such analysis is qualitative---no experimental validation. The proposed approach is clear and contributes a solution to a well-defined problem.

缺点

The main weakness of the paper results from the experimental results (and validation) having been obtained for a single application: CenterPoint (in both versions, Pillar and Voxel). This is too specific, and it is not clear that similar results could be obtained with other applications using 3D Lidar data, e.g., segmentation.

问题

Two main questions: --Would similar results be obtained with other object detectors based on Lidar data? --Would similar results be obtained with other applications using Lidar data, e.g., segmentation?

伦理问题详情

No ethics concerns.

评论

Q1: Would be similar results on other object detectors?

A1: 1. CenterPoint integrates two milestone works in LiDAR-based BEV detection, VoxelNet and PointPillars as CP-Pillar and CP-Voxel. In particular, CP-Pillar and CP-Voxel have different network designs. The CP-Pillar model is a fully dense convolutional network, while the CP-Voxel model includes SPConv and dense convolution. Our results on CenterPoint (-pillar and -voxel) demonstrate that:

  • Lidar-PTQ is applicable to pillar-based and voxel-based detectors.
  • Lidar-PTQ is applicable to SPConv and dense convolution operations.
  1. Except CP-Pillar and CP-Voxel models, to evaluate the generalization ability of our method, we also evaluate our LiDAR-PTQ on FSD, a fully sparse 3D detector. The experiments demonstrate that our method is also applicable to fully sparse detectors.
ModelsMethodsBits(W/A)Mean mAP/L2Mean mAPH/L2Veh mAP/L2Veh mAPH/L2Ped mAP/L2Ped mAPH/L2
CP-PillarFull Prec.32/3265.7860.3265.9265.4265.6555.23
CP-PillarLiDAR-PTQ8/865.6060.1265.6465.1465.5555.11
CP-VoxelFull Prec.32/3267.6765.2566.2965.7968.0462.35
CP-VoxelLiDAR-PTQ8/867.6065.1866.2765.7867.9562.24
FSDFull Prec.32/3273.0170.9470.3469.9873.9569.16
FSDLiDAR-PTQ8/872.8470.7369.9569.6273.8568.95
ModelsRepresentationBackboneNeckHead
CP-PillarPillardensedensedense
CP-VoxelVoxelsparsedensedense
FSDPoint+Voxelsparsesparsesparse

Q2: Would be similar results on other object applications, e.g., segmentation?

A2: Thanks for your constructive suggestions.

  1. For a timely response, we have not enough time to provide experiments for point cloud segmentation task due to the original intention of this paper is to focus on the LiDAR-based BEV detection task. LiDAR point cloud segmentation is another topic, which beyond the original scope of this study.
  2. Furthermore, we are actively conducting experiments in point cloud segmentation task, and if time permits, we will provide the related results before the rebuttal deadline to better evaluate our LiDAR-PTQ method.
评论

Q1: Would be similar results on other applications using Lidar data, e.g., segmentation.

A1: Given your prior constructive suggestions, we conducted experiments on SemanticKITTI [1] dataset for point cloud segmentation to further evaluate the generalization of LiDAR-PTQ. Specifically, we utilize SPVNAS [2] as our baseline, which is a representative work in point cloud segmentation. As shown in table below, adopting entropy calibration leads to a significant accuracy drop of 18.09 mIOU. As for a vanilla max-min calibration, there is still a performance drop 2.64 mIOU for quantized SPVNAS. However, LiDAR-PTQ can further achieve comparable accuracy to its float counterpart. This demonstrates the effectiveness of LiDAR-PTQ on point cloud segmentation tasks as well. For more detailed performance of different classes refer to Tab. 2.

ModelsMethodsBits(W/A)mIOU
Full Prec.32/3265.01
SPVNAS-TinyEntropy8/846.92 (-18.09)
Max-min8/862.37 (-2.64)
LiDAR-PTQ8/864.86
MethodmIoUcarbicyclemotorcycletruckother-vehiclepersonbicyclistmotorcyclistroadparkingsidewalkother-groundbuildingfencevegetationtrunkterrainpoletraffic sign
Full Prec.65.096.349.077.674.451.875.288.25.793.444.681.03.589.556.587.868.475.167.149.6
Entropy46.992.934.772.120.437.248.580.951.547.816.928.70.279.947.582.957.044.055.938.8
Max-min62.494.546.275.373.050.273.686.3957.292.341.578.92.187.353.485.165.371.863.048.6
LiDAR-PTQ64.996.348.778.074.352.274.587.95.993.344.080.93.589.456.487.668.374.567.249.5

We hope to resolve your doubts with our best efforts. Please share your thoughts on viewing our reply.

References:

  1. SemanticKITTI: A dataset for semantic scene understanding of lidar sequence, ICCV 2019.
  2. Searching efficient 3d architectures with sparse point-voxel convolution, ECCV 2020.
评论

Your answers clarified the issues related with object detectors. You have also added additional results specifically for segmentation (which was not the focus of the paper) ---thank you for the additional results. These additional results clearly validate your approach and therefore I think that the contribution of your approach is solid and clear.

评论

Dear reviewers,

We thank everyone's effort for the precious review. The paper has been moderately revised according to the suggestions received. We've carefully examined all the questions to write the answers. Please feel encouraged to discuss actively with us if any new doubts occur. All changes in manuscript are marked with blue.

Here is a list of updates (by page order) in the paper and appendix:

  • Revised the description of LEVEL_2 mAPH in Tab.1 (suggested by Reviewer Dy1B and 3wrT) to make it clearer. (Section 3.1)
  • Added more algorithm details in Algorithm 1 to make it clearer. (Section 3.1)
  • Added relative performance drop in Table 2 to make it clearer. (Section 3.2)
  • Revised the sentence at the end of page 6 (suggested by Reviewer 3wrT) to make it clearer. (Section 3.4)
  • Added inference acceleration performance with other baselines in Tab.3 (suggested by Reviewer Dy1B). (Section 4.1)
  • Added more implementation details at the beginning of page 7 (suggested by Reviewer Dy1B and 3wrT) to make it clearer. (Section 5.1)
  • Fixed typos mentioned by Reviewer 3wrT of Page 5 and page 7. (Section 5.2)
  • Added experiments on other type detectors (fully sparse detector) on Sup.

Sincerely,

The Authors

评论

Dear reviewers,

Could you please read the authors' responses and give your feed back? The period of Author-Reviewer Discussions is Fri, Nov 10 – Wed, Nov 22.

Many thanks,

AC

AC 元评审

This paper proposes an effective Post-Training Quantization (PTQ) method called LiDAR-PTQ. The proposed method is particularly curated for 3D lidar detection. Extensive experimental results confirm that the LiDAR-PTQ achieves state-of-the-art quantization performance when applied to CenterPoint. And the LiDAR-PTQ is cost-effective.

This paper receives four “marginally above the acceptance threshold” ratings.

Reviewer szHU gives “marginally above the acceptance threshold”. Reviewer szHU is satisfied with the performance and the efficiency of the proposed method. Reviewer szHU thinks the proposed approach is clear. But also, Reviewer szHU thinks the experiments are too application-specific. The authors gave their rebuttals. After reading them, Reviewer szHU has no further questions.

Reviewer Dy1B gives “marginally above the acceptance threshold”. Reviewer Dy1B thinks this paper is easy to understand. Reviewer Dy1B is satisfied with the performance of the proposed method. But also, Reviewer Dy1B wants to see the effect for other types of detectors. Reviewer Dy1B points out that the results on NuScenes seem not available. The authors gave their rebuttals. Reviewer Dy1B still gives “marginally above the acceptance threshold”.

Reviewer 3wrT gives “marginally above the acceptance threshold”. Reviewer 3wrT thinks this paper is well-written. Since there is relatively little work on LIDAR network quantization, Reviewer 3wrT thinks the paper is timely and important. But also, Reviewer 3wrT wants to see more details about optimization steps. The authors gave their rebuttals. Reviewer 3wrT still gives “marginally above the acceptance threshold”.

Reviewer fLuR gives “marginally above the acceptance threshold”. Reviewer fLuR thinks this paper achieves competitive results compared with other quantization methods. Reviewer fLuR thinks the writing is clear. Reviewer fLuR thinks the proposed method is practical. But also, Reviewer fLuR thinks the paper lacks novelty. Reviewer fLuR thinks some claims are inaccurate. The authors’ rebuttals partially address the concerns of Reviewer fLuR.

Therefore, based on the reviewers’ comments, the paper can be accepted by ICLR.

为何不给更高分

The idea of this paper is not very novel. Additional experiments mentioned by reviewers should be incorporated into the revised paper.

为何不给更低分

The proposed method is practical. Both the performance and the efficiency of the proposed method are satisfactory.

最终决定

Accept (poster)