-
Submanifold sparse convolution, 2017
- sparse convolution SC(m, n, f, s): operate on active sites
- small change to the convolution operation, it may bring computational benefits
- pooling are variants of SC(·, ·, f, s)
- submanifold sparse convolution SSC(m, n, f): |active sites| of input & output of SSC are the same,
- i.e. restrict an output site to be active if and only if the site at the corresponding site in the input is active
- Deconvolution: also keep |active sites| of input & output layers.
- 太硬,receptive field的增大全靠pooling
-
Spatial pruned sparse convolution, NeurIPS 22
- SPSC only operates on foreground sites, measured by feature magnitude [sparser than SSC]
- Dilation on important voxels instead of no dilation during downsampling [denser than SSC]
-
Part-A2, tpami 21 关注instance内部辅助信息及特征的学习。
- intra-object part locations
- ROI-aware Point cloud pooling
- 从了解、对比经典的角度,也值得一看。
-
VoxelNeXt, cvpr23
- center-missing issue can also be simply skipped through sparse networks that have large receptive fields. How?
- Additional 2 down-samplings & combine active voxels from the last 3 feature levels.
- Spatially Voxel Pruning: downsampling中适当增加receptive field
- predicate boxes from active voxels with large scores of classification.
-
HEDNet, NeurIPS 23
- hybrid detector
-
SAFDNet, CVPR 24 oral simple & better than FSD V2, 23
- HEDNet + adaptive feature diffusion / dilation ⇒ fully sparse
- AFD enlarges receptive field effectively
- uniform dilation ⇒ good performance but inefficiency
- no dilation (submanifold spare CNN) ⇒ efficiency but inferior performance
- solve center feature missing via nearest heatmap instead of center heatmap.