Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Deep neural networks have achieved great success in the last decade. When
designing neural networks to handle the ubiquitous geometric data such as point
clouds and graphs, it is critical that the model can maintain invariance
towards various transformations such as translation, rotation, and scaling.
Most existing graph neural network (GNN) approaches can only maintain
permutation-invariance, failing to guarantee invariance with respect to other
transformations. Besides GNNs, other works design sophisticated
transformation-invariant layers, which are computationally expensive and
difficult to be extended. In this paper, we revisit why general neural networks
cannot maintain transformation invariance. Our findings show that
transformation-invariant and distance-preserving initial point representations
are sufficient to achieve transformation invariance rather than needing
sophisticated neural layer designs. Motivated by these findings, we propose
Transformation Invariant Neural Networks (TinvNN), a straightforward and
general plug-in for geometric data. Specifically, we realize transformation
invariant and distance-preserving initial point representations by modifying
multi-dimensional scaling and feed the representations into existing neural
networks. We prove that TinvNN can strictly guarantee transformation
invariance, being general and flexible enough to be combined with the existing
neural networks. Extensive experimental results on point cloud analysis and
combinatorial optimization demonstrate the effectiveness and general
applicability of our method. We also extend our method into equivariance cases.
Based on the results, we advocate that TinvNN should be considered as an
essential baseline for further studies of transformation-invariant geometric
deep learning.
Authors (5)
Ziwei Zhang
Xin Wang
Zeyang Zhang
Peng Cui
Wenwu Zhu
Submitted
December 23, 2021
Key Contributions
This paper revisits transformation invariance in geometric deep learning, arguing that achieving invariance (to translation, rotation, scaling) is primarily dependent on the initial point representations rather than solely on sophisticated layer designs. It demonstrates that transformation-invariant and distance-preserving initial representations are sufficient for achieving transformation invariance, offering a simpler and potentially more effective approach than current GNNs or complex invariant layers.
Business Value
Enables more robust and efficient processing of 3D data, crucial for applications in autonomous driving, robotics, augmented/virtual reality, and medical imaging, by ensuring models are not sensitive to object orientation or position.