Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 92% Match Research Paper Geometric Deep Learning Researchers,GNN Researchers,Computer Vision Engineers,Robotics Engineers 1 week ago

Revisiting Transformation Invariant Geometric Deep Learning: An Initial Representation Perspective

graph-neural-networks › graph-learning
📄 Abstract

Abstract: Deep neural networks have achieved great success in the last decade. When designing neural networks to handle the ubiquitous geometric data such as point clouds and graphs, it is critical that the model can maintain invariance towards various transformations such as translation, rotation, and scaling. Most existing graph neural network (GNN) approaches can only maintain permutation-invariance, failing to guarantee invariance with respect to other transformations. Besides GNNs, other works design sophisticated transformation-invariant layers, which are computationally expensive and difficult to be extended. In this paper, we revisit why general neural networks cannot maintain transformation invariance. Our findings show that transformation-invariant and distance-preserving initial point representations are sufficient to achieve transformation invariance rather than needing sophisticated neural layer designs. Motivated by these findings, we propose Transformation Invariant Neural Networks (TinvNN), a straightforward and general plug-in for geometric data. Specifically, we realize transformation invariant and distance-preserving initial point representations by modifying multi-dimensional scaling and feed the representations into existing neural networks. We prove that TinvNN can strictly guarantee transformation invariance, being general and flexible enough to be combined with the existing neural networks. Extensive experimental results on point cloud analysis and combinatorial optimization demonstrate the effectiveness and general applicability of our method. We also extend our method into equivariance cases. Based on the results, we advocate that TinvNN should be considered as an essential baseline for further studies of transformation-invariant geometric deep learning.
Authors (5)
Ziwei Zhang
Xin Wang
Zeyang Zhang
Peng Cui
Wenwu Zhu
Submitted
December 23, 2021
arXiv Category
cs.CV
arXiv PDF

Key Contributions

This paper revisits transformation invariance in geometric deep learning, arguing that achieving invariance (to translation, rotation, scaling) is primarily dependent on the initial point representations rather than solely on sophisticated layer designs. It demonstrates that transformation-invariant and distance-preserving initial representations are sufficient for achieving transformation invariance, offering a simpler and potentially more effective approach than current GNNs or complex invariant layers.

Business Value

Enables more robust and efficient processing of 3D data, crucial for applications in autonomous driving, robotics, augmented/virtual reality, and medical imaging, by ensuring models are not sensitive to object orientation or position.