Equivariant Deep Learning via Morphological and Linear Scale Space PDEs on the Space of Positions and Orientations
| Authors |
|
|---|---|
| Publication date | 2021 |
| Host editors |
|
| Book title | Scale Space and Variational Methods in Computer Vision |
| Book subtitle | 8th International Conference, SSVM 2021, Virtual Event, May 16–20, 2021 : proceedings |
| ISBN |
|
| ISBN (electronic) |
|
| Series | Lecture Notes in Computer Science |
| Event | 8th International Conference on Scale Space and Variational Methods in Computer Vision, SSVM 2021 |
| Pages (from-to) | 27-39 |
| Number of pages | 13 |
| Publisher | Cham: Springer |
| Organisations |
|
| Abstract |
We present PDE-based Group Convolutional Neural Networks (PDE-G-CNNs) that generalize Group equivariant Convolutional Neural Networks (G-CNNs). In PDE-G-CNNs a network layer is a set of PDE-solvers where geometrically meaningful PDE-coefficients become trainable weights. The underlying PDEs are morphological and linear scale space PDEs on the homogeneous space Md of positions and orientations. They provide an equivariant, geometrical PDE-design and model interpretability of the network. The network is implemented by morphological convolutions with approximations to kernels solving morphological α -scale-space PDEs, and to linear convolutions solving linear α -scale-space PDEs. In the morphological setting, the parameter α regulates soft max-pooling over balls, whereas in the linear setting the cases α= 1/2 and α= 1 correspond to Poisson and Gaussian scale spaces respectively. We show that our analytic approximation kernels are accurate and practical. We build on techniques introduced by Weickert and Burgeth who revealed a key isomorphism between linear and morphological scale spaces via the Fourier-Cramér transform. It maps linear α -stable Lévy processes to Bellman processes. We generalize this to Md and exploit this relation between linear and morphological scale-space kernels. We present blood vessel segmentation experiments that show the benefits of PDE-G-CNNs compared to state-of-the-art G-CNNs: increase of performance along with a huge reduction in network parameters. |
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1007/978-3-030-75549-2_3 |
| Other links | https://www.scopus.com/pages/publications/85106399233 |
| Permalink to this page | |