Geometric Deep Learning: A Quick Tour

栏目: IT技术 · 发布时间: 5年前

内容简介:The following document provides a whirlwind tour of some fundamental concepts in geometric deep learning.Find the latex-written version of this articleThe following document provides a whirlwind tour of some fundamental concepts in geometric deep learning.

The following document provides a whirlwind tour of some fundamental concepts in geometric deep learning.

Geometric Deep Learning: A Quick Tour

Photo by Clint Adair onUnsplash

Find the latex-written version of this article here

The following document provides a whirlwind tour of some fundamental concepts in geometric deep learning. The mathematical derivations might not be as rigorously shown and some equations are stated without proofs. This is done intentionally to keep the document short yet comprehensive enough. For deeper analysis, please refer to Bronstein et al. (2017), Gong (2018), Kipf et al. (2017), Hammond et al. (2011), Shuman et al. (2011) and http://geometricdeeplearning.com/ . This is a living document, so please let me know if you find any errors or inconsistency, I will fix them as soon as possible. I am by no means an expert in this field as this note is merely written for my own learning purpose.

We first start by reviewing the concept of a Laplacian matrix in a graph and its’ eigenvalue decomposition. Next, we define convolution operation in the graph and show that it is equivalent to applying a filter in the spectral domain of the graph, where spectrum here refers to the eigenvalues of the laplacian. We show that spectral convolution, by construction, is similar to applying Laplacian operator to function. Lastly, we show how to approximate the Laplacian filter with the sum of Chebyshev polynomials to reduce the algorithm complexity (as performing the complete eigenvalue decomposition is O(n²) where n is the number of vertices in the graph.

Graph Laplacian

Assume we have an undirected, connected and weighted graph G = (V, E, W) where V is a set of |V| = n vertices, E is a set of edges and W is a set of weights W ᵢⱼ for each edge i~j . Define D as the degree matrix where D = diag(Σⱼ W ᵢⱼ). The normalized graph laplacian Δ can be defined as the following.

We can perform eigenvalue decomposition to the laplacian matrix Δ as the following.

Spectral Convolution

Motivated by the convolution operation in Fourier Transform, we define graph convolution as applying a filter to the laplacian spectral components. In a nutshell, the idea is to project the input function f to its’ “Fourier” basis, or here laplacian eigenvectors by means of Φf . Multiply the basis with filtered eigenvalues ĥ(Λ) , resulting in ĥ(Λ) Φf . Lastly, apply an “inverse Fourier transform” by dot product with Φ , resulting in Φ ĥ(Λ) Φf .

Geometric Deep Learning: A Quick Tour

After applying some rearrangements, we can see that the right hand side of the equation can be converted to a function of the normalized laplacian matrix Δ = Φ Λ Φᵀ from equation 5.

Now, similar to a convolutional neural network, we would like to apply more than 1 convolutional filters and also add a nonlinear output transformation. Suppose we use p filters with ξ nonlinear transformation, the spectral convolution network is defined as the following.

Geometric Deep Learning: A Quick Tour

Approximation with Chebyshev Polynomials

ChebNets

The main problem with equation 12 is that we need to calculate all eigenvectors which is O(|V|²) in complexity. This is not scalable when we deal with a very large graph with |V|>>. The main idea in ChebNets is to approximate the Φ ĥ (Λ) Φᵀ with a sum of orthonormal Chebyshev polynomials that can be calculated iteratively using the following formula.

We first approximate the eigenvalue filter with a weighted sum of r Chebyshev polynomials that takes in the transformed eigenvalue matrix Λ^. This transformation is needed because Chebyshev polynomials form an orthogonal basis only when the input domain interval is in [-1, 1] . Meanwhile, the eigenvalues range from [0, λ ₘₐₓ ] . Hence to transform the latter to the former, we can apply 2x/λ ₘₐₓ -1 such that for x = 0, (2(0)/λₘₐₓ-1 = -1) and for x = λₘₐₓ, (2λₘₐₓ/λₘₐₓ -1 = 1).

We can now take equation 12 and substitute the filter with its’ Chebyshev polynomials. Note that Λ^ = 2 Λ / λₘₐₓ - I and similarly, Δ^ = 2 Δ/ λₘₐₓ - I . By means of rearrangement, we can transform the right hand side into a sum of Chebyshev polynomials applied on the normalized laplacian Δ^ .

Geometric Deep Learning: A Quick Tour

GraphConvNets

GraphConvNets is a special case of ChebNets where we only take 2 Chebyshev polynomial terms r = 2 . Since the normalized laplacian Δ is used, it has been proven that λₘₐₓ = 2 . We can take equation 21 and substitute the given assumptions.

Geometric Deep Learning: A Quick Tour

We now have a graph convolution network that is defined purely using degree matrix D and weight matrix W . To further simplify, we assume α₀ = -α₁ = α.

Geometric Deep Learning: A Quick Tour

Although in practice, since the eigenvalues are in range [0, 2], repeated application of this multiplication could result in numerical instabilities. Hence we apply further normalization by letting Ŵ = W + I and D̂ = diag(Σ{j ≠i} ŵᵢⱼ).

Geometric Deep Learning: A Quick Tour

Using these assumptions, GraphConvNets bypasses the need for explicit calculation of Chebyshev polynomials and results in the pure application of weighted sum operation on each vertex’s neighbors.

References

Bronstein, M. M., Bruna, J., LeCun, Y., Szlam, A., and Vandergheynst, P. (2017). Geometric deep learning: going beyond Euclidean data. IEEE Signal Processing Magazine, 34(4):18{42. arXiv: 1611.08097.

Gong, S. (2018). Geometric Deep Learning. Imperial College London MSc Thesis.

Hammond, D. K., Vandergheynst, P., and Gribonval, R. (2011). Wavelets on graphs via spectral graph theory. Applied and Computational Harmonic Analysis, 30(2):129{150.

Shuman, D. I., Vandergheynst, P., and Frossard, P. (2011). Chebyshev Polynomial Approximation for Distributed Signal Processing. 2011 International Conference on Distributed Computing in Sensor Systems and Workshops (DCOSS), pages 1-8. arXiv: 1105.1891.

T. Kipf, M. Welling, Semi-supervised Classification with Graph Convolutional Networks , ICLR 2017


以上所述就是小编给大家介绍的《Geometric Deep Learning: A Quick Tour》,希望对大家有所帮助,如果大家有任何疑问请给我留言,小编会及时回复大家的。在此也非常感谢大家对 码农网 的支持!

查看所有标签

猜你喜欢:

本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们

在线

在线

王坚 / 中信出版集团 / 2018-5-21 / 59.00元

50多万年前的关键词是光明与黑暗, 50多年前的关键词是数字和模拟, 而今天的关键词是在线与离线。 移动互联网是比传统互联网在线程度更深的互联网。对于真正成熟的互联网来说,手机只是诸多的在线设备之一,慢慢地,每一个设备都会变成互联网的终端。 真正的竞争力,是把所有人都可能拥有的东西变成财富,让沙子变成硅。大家都把大数据当作金矿,想要掘金。但在王坚看来,大数据的厉害之处是把沙......一起来看看 《在线》 这本书的介绍吧!

RGB转16进制工具
RGB转16进制工具

RGB HEX 互转工具

随机密码生成器
随机密码生成器

多种字符组合密码

RGB CMYK 转换工具
RGB CMYK 转换工具

RGB CMYK 互转工具