Abstract
Deep Affine Normalizing Flows are efficient and powerful models for high-dimensional density estimation and sample generation. Yet little is known about how they succeed in approximating complex distributions, given the seemingly limited expressiveness of individual affine layers. In this work, we take a first step towards theoretical understanding by analyzing the behaviour of a single affine coupling layer under maximum likelihood loss. We show that such a layer estimates and normalizes conditional moments of the data distribution, and derive a tight lower bound on the loss depending on the orthogonal transformation of the data before the affine coupling. This bound can be used to identify the optimal orthogonal transform, yielding a layer-wise training algorithm for deep affine flows. Toy examples confirm our findings and stimulate further research by highlighting the remaining gap between layer-wise and end-to-end training of deep affine flows.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
1 Introduction
Affine Normalizing Flows such as RealNVP [4] are widespread and successful tools for density estimation. They have seen recent success in generative modeling [3, 4, 9], solving inverse problems [1], lossless compression [6], out-of-distribution detection [12], better understanding adversarial examples [7] and sampling from Boltzmann distributions [13].
These flows approximate arbitrary data distributions \(\mu (\mathrm {\mathbf {x}})\) by learning an invertible mapping \(T(\mathrm {\mathbf {x}})\) such that given samples are mapped to normally distributed latent codes \(\mathrm {\mathbf {z}}:= T(\mathrm {\mathbf {x}})\). In other words, they reshape the data density \(\mu \) to form a normal distribution.
While being simple to implement and fast to evaluate, affine flows appear not very expressive at first glance. They consist of invertible layers called coupling blocks. Each block leaves half of the dimensions untouched and subjects the other half to just parameterized translations and scalings.
Explaining the gap between theory and applications remains an unsolved challenge. Taking the problem apart, a single layer consists of a rotation and an affine nonlinearity. It is often hand-wavingly argued that the deep model’s expressivity comes from the rotations between the couplings by allowing different dimensions to influence one another [4].
In this work, we open a rigorous branch of explanation by characterizing the normalizing flow generated by a single affine layer. More precisely, we contribute:
-
A single affine layer under maximum likelihood (ML) loss learns first- and second-order moments of the conditional distribution of the changed (active) dimensions given the unchanged (passive) dimensions (Sect. 3.2).
-
From this insight, we derive a tight lower bound on how much the affine nonlinearity can reduce the loss for a given rotation (Sect. 3.3). This is visualized in Fig. 1 where the bound is evaluated for different rotations of the data.
-
We formulate a layer-wise training algorithm that determines rotations using the lower bound and nonlinearities using gradient descent in turn (Sect. 3.4).
-
We show that such a single affine layer under ML loss makes the active independent of the passive dimensions if they are generated by a certain rule (Sect. 3.5).
Finally, we show empirically in Sect. 4 that while improving the training of shallow flows, the above new findings do not yet explain the success of deep affine flows and stimulate further research.
2 Related Work
The connection between affine transformations and the first two moments of a distribution is well-known in the Optimal Transport literature. When the function space of an Optimal Transport (OT) problem with quadratic ground cost is reduced to affine maps, the best possible transport matches mean and covariance of the involved distributions [17]. In the case of conditional distributions, affine maps become conditional affine maps [16]. We show such maps to have the same minimizer under maximum likelihood loss (KL divergence) as under OT costs.
It has been argued before that a single coupling or autoregressive block [14] can capture the moments of conditional distributions. This is one of the motivations for the SOS flow [8], based on a classical result on degree-3 polynomials by [5]. However, they do not make this connection explicit. We are able to give a direct correspondence between the function learnt by an affine coupling and the first two moments of the distribution to be approximated.
Rotations in affine flows are typically chosen at random at initialization and left fixed during training [3, 4]. Others have tried training them via some parameterization like a series of Householder reflections [15]. The stream of work most closely related to ours explores the idea to perform layer-wise training. This allows an informed choice of the rotation based on the current estimate of the latent normal distribution. Most of these works propose to choose the least Gaussian dimensions as the active subspace [2, 11]. We argue that this is inapplicable to affine flows due to their limited expressivity when the passive dimensions are not informative. To the best of our knowledge, our approach is the first to take the specific structure of the coupling layer into account and derive a tight lower bound on the loss as a function of the rotation.
3 Single Affine Coupling Layer
3.1 Architecture
Normalizing flows approximate data distributions \(\mu \) available through samples \(\mathrm {\mathbf {x}}\in \mathbb {R}^D \sim \mu \) by learning an invertible function \(T(\mathrm {\mathbf {x}})\) such the latent codes \(\mathrm {\mathbf {z}}:= T(\mathrm {\mathbf {x}})\) follow an isotropic normal distribution \(\mathrm {\mathbf {z}}\in \mathbb {R}^D \sim \mathcal {N}(0, \mathrm {\mathbf {1}})\). When such a function is found, the data distribution \(\mu (\mathrm {\mathbf {x}})\) can be approximated using the change-of-variables formula:
where \(\mathrm {\mathbf {J}} = \nabla T(\mathrm {\mathbf {x}})\) is the Jacobian of the invertible function, and “\(\cdot _\sharp \)” is the push-forward operator. New samples \(\mathrm {\mathbf {x}}\sim \mu \) can be easily generated by drawing \(\mathrm {\mathbf {z}}\) from the latent Gaussian and transporting them backward through the invertible function:
Affine Normalizing Flows are a particularly efficient way to parameterize such an invertible function \(T\): They are simple to implement and fast to evaluate in both directions \(T(\mathrm {\mathbf {x}})\) and \(T^{-1}(\mathrm {\mathbf {z}})\), along with the Jacobian determinant \(\det \mathrm {\mathbf {J}}\) [1]. Like most normalizing flow models, they consist of the composition of several invertible layers \(T(\mathrm {\mathbf {x}}) = (T_L \circ \cdots \circ T_1)(\mathrm {\mathbf {x}})\). The layers are called coupling blocks and modify the distribution sequentially. We recursively define the push-forward of the first l blocks as
Each block \(T_l, l=1, \dots , L\) contains a rotation \(\mathrm {\mathbf {Q}}_l \in \text {SO}(D)\) and a nonlinear transformation \(\tau _l\):
The nonlinear transformation \(\tau _l\) is given by:
Here, \(\mathrm {\mathbf {y}}= \mathrm {\mathbf {Q}}_l \mathrm {\mathbf {x}}_{l-1} \sim (\mathrm {\mathbf {Q}}_l)_\sharp \mu _{l-1}\) is the rotated input to the nonlinearity (dropping the index l on \(\mathrm {\mathbf {y}}\) for simplicity) and \(\odot \) is element-wise multiplication. An affine nonlinearity first splits its input into passive and active dimensions \(\mathrm {\mathbf {p}}\in \mathbb {R}^{D_P}\) and \(\mathrm {\mathbf {a}}\in \mathbb {R}^{D_A}\). The passive subspace is copied without modification to the output of the coupling. The active subspace is scaled and shifted as a function of the passive subspace, where \(s_l\) and \(t_l : \mathbb {R}^{D_P}\rightarrow \mathbb {R}^{D_A}\) are represented by a single generic feed forward neural network [9] and need not be invertible themselves. The affine coupling design makes inversion trivial by transposing \(\mathrm {\mathbf {Q}}_l\) and rearranging terms in \(\tau _l\).
Normalizing Flows, and affine flows in particular, are typically trained using the Maximum Likelihood (ML) loss [3]. It is equivalent to the Kullback-Leibler (KL) divergence between the push-forward of the data distribution \(\mu \) and the latent normal distribution [10]:
The two differ only by terms independent of the trained model (the typically unknown entropy \(H[\mu ]\) and the normalization of the normal distribution).
It is unknown whether affine normalizing flows can push arbitrarily complex distributions to a normal distribution [14]. In the remainder of the section, we shed light on this by considering an affine flow that consists of just a single coupling as defined in Eq. (5). Since we only consider one layer, we’re dropping the layer index l for the remainder of the section. In Sect. 4, we will discuss how these insights on isolated affine layers transfer to deep flows.
3.2 KL Divergence Minimizer
We first derive the exact form of the ML loss in Eq. (6) for an isolated affine coupling with a fixed rotation \(\mathrm {\mathbf {Q}}\) as in Eq. (4).
The Jacobian for this coupling has a very simple structure: It is a triangular matrix whose diagonal elements are \(\mathrm {\mathbf {J}}_{ii}=1\) if i is a passive dimension and \(\mathrm {\mathbf {J}}_{ii}=\exp (s_i(\mathrm {\mathbf {p}}))\) if i is active. Its determinant is the product of the diagonal elements, so that \(\det \mathrm {\mathbf {J}}(\mathrm {\mathbf {x}})>0\) and \(\log \det \mathrm {\mathbf {J}}(\mathrm {\mathbf {x}})=\sum _{i=1}^{D_A}s_i(\mathrm {\mathbf {p}})\). The ML loss thus reads:
We now derive the minimizer of this loss:
Lemma 1 (Optimal single affine coupling)
Given a distribution \(\mu \) and a single affine coupling layer \(T\) with a fixed rotation \(\mathrm {\mathbf {Q}}\). Like in Eq. (5), call \((\mathrm {\mathbf {a}}, \mathrm {\mathbf {p}}) = \mathrm {\mathbf {Q}}\mathrm {\mathbf {x}}\) the rotated versions of \(\mathrm {\mathbf {x}}\sim \mu \). Then, at the unique minimum of the ML loss (Eq. (8)), the functions \(s, t : \mathbb {R}^{D_P}\rightarrow \mathbb {R}^{D_A}\) as in Eq. (4) take the following value:
We derive this by optimizing for \(s(\mathrm {\mathbf {p}}), t(\mathrm {\mathbf {p}})\) in Eq. (8) for each value of \(\mathrm {\mathbf {p}}\) separately. The full proof can be found in Appendix A.1.
We insert the optimal s and t to find the active part of the globally optimal affine nonlinearity:
It normalizes \(\mathrm {\mathbf {a}}\) for each \(\mathrm {\mathbf {p}}\) by shifting the mean of \(\mu (\mathrm {\mathbf {a}}|\mathrm {\mathbf {p}})\) to zero and rescaling the individual standard deviations to one.
Example 1
Consider a distribution where the first variable \(p\) is uniformly distributed on the interval \([-2, 2]\). The distribution of the second variable \(a\) is normal, but its mean \(m(p)\) and standard deviation \(\sigma (p)\) are varying depending on \(p\):
We call this distribution “W density”. It is shown in Fig. 2a.
We now train a single affine nonlinearity \(\tau \) by minimizing the ML loss, setting \(\mathrm {\mathbf {Q}}= \mathrm {\mathbf {1}}\). As hyperparameters, we choose a subnet for s, t with one hidden layer and a width of 256, a learning rate of \(10^{-1}\), a learning rate decay with factor 0.9 every 100 epochs, and a weight decay of 0. We train for 4096 epochs with 4096 i.i.d. samples from \(\mu \) each using the Adam optimizer.
We solve s, t in Lemma 1 for the estimated mean \(\hat{m}(p)\) and standard deviation \(\hat{\sigma }(p)\) as predicted by the learnt \(\hat{s}\) and \(\hat{t}\). Upon convergence of the model, they closely follow their true counterparts \(m(p)\) and \(\sigma (p)\) as shown in Fig. 2b.
Example 2
This example modifies the previous to illustrate that the learnt conditional density \(\tau _\sharp \mu (\mathrm {\mathbf {a}}|\mathrm {\mathbf {p}})\) is not necessarily Gaussian at the minimum of the loss.
The W density from above is transformed to the “WU density” by replacing the conditional normal distribution by a conditional uniform distribution with the same conditional mean \(m(p)\) and standard deviation \(\sigma (p)\) as before.
One might wrongly believe that the KL divergence favours building a distribution that is marginally normal while ignoring the conditionals, i.e. \(\tau _\sharp \mu (p) = \mathcal {N}\). Lemma 1 predicts the correct result, resulting in the following uniform push-forward density depicted in Fig. 2d:
Note how \(\tau _\sharp \mu (a|p)\) does not depend on p, which we later generalize in Lemma 2.
3.3 Tight Bound on Loss
Knowing that a single affine layer learns the mean and standard deviation of \(\mu (a_i|\mathrm {\mathbf {p}})\) for each \(\mathrm {\mathbf {p}}\), we can insert this minimizer into the KL divergence. This yields a tight lower bound on the loss after training. Even more, it allows us to compute a tight upper bound on the loss improvement by the layer, which we denote \(\varDelta \ge 0\). This loss reduction can be approximated using samples without training.
Theorem 1 (Improvement by single affine layer)
Given a distribution \(\mu \) and a single affine coupling layer \(T\) with a fixed rotation \(\mathrm {\mathbf {Q}}\). Like in Eq. (5), call \((\mathrm {\mathbf {a}}, \mathrm {\mathbf {p}}) = \mathrm {\mathbf {Q}}\mathrm {\mathbf {x}}\) the rotated versions of \(\mathrm {\mathbf {x}}\sim \mu \). Then, the KL divergence has the following minimal value:
The loss improvement by the optimal affine coupling as in Lemma 1 is:
To proof, insert the minimizer s, t from Lemma 1 into Eq. (8). Then evaluate \(\varDelta = \mathcal {D}_\text {KL}(\mu || \mathcal {N}) - \mathcal {D}_\text {KL}(T_\sharp \mu || \mathcal {N})\) to obtain the statement. The detailed proof can be found in Appendix A.2.
The loss reduction by a single affine layer depends solely on the moments of the distribution of the active dimensions conditioned on the passive subspace. Higher order moments are ignored by this coupling design. Together with Lemma 1, this paints the following picture of an affine coupling layer: It fits a Gaussian distribution to each conditional \(\mu (a_i|\mathrm {\mathbf {p}})\) and normalizes this Gaussian’s moments. The gap in entropy between the fit Gaussian and the true conditional distribution cannot be reduced by the affine transformation. This makes up the remaining KL divergence in Eq. (18).
We now make the connection explicit that a single affine layer can only achieve zero loss on the active subspace iff the conditional distribution is Gaussian with diagonal covariance:
Corollary 1
If and only if \((\mathrm {\mathbf {Q}}_\sharp \mu )(\mathrm {\mathbf {a}}|\mathrm {\mathbf {p}})\) is normally distributed for all \(p\) with diagonal covariance, that is:
a single affine block can reduce the KL divergence on the active subspace to zero:
The proof can be found in Appendix A.3.
Example 3
We revisit the Examples 1 and 2 and confirm that the minimal loss achieved by a single affine coupling layer on the W-shaped densities matches the predicted lower bound. This is the case for both densities. Figure 3 shows the contribution of the conditional part of the KL divergence \( \mathcal {D}_\text {KL}((T_\sharp \mu )(a|p) || \mathcal {N}) \) as a function of \(p\):
For the W density, the conditional \(\mu (a|p)\) is normally distributed. This is the situation of Corollary 1 and the remaining conditional KL divergence is zero. The remaining loss for the WU density is the negentropy of a uniform distribution with unit variance.
3.4 Determining the Optimal Rotation
The rotation \(\mathrm {\mathbf {Q}}\) of the isolated coupling layer determines the splitting into active and passive dimensions and the axes of the active dimensions (the rotation within the passive subspace only rotates the input into s, t and is irrelevant). The bounds in Theorem 1 heavily depend on these choices and thus depend on the chosen rotation \(\mathrm {\mathbf {Q}}\). This makes it natural to consider the loss improvement as a function of the rotation: \(\varDelta (\mathrm {\mathbf {Q}})\). When aiming to maximally reduce the loss with a single affine layer, one should choose the subspace maximizing this tight upper bound in Eq. (20):
We propose to approximate this maximization by evaluating the loss improvement for a finite set of candidate rotations in Algorithm 1 “Optimal Affine Subspace (OAS)”. Note that Step 5 requires approximating \(\varDelta \) from samples. In the regime of low \({D_P}\), one can discretize this by binning samples by their passive coordinate \(\mathrm {\mathbf {p}}\). Then, one computes mean and variance empirically for each bin. We leave the general solution of Eq. (23) for future work.
Example 4
Consider the following two-component 2D Gaussian Mixture Model:
We choose \(\delta = 0.95, \sigma = \sqrt{1 - \delta ^2} = 0.3122...\) so that the mean is zero and the standard deviation along the first axis is one. We now evaluate the loss improvement \(\varDelta (\theta )\) in Eq. (20) as a function of the angle \(\theta \) with which we rotate the above distribution:
Analytically, this can be done pointwise for a given \(p\) and then integrated numerically. This will not be possible for applications where only samples are available. As a proof of concept, we employ the previously mentioned binning approach. It groups N samples from \(\mu \) by their \(p\) value into B bins. Then, we compute \(m_{A|p_b}\) and \(\sigma _{A|p_b}\) using the samples in each bin \(b = 1, \dots , B\).
Figure 4 shows the upper bound as a function of the rotation angle, as obtained from the two approaches. Here, we used \(B=32\) bins and a maximum of \(N=2^{13}=8192\) samples. Around \(N \approx 256\) samples are sufficient for a good agreement between the analytic and empiric bound on the loss improvement and the corresponding angle at the maximum.
Note: For getting a good density estimate using a single coupling, it is crucial to identify the right rotation. If we naively or by chance decide for \(\theta = 90^\circ \), the distribution is left unchanged.
3.5 Independent Outputs
An important step towards pushing a multivariate distribution to a normal distribution is making the dimensions independent of one another. Then, the residual to a global latent normal distribution can be solved with one sufficiently expressive 1D flow per dimension, pushing each distribution independently to a normal distribution. The following lemma shows for which data sets a single affine layer can make the active and passive dimensions independent.
Lemma 2
Given a distribution \(\mu \) and a single affine coupling layer \(T\) with a fixed rotation \(\mathrm {\mathbf {Q}}\). Like in Eq. 5, call \((\mathrm {\mathbf {a}}, \mathrm {\mathbf {p}}) = \mathrm {\mathbf {Q}}\mathrm {\mathbf {x}}\) the rotated versions of \(\mathrm {\mathbf {x}}\sim \mu \). Then, the following are equivalent:
-
1.
\( \mathrm {\mathbf {a}}' := \tau (\mathrm {\mathbf {a}}|\mathrm {\mathbf {p}}) \perp \mathrm {\mathbf {p}}\) for \(\tau (\mathrm {\mathbf {a}}|\mathrm {\mathbf {p}})\) minimizing the ML loss in Eq. (8),
-
2.
There exists \(\mathrm {\mathbf {n}}\perp \mathrm {\mathbf {p}}\) such that \( \mathrm {\mathbf {a}}= f(\mathrm {\mathbf {p}}) + \mathrm {\mathbf {n}}\odot g(\mathrm {\mathbf {p}})\), where \(f, g: \mathbb {R}^{D_P}\rightarrow \mathbb {R}^{D_A}\).
The proof can be found in Appendix A.4.
This results shows what our theory can explain about deep affine flows: It is easy to see that \(D-1\) coupling blocks with \({D_A}=1, {D_P}= D - 1\) can make all variables independent if the data set can be written in the form of \(x_i = f(\mathrm {\mathbf {x}}_{\ne i}) + x_i g(\mathrm {\mathbf {x}}_{\ne i})\). Then, only the aforementioned independent 1D flows are necessary for a push-forward to the normal distribution.
Example 5
Consider again the W-shaped densities from the previous Examples 1 and 2. After optimizing the single affine layer, the two variables \(p, a'\) are independent (compare Fig. 2c, d):
4 Layer-Wise Learning
Do the above single-layer results explain the expressivity of deep affine flows? To answer this question, we construct a deep flow layer by layer using the optimal affine subspace (OAS) algorithm Algorithm 1. Each layer l being added to the flow is trained to minimize the residuum between the current push-forward \(\mu _{l-1}\) and the latent \(\mathcal {N}\). The corresponding rotation \(\mathrm {\mathbf {Q}}_l\) is chosen by maximizing \(\varDelta (\mathrm {\mathbf {Q}}_l)\) and the nonlinearities \(\tau _l\) are trained by gradient descent, see Algorithm 2.
Can this ansatz reach the quality of end-to-end affine flows? An analytic answer is out of the scope of this work, and we consider toy examples.
Example 6
We consider a uniform 2D distribution \(\mu = \mathcal {U}([-1, 1]^2)\). Figure 5 compares the flow learnt layer-wise using Algorithm 2 to flows learnt layer-wise and end-to-end, but with fixed random rotations. Our proposed layer-wise algorithm performs on-par with end-to-end training despite optimizing only the respective last layer in each iteration, and beats layer-wise random subspaces.
Example 7
We now provide more examples on a set of toy distributions. As before, we train layer-wise using OAS and randomly selected rotations, and end-to-end. Additionally, we train a mixed variant of OAS and end-to-end: New layers are still added one by one, but Algorithm 2 is modified such that iteration l optimizes all layers 1 through l in an end-to-end fashion. We call this training “progressive” as layers are progressively activated and never turned off again.
We obtain the following results: Optimal rotations always outperform random rotations in layer-wise training. With only a few layers, they also outperform end-to-end training, but are eventually overtaken as the network depth increases. Progressive training continues to be competitive also for deep networks.
Figure 6 shows the density estimates after twelve layers. At this point, none of the methods show a significant improvement by adding layers. Hyperparameters were optimized for each training configuration to obtain a fair comparison. Densities obtained by layer-wise training exhibit significant spurious structure for both optimal and random rotations, with an advantage for optimally chosen subspaces.
5 Conclusion
In this work, we showed that an isolated affine coupling learns the first two moments of the conditioned data distribution \(\mu (\mathrm {\mathbf {a}}|\mathrm {\mathbf {p}})\). Using this result, we derived a tight upper bound on the loss reduction that can be achieved by such a layer. We then used this to choose the best rotation of the coupling.
We regard our results as a first step towards a better understanding of deep affine flows. We provided sufficient conditions for a data set that can be exactly solved with layer-wise trained affine couplings and a single layer of D independent 1D flows.
Our results can be seen analogously to the classification layer at the end of a multi-layer classification network: The results from Sect. 3 directly apply to the last coupling in a deep normalizing flow. This raises a key question for future work: How do the first \(L-1\) layers prepare the distribution \(\mu _{L-1}\) such that the final layer can perfectly push the data to a Gaussian?
References
Ardizzone, L., Kruse, J., Rother, C., Köthe, U.: Analyzing inverse problems with invertible neural networks. In: International Conference on Learning Representations (2018)
Bigoni, D., Zahm, O., Spantini, A., Marzouk, Y.: Greedy inference with layers of lazy maps. arXiv preprint arXiv:1906.00031 (2019)
Dinh, L., Krueger, D., Bengio, Y.: NICE: non-linear independent components estimation. arXiv preprint arXiv:1410.8516 (2014)
Dinh, L., Sohl-Dickstein, J., Bengio, S.: Density estimation using real nvp. arXiv preprint arXiv:1605.08803 (2016)
Fleishman, A.I.: A method for simulating non-normal distributions. Psychometrika 43(4), 521–532 (1978)
Hoogeboom, E., Peters, J., van den Berg, R., Welling, M.: Integer discrete flows and lossless compression. In: Advances in Neural Information Processing Systems, pp. 12134–12144 (2019)
Jacobsen, J.H., Behrmann, J., Zemel, R., Bethge, M.: Excessive invariance causes adversarial vulnerability. arXiv preprint arXiv:1811.00401 (2018)
Jaini, P., Selby, K.A., Yu, Y.: Sum-of-squares polynomial flow. arXiv preprint arXiv:1905.02325 (2019)
Kingma, D.P., Dhariwal, P.: Glow: generative flow with invertible 1x1 convolutions. In: Advances in Neural Information Processing Systems, pp. 10215–10224 (2018)
Marzouk, Y., Moselhy, T., Parno, M., Spantini, A.: Sampling via measure transport: an introduction. In: Ghanem, R., Higdon, D., Owhadi, H. (eds.) Handbook of Uncertainty Quantification, pp. 1–41. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-11259-6_23-1
Meng, C., Ke, Y., Zhang, J., Zhang, M., Zhong, W., Ma, P.: Large-scale optimal transport map estimation using projection pursuit. In: Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 32, pp. 8116–8127. Curran Associates, Inc. (2019)
Nalisnick, E., Matsukawa, A., Teh, Y.W., Lakshminarayanan, B.: Detecting out-of-distribution inputs to deep generative models using a test for typicality. arXiv preprint arXiv:1906.02994 (2019)
Noé, F., Olsson, S., Köhler, J., Wu, H.: Boltzmann generators: sampling equilibrium states of many-body systems with deep learning. Science 365(6457), eaaw1147 (2019)
Papamakarios, G., Nalisnick, E., Rezende, D.J., Mohamed, S., Lakshminarayanan, B.: Normalizing flows for probabilistic modeling and inference. arXiv preprint arXiv:1912.02762 (2019)
Putzky, P., Welling, M.: Invert to learn to invert. In: Advances in Neural Information Processing Systems, pp. 446–456 (2019)
Tabak, E.G., Trigila, G.: Conditional expectation estimation through attributable components. Inf. Infer. J. IMA 7(4), 727–754 (2018)
Trigila, G., Tabak, E.G.: Data-driven optimal transport. Commun. Pure Appl. Math. 69(4), 613–648 (2016)
Acknowledgement
This work is supported by Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy EXC-2181/1 – 390900948 (the Heidelberg STRUCTURES Cluster of Excellence).
Furthermore, we thank our colleagues Lynton Ardizzone, Jakob Kruse, Jens Müller, and Peter Sorrenson for their help, support and fruitful discussions.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
1 Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Draxler, F., Schwarz, J., Schnörr, C., Köthe, U. (2021). Characterizing the Role of a Single Coupling Layer in Affine Normalizing Flows. In: Akata, Z., Geiger, A., Sattler, T. (eds) Pattern Recognition. DAGM GCPR 2020. Lecture Notes in Computer Science(), vol 12544. Springer, Cham. https://doi.org/10.1007/978-3-030-71278-5_1
Download citation
DOI: https://doi.org/10.1007/978-3-030-71278-5_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-71277-8
Online ISBN: 978-3-030-71278-5
eBook Packages: Computer ScienceComputer Science (R0)