In generative modeling, the Wasserstein distance (WD) has emerged as a useful metric to measure the discrepancy between generated and real data distributions. Unfortunately, it is challenging to approximate the WD of high-dimensional distributions. In contrast, the sliced Wasserstein distance (SWD) factorizes high-dimensional distributions into their multiple one-dimensional marginal distributions and is thus easier to approximate. In this paper, we introduce novel approximations of the primal and dual SWD. Instead of using a large number of random projections, as it is done by conventional SWD approximation methods, we propose to approximate SWDs with a small number of parameterized orthogonal projections in an end-to-end deep learning fashion. As concrete applications of our SWD approximations, we design two types of differentiable SWD blocks to equip modern generative frameworks---Auto-Encoders (AE) and Generative Adversarial Networks (GAN). In the experiments, we not only show the superiority of the proposed generative models on standard image synthesis benchmarks, but also demonstrate the state-of-the-art performance on challenging high resolution image and video generation in an unsupervised manner.
We study the perturbation bound for the spectral radius of an m th-order n -dimensional nonnegative tensor A. The main contribution of this paper is to show that when A is perturbed to a nonnegative tensor A~ by ΔA, the absolute difference between the spectral radii of A and A~ is bounded by the largest magnitude of the ratio of the i th component of ΔAxm-1 and the i th component xm-1, where x is an eigenvector associated with the largest eigenvalue of A in magnitude and its entries are positive. We further derive the bound in terms of the entries of A only when x is not known in advance. Based on the perturbation analysis, we make use of the NQZ algorithm to estimate the spectral radius of a nonnegative tensor in general. On the other hand, we study the backward error matrix ΔA and obtain its smallest error bound for its perturbed largest eigenvalue and associated eigenvector of an irreducible nonnegative tensor. Based on the backward error analysis, we can estimate the stability of computation of the largest eigenvalue of an irreducible nonnegative tensor by the NQZ algorithm. Numerical examples are presented to illustrate the theoretical results of our perturbation analysis.
How to handle large multidimensional datasets, such as hyperspectral images and video information, efficiently and effectively plays a critical role in big-data processing. The characteristics of low-rank tensor decomposition in recent years demonstrate the essentials in describing the tensor rank, which often leads to promising approaches. However, most current tensor decomposition models consider the rank-1 component simply to be the vector outer product, which may not fully capture the correlated spatial information effectively for large-scale and high-order multidimensional datasets. In this article, we develop a new novel tensor decomposition model by extending it to the matrix outer product or called Bhattacharya-Mesner product, to form an effective dataset decomposition. The fundamental idea is to decompose tensors structurally in a compact manner as much as possible while retaining data spatial characteristics in a tractable way. By incorporating the framework of the Bayesian inference, a new tensor decomposition model on the subtle matrix unfolding outer product is established for both tensor completion and robust principal component analysis problems, including hyperspectral image completion and denoising, traffic data imputation, and video background subtraction. Numerical experiments on real-world datasets demonstrate the highly desirable effectiveness of the proposed approach.
Summary In this paper, we propose several relaxation algorithms for solving the tensor equation arising from the higher‐order Markov chain and the multilinear PageRank. The semi‐symmetrization technique on the original equation is also employed to modify the proposed algorithms. The convergence analysis is given for the proposed algorithms. It is shown that the new algorithms are more efficient than the existing ones by some numerical experiments when relaxation parameters are chosen suitably.