Searched for: author%3A%22Batselier%2C+K.%22
(1 - 7 of 7)
document
Batselier, K. (author), Ko, Ching Yun (author), Wong, Ngai (author)
This article reformulates the multiple-input-multiple-output Volterra system identification problem as an extended Kalman filtering problem. This reformulation has two advantages. First, it results in a simplification of the solution compared to the Tensor Network Kalman filter as no tensor filtering equations are required anymore. The second...
conference paper 2019
document
Chen, Cong (author), Batselier, K. (author), Ko, Ching Yun (author), Wong, Ngai (author)
A restricted Boltzmann machine (RBM) learns a probability distribution over its input samples and has numerous uses like dimensionality reduction, classification and generative modeling. Conventional RBMs accept vectorized data that dismiss potentially important structural information in the original tensor (multi-way) input. Matrix-variate...
conference paper 2019
document
Chen, Cong (author), Batselier, K. (author), Ko, Ching Yun (author), Wong, Ngai (author)
There has been growing interest in extending traditional vector-based machine learning techniques to their tensor forms. Support tensor machine (STM) and support Tucker machine (STuM) are two typical tensor generalization of the conventional support vector machine (SVM). However, the expressive power of STM is restrictive due to its rank-one...
conference paper 2019
document
Ko, Ching Yun (author), Batselier, K. (author), Daniel, Luca (author), Yu, Wenjian (author), Wong, Ngai (author)
We propose a new tensor completion method based on tensor trains. The to-be-completed tensor is modeled as a low-rank tensor train, where we use the known tensor entries and their coordinates to update the tensor train. A novel tensor train initialization procedure is proposed specifically for image and video completion, which is demonstrated...
journal article 2020
document
Ko, Ching Yun (author), Chen, Cong (author), He, Zhuolun (author), Zhang, Yuke (author), Batselier, K. (author), Wong, Ngai (author)
Sum-product networks (SPNs) constitute an emerging class of neural networks with clear probabilistic semantics and superior inference speed over other graphical models. This brief reveals an important connection between SPNs and tensor trains (TTs), leading to a new canonical form which we call tensor SPNs (tSPNs). Specifically, we...
journal article 2020
document
Batselier, K. (author), Cichocki, Andrzej (author), Wong, Ngai (author)
In this article, two new algorithms are presented that convert a given data tensor train into either a Tucker decomposition with orthogonal matrix factors or a multi-scale entanglement renormalization ansatz (MERA). The Tucker core tensor is never explicitly computed but stored as a tensor train instead, resulting in both computationally and...
journal article 2021
document
Chen, Cong (author), Batselier, K. (author), Yu, Wenjian (author), Wong, Ngai (author)
Tensor, a multi-dimensional data structure, has been exploited recently in the machine learning community. Traditional machine learning approaches are vector- or matrix-based, and cannot handle tensorial data directly. In this paper, we propose a tensor train (TT)-based kernel technique for the first time, and apply it to the conventional...
journal article 2022
Searched for: author%3A%22Batselier%2C+K.%22
(1 - 7 of 7)