Matrix Logarithm Parametrizations for Neural Network Covarious Models

Williams, Peter (1999) Matrix Logarithm Parametrizations for Neural Network Covarious Models. Neural Networks, 12 (2). pp. 299-308. ISSN 08936080

Full text not available from this repository.


Neural networks are commonly used to model conditional probability distributions. The idea is to represent distributional parameters as functions of conditioning events, where the function is determined by the architecture and weights of the network. An issue to be resolved is the link between distributional parameters and network outputs. The latter are unconstrained real numbers whereas distributional parameters may be required to lie in proper subsets, or be mutually constrained, e.g. by the positive definiteness requirement for a covariance matrix. The paper explores the matrix-logarithm parametrization of covariance matrices for multivariate normal distributions. From a Bayesian point of view the choice of parametrization is linked to the choice of prior. This is treated by investigating the invariance of predictive distributions, for the chosen parametrization, with respect to an important class of priors.

Item Type: Article
Schools and Departments: School of Engineering and Informatics > Informatics
Depositing User: EPrints Services
Date Deposited: 06 Feb 2012 19:52
Last Modified: 07 Jun 2012 13:24
📧 Request an update