Gaussian Processes

Current Understandings

GPs combine the flexibility of being capable of modelling arbitrary smooth functions if given enough data, with the simplicity of a Bayesian specification that only requires inference over a small number of readily interpretable hyperparameters (in contrast to deep neural networks) -- such as the length-scales by which the function varies along different dimensions, the contributions of signal and noise to the variance in the data, etc [Infinite Mixtures of Gaussian Process Experts, NIPS 2012].

As a probabilistic non-parametric framework, it can model middle-scale data very well. Of course, we can use a varieties of sparse (Neial Lawrence once argued that we should call it low-rank) approximation methods to enable it to handle millions of datapoints. But in that case, it does not seems capable of competing with deep neural networks.

results matching ""

    No results matching ""