Talk:Variable kernel density estimation

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia

There seems to be a factor 1/h missing in the first section called Rationale. Compare here: http://en.wikipedia.org/wiki/Kernel_density_estimation#Definition

Good call. In my own paper, the h is a sigma and it is absorbed into K while the K is not normalized and there is a separate normalization coefficient. However, I decided in this article to pull out the bandwidth, change the symbol and normalize K to bring my notation more in line with others, not realizing that you also have to pull h out of the normalization coefficient. Thanks. Peteymills (talk) 23:16, 13 May 2013 (UTC)[reply]

nested kernel estimators :[edit]

nested kernel estimators = multilayer perceptron (or feedforward neural networks)

the "multi-variate kernel density estimation" is the special case where the kernel's parameter depends on space directions, to be general in fact the kernel could depend on the point of space, where it's parameters are given by another "kernel density estimator",

this way each layer (each kernel density estimator) is the layer of a (feed-forward) neural network,

and we also include hidden markov models (for discrete random variables)

78.227.78.135 (talk) 01:10, 30 October 2015 (UTC)[reply]

Kernel shape variation[edit]

"For multivariate estimators, the parameter, h, can be generalized to vary not just the size, but also the shape of the kernel. This more complicated approach will not be covered here."

It would be helpful if the article either covered the part about shape variation, or provided a reference on where to find more information about that.

2001:16B8:A07F:C600:D80B:18C9:FA99:B492 (talk) 07:39, 13 May 2020 (UTC)[reply]