Gaussian Transformations, Recursive Learning, and the Maximum Likelihood Estimation of Conditional Distribution Functions (joint with Sami Stouli, Bristol)

We propose an algorithm for machine learning of conditional distribution functions for a dependent variable (Y ) with continuous support. The algorithm produces a complete description of the conditional distribution function at all observed points in the covariate (X) space, and provides a similar estimate for other possible covariate values. The descriptions it provides are quite general and are globally valid conditional densities.The algorithm is multilayered and feed-forward. Each layer has the same statistical interpretation: Layer k takes a vector e(k-1) that is nearly perfectly marginally Gaussian and makes it more marginally Gaussian and more independent of X. It does this by applying a continuous monotonic transformation that varies depending on an observation’s X value. Each layer is estimated by an elastic net regularization of maximum likelihood. We demonstrate Wilks’ phenomenon for the composite algorithm and show how to calculate the algorithm’s effective dimension.

Please sign up for meetings here: docs.google.com/spreadsheets/d/1qPQqXivNYBDNJY_0OdHcZfjslHLu5UtSVQMd0LpETqc/edit#gid=0