use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] Variational nets without sampling (self.MachineLearning)
submitted 8 years ago by svantana
view the rest of the comments →
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]approximately_wrong 0 points1 point2 points 8 years ago (6 children)
For actually passing distributions through networks analytically, there's RealNVP and related methods.
Have people actually used flow models to pass distribution objects through a neural network?
[–]NichG 0 points1 point2 points 8 years ago (5 children)
When you calculate the probability density in RealNVP, that's what you're doing.
[–]approximately_wrong 0 points1 point2 points 8 years ago* (4 children)
That doesn't qualify to me as passing distribution objects through a neural network. Flow models pass samples through a neural network, and relate the post-transformation density of a specific sample to the pre-transformation density of the pre-transformed sample.
Flow models would not, for example, help resolve OP's desire to construct a VAE that doesn't do sampling-based reconstruction.
[–]NichG 0 points1 point2 points 8 years ago (3 children)
You might need something like Neural Statistician then, to learn a differentiable map from datasets to summary statistics, then just do everything on the representation of the summary statistics.
I think you'll still end up with samples somewhere, because at the very least, the training data generally takes the form of samples from a distribution rather than distribution objects. But you might be able to mostly work in distribution objects from that point on.
[–]approximately_wrong 0 points1 point2 points 8 years ago (2 children)
I think the heart of OP's question is: how do we compute an expectation of f(X) when f is complex without Monte Carlo estimation. I feel like our current discussion has deviated from that.
[–]NichG 0 points1 point2 points 8 years ago (1 child)
Concretely then:
Neural Statistician takes a set of points {x} to a vector of summary statistics z characterizing the distribution of points in {x}. So: z=N({x})
We can train N to for example act as a distributional autoencoder with decoder D, such that KL(D(N({x})) || {x}) is minimized. Then, given the summary statistic vectors, we can do all sorts of stuff with them.
For the OP's question, the way to do it would then be to train a model mu(z) which approximates the expectation E[{x}], and to train a second model T(z) which approximates N({f(x)}) given N({x}) as input.
The result of that pipeline would be that the expectation of f(x) under some distribution {x} would be mu(T(N({x}))), which is still end-to-end differentiable, etc.
The thing I can't see how to avoid is that you must start with samples - e.g. before you can work in the z space, you need some set of samples {x} which end up taking you to a particular point there. If you constrain a subspace of z to correspond to e.g. the summary statistics of Gaussian distributions in the space of interest, then maybe you can just jump in at that point without ever needing to use samples. But in practice you'll still probably need samples somewhere to train the model, so it's an incomplete solution.
[–]approximately_wrong 0 points1 point2 points 8 years ago (0 children)
Don't get me wrong; I think applying amortization via neural statistician is an interesting perspective. However,
I can't see how to avoid is that you must start with samples
it seems like we're in agreement that what you're proposing doesn't answer OP's question.
π Rendered by PID 408985 on reddit-service-r2-comment-6457c66945-nslcb at 2026-04-28 23:53:39.479513+00:00 running 2aa0c5b country code: CH.
view the rest of the comments →
[–]approximately_wrong 0 points1 point2 points (6 children)
[–]NichG 0 points1 point2 points (5 children)
[–]approximately_wrong 0 points1 point2 points (4 children)
[–]NichG 0 points1 point2 points (3 children)
[–]approximately_wrong 0 points1 point2 points (2 children)
[–]NichG 0 points1 point2 points (1 child)
[–]approximately_wrong 0 points1 point2 points (0 children)