Profile cover photo
Profile photo
Jedd Haberstro
148 followers
148 followers
About
Posts

Post has attachment
Learning vs Inference
This will be a short post. In machine learning, the terms "learning" and "inference" are used often and and it's not always clear what is meant. For example, is "variational inference" and neural network "inferencing" the same thing? Usually not! When the d...
Add a comment...

Post has attachment
Why is the ELBO difficult to optimize?
The task of bayesian inference is to compute the posterior p(z | x) of the model p(x, z) = p(x|z)p(z) where z is a latent variable. This is often intractable, so a bayesian may resort to approximating it with some easier distribution q(z) — this method is c...
Add a comment...

Post has attachment
Expectation Maximization vs Variational Bayes
I constantly find myself forgetting the details of the EM algorithm, variational bayes, and what exactly the difference is between the two. To avoid confusion in the future, I wrote the following note. Q: What is the difference between EM and Variational Ba...
Add a comment...

Post has attachment
Frequentists vs Bayesians
A really wonderful aspect of learning about machine learning is that you can't help but learn about the field statistics as well. As a computer scientist (or really, a software engineer -- I have a hard time calling myself a computer scientist), one of the ...
Add a comment...

Post has attachment
An intuition for Newton's method
During my lazy weekend afternoons (and all the other days of the week) I've been going through Nando de Freitas' undergraduate machine learning course on youtube . In lecture 26 he introduces gradient descent, an iterative algorithm for optimizing (potentia...
Add a comment...

Post has attachment
On optimizing high-dimensional non-convex functions
Excuse me for the (in)completeness of this post. What follows is merely a thought, inspired by two independent statements, about a domain of science (or math, really) with which I am barely initiated. Let me give you these two statements first. In the book ...
Add a comment...

Post has attachment
TIL: premultiplied-alpha colors
Alpha is the measure of how translucent an object is. An alpha of 0.0 means the object is entirely transparent, an alpha of 1.0 means the object is entirely opaque, and an alpha in the middle means a fraction of the total light may passthrough the object. T...
Add a comment...

Post has attachment
TIL: The column space and null space
A vector space is a set of vectors that is closed under addition and scalar multiplication. In other words, given two vectors, v  and w ,  a vector space is formed by the set of all linear combinations formed between  v  and w , namely  c v  + d w  for arbi...
Add a comment...

Post has attachment
TIL: a principled approach to dynamic programming
Dynamic programming has always been a topic I understood at a surface level ( it's just memoization, right?!) , but ultimately feared for lack of real-world experience solving such problems. I read today a superb explanation of a principled approach to solv...
Add a comment...

Post has attachment
TIL: Alpha-to-coverage for order-independent transparency
Alpha-to-coverage is a computer graphics technique, supported by most (all?) modern GPUs, for rendering translucent multi-sampled anti-aliased primitives. Given an alpha value in the range of [0.0, 1.0] and N  samples of color stored per pixel, the alpha ch...
Add a comment...
Wait while more posts are being loaded