For a long time, we were training machine learning models by optimizing surrogate functions or log-likelihood rather than evaluation functions. For example, while the evaluation is the 0-1 loss, the objective function
Link: Authors: Marco Tulio Ribeiro, Sameer Singh, Carlos Guestrin I have been busy with finals for the past few days so this post is to make up for last week.
Link: Authors: William Yang Wang, Yashar Mehdad, Dragomir R. Radev, Amanda Stent Summary The paper tackles the problem of timeline summarization - extracting milestones of a news story and arranging
Link: Authors: Lu Wang, Wang Ling Summary Generate opinion summary using an encoder-decoder neural model. The problem can be formulated as given an input, which contains a set of reviews/arguments, construct a
I came across a very interesting paper by Fraenkel and Schul on how negative phrases are constructed. It was a delighting moment when I realized that "these things are so true
Suppose we are evaluating a class of predictors on a data distribution $D$. This distribution generates data pairs $(x, y)$ where $y = f(x) + \epsilon$, $\epsilon$ is a zero-mean
This post relates to "Gibbs sampling for the uninitiated" (Resnik and Hardisty, 2010), a very helpful material for those who just began to learn about Bayesian inference (like me). The