I am a third-year research student working with Arthur Gretton at Gatsby Computational Neuroscience Unit, UCL. I received M.Eng. from Tokyo Institute of Technology where I worked with Masashi Sugiyama on supervised feature selection using squared-loss mutual information. Before that I was a research assistant working with Thanaruk Theeramunkong on a Thai news relations discovery project. I received B.Sc. in Computer Science from SIIT, Thammasat university, Thailand.
Contact: Wittawat Jitkrittum (วิทวัส จิตกฤตธรรม) ( )
My works are listed on this page. I occasionally update my blog summarizing what I learn. Some photos I have taken are on Flickr. I also maintain a web site for Gatsby's machine learning journal club. The RSS feed url for the blog posts is here.
24 May 2016. Interpretable Distribution Features with Maximum Testing Power: a linear-time nonparametric two-sample test which returns a set of local features indicating why the null hypothesis is rejected. Python code available on Github.
Dec 2015. K2-ABC: Approximate Bayesian Computation with Infinite Dimensional Summary Statistics via Kernel Embeddings: summary statistic free approximate Bayesian computation with kernel embeddings. Accepted to AISTATS 2016.
13 Nov 2015. We released the code for Locally Linear Latent Variable Model (LL-LVM), which was accepted to NIPS 2015. Check our Matlab code here on Github.
8 Jun 2015. Bayesian Manifold Learning: Locally Linear Latent Variable Model (LL-LVM): a probabilistic model for non-linear manifold discovery.
9 Mar 2015. Kernel-Based Just-In-Time Learning for Passing Expectation Propagation Messages: a fast, online algorithm for nonparametric learning of EP message updates. Source code available here.