05/10/2017 – Mutual Information
After discussing the source and uses of Mutual Information, Erik-Jan introduced the concept of Maximal Information Criterion (MIC): a relatively new measure of dependence designed to detect both linear and nonlinear dependencies between two variables.
We held a competition to find a simple (maximum 2 operations) functional relationship that the Maximal Information Coefficient could not find given 300 data points. This competition was won by Daniel using the function f(x) = 9 * (x < 9), which yielded an MIC of 0.
Photo: Everyone hard at work finding functions of data which are hard to detect.
Discussing the results of our competition gave us new insights into how the MIC determines whether a pattern can be considered a relevant dependence, and how the choice of N = 300 limits its ability to detect complex functional relations!
If you want to view the presentation at your own leisure, click here (PDF).
In this meeting, Erik-Jan will present on mutual information, leading up to a non-parametric measure of association that performs well in many different situations & functional forms. Here is a nice associated image to pique your interest:
Interactive component: We will have a competition to try to break down on this measure! The winner will receive a prize. Bring your laptop (with R)!
Preparation: none 🙂
You must be logged in to post a comment.