<?xml version="1.0" encoding="utf-8" standalone="no"?><feed xmlns="http://www.w3.org/2005/Atom">
    <title type="text">The Grand Locus</title>
    <subtitle type="html">My name is Guillaume Filion. I am a scientist who loves biology and mathematics. As of late I also got into computers and the Internet. I intend my blog to be recreational, and not academic nor educational. I wish you will find some of the posts inspiring for your own reflection.</subtitle>
    <updated>2023-02-08T16:47:00Z</updated>
    <id>tag:blog.thegrandlocus.com,2023-02-08:atom.xml</id>
    <link href="http://blog.thegrandlocus.com/" hreflang="en" rel="alternate" type="text/html"/>
    <link href="http://blog.thegrandlocus.com/feed/atom.xml" rel="self" type="application/atom+xml"/>
    <link href="http://pubsubhubbub.appspot.com/" rel="hub"/>
    <rights>Copyright (c) 2023</rights>
    <generator uri="http://blog.thegrandlocus.com/" version="1.0">
        gui11aume's fork of Bloggart
    </generator>
    
    <entry>
        <title>On PhD advisory committees</title>
        <link href="http://blog.thegrandlocus.com/2023/02/on-phd-advisory-committees" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2023-02-08:post:176b58de4ab777d5329800d0f14b6ae8af4935e3</id>
	<updated>2023-02-08T16:46:20Z</updated>
	<published>2023-02-08T16:46:20Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;img src="http://blog.thegrandlocus.com/img/medicine.png" style="float:left; margin-right:20px;" title="Hygieia (detail of 'Medicine') by Gustav Klimt, 1900–1907" /&gt;


&lt;h3&gt;Joana, year 1&lt;/h3&gt;
&lt;p&gt;Joana is quite tense, she smiles a little too much. She brought some cookies for the committee, which I find very nice. I eat a cookie and thank her for the attention. I hope it helps her relax. A few minutes later, the chair of the committee thanks everybody for coming and goes through the protocol: Joana will give a forty minute presentation of her work, then we will discuss her project together, then she will leave the room so that we can speak with her adviser in private, and finally she will come back into the room without her adviser so that we can speak with her in private. “You can start whenever you are ready Joana” says the chair. Joana breathes deeply and she starts.
&lt;/p&gt;
&lt;p&gt;It is the middle of the presentation. Joana is still answering the question of the chair. Her answer does not make sense to me but I nod reassuringly. I am curious as to whether her adviser will correct her. I met him at a congress but I do not know his supervision style.
&lt;/p&gt;
&lt;p&gt;“What Joana meant to say is that...”
&lt;/p&gt;
&lt;p&gt;“A protective micromanager” I think to myself. This...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2023/02/on-phd-advisory-committees"&gt;On PhD advisory committees&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>Fisher information (with a cat)</title>
        <link href="http://blog.thegrandlocus.com/2022/12/fisher-information-with-a-cat" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2022-12-13:post:117c2892bd1279a053dd32263056ba461fe238b4</id>
	<updated>2022-12-13T15:02:56Z</updated>
	<published>2022-12-13T15:02:56Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;&lt;img src="http://blog.thegrandlocus.com/img/diffuser_cat_1.png" style="float:left; margin-right:20px; margin-bottom:20px;" title="Diffuser cat. licence: https://huggingface.co/spaces/CompVis/stable-diffusion-license"&gt;&lt;i&gt;It is still summer but the days are getting shorter (p &amp;lt; 0.05). Edgar and Sofia are playing chess, Immanuel purrs in a sofa next to them. Edgar has been holding his head for a while, thinking about his next move. Sofia starts:&lt;/i&gt;
&lt;/p&gt;
&lt;p&gt;“Something bothers me Immanuel. In the &lt;a href="http://blog.thegrandlocus.com/2021/11/a-gentle-introduction-to-the-cramer-rao-lower-bound-with-a-cat"&gt;last post&lt;/a&gt;, you told us that Fisher information could be defined as a variance, but that is not what I remember from my classes of mathematical statistics.”
&lt;/p&gt;
&lt;p&gt;“What do you remember, Sofia?”
&lt;/p&gt;
&lt;p&gt;“Our teacher said it was the curvature of the log-likelihood function around the maximum. More specifically, consider a parametric model $(f(X;\theta))$ where $(X)$ is a random variable and $(\theta)$ is a parameter. Say that the true (but unknown) value of the parameter is $(\theta^*)$. The first terms of the Taylor expansion of the log-likelihood $(\log f(X;\theta))$ around $(\theta^*)$ are
&lt;/p&gt;
&lt;p&gt;$$\log f(X;\theta^*) + (\theta - \theta^*) \cdot \frac{\partial}{\partial \theta} \log f(X;\theta^*)
   + \frac{1}{2}(\theta - \theta^*)^2 \cdot \frac{\partial^2}{\partial \theta^2} \log f(X;\theta^*).$$
&lt;/p&gt;
&lt;p&gt;Now compute the expected value and obtain the approximation below. We call it $(\varphi(\theta))$ to emphasize that it is...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2022/12/fisher-information-with-a-cat"&gt;Fisher information (with a cat)&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>Does science need statistical tests?</title>
        <link href="http://blog.thegrandlocus.com/2022/10/does-science-need-statistical-tests" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2022-10-14:post:c85b511d1a51e960cd5bd7f96009cedc18865eff</id>
	<updated>2022-10-14T09:22:06Z</updated>
	<published>2022-10-14T09:22:06Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;&lt;img src="http://blog.thegrandlocus.com/img/kandinsky_1923.png" style="float:left; margin-right:20px; margin-bottom:20px" title="Circles in a circle, Wassily Kandinsky (1293)" /&gt;Some time ago, my colleague &lt;a href="2014/09/mis-using-the-ks-test-for-p-hacking"&gt;John&lt;/a&gt; asked for help with the statistics for one of his manuscripts.
&lt;/p&gt;
&lt;p&gt;“We have this situation where we knocked out a gene with CRISPR and I want to test if it affects viability. I know that you are supposed to use a non-parametric test when the sample is small, but I have heard that you can still use the &lt;i&gt;t&lt;/i&gt; test if the variables are Gaussian. So now I am genuinely confused. Which test should I use?”
&lt;/p&gt;
&lt;p&gt;“I agree. It’s confusing. Why do you want to make a statistical test by the way?”
&lt;/p&gt;
&lt;p&gt;“Same as everyone. I want to know if the effect is significant. Plus, I’m hundred percent sure that the reviewers will ask for it.”
&lt;/p&gt;
&lt;p&gt;“I see. I will rephrase my question then. What decision do you have to make?”
&lt;/p&gt;
&lt;p&gt;“I can give you all the details of our experiments if you want, but I’m surprised. Nobody has ever asked me that before and I thought that experimental details do not really matter so much for a statistical test. So what kind of details do you need?”
&lt;/p&gt;
&lt;p&gt;“Nothing in particular. I just want to know whether you...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2022/10/does-science-need-statistical-tests"&gt;Does science need statistical tests?&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>Journal clubs, ranked from worst to best</title>
        <link href="http://blog.thegrandlocus.com/2022/03/journal-clubs-ranked-from-worst-to-best" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2022-03-07:post:9243354baa399973cfc2b50270ea638a99f9af4c</id>
	<updated>2022-03-07T20:35:16Z</updated>
	<published>2022-03-07T20:35:16Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;One of the most difficult tasks for an academic is to know the literature. Most labs run some kind of literature discussion, which are usually referred to as “journal clubs” in biology. We tested several variants when I created my lab in 2012 and we have learned a good deal about what works and what does not. So here is my personal view on the different types of journal clubs, from worst to best.
&lt;/p&gt;

&lt;h3&gt;5. Online discussion forum&lt;/h3&gt;
&lt;p&gt;Live meetings have several disadvantages: they disrupt the workflow of experiments, there are no records, and some people speak too much. I reasoned that a good way of addressing all this would be to put up an online forum where we could upload a paper and share our thoughts. I invited all the ~300 researchers of the institute to participate with the plan to go to social media if it gained momentum.
&lt;/p&gt;
&lt;img src="http://blog.thegrandlocus.com/img/book_transforming_itself.png" style="display:block; margin:auto;" title="Book transforming itself into nude woman, Salvador Dalí (1940)" /&gt;

&lt;p&gt;That did not work at all. We never got past the first paper, which had only one comment (mine). It was not appealing at all to formalize your thoughts in order to write a paragraph and post it for people who work next door. The written form is...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2022/03/journal-clubs-ranked-from-worst-to-best"&gt;Journal clubs, ranked from worst to best&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>A gentle introduction to the Cramér-Rao lower bound (with a cat)</title>
        <link href="http://blog.thegrandlocus.com/2021/11/a-gentle-introduction-to-the-cramer-rao-lower-bound-with-a-cat" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2021-11-22:post:fb7693166e9f8bd6ad96d061086ba6a5802f8f73</id>
	<updated>2021-11-22T16:22:01Z</updated>
	<published>2021-11-22T16:22:01Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;&lt;img src="http://blog.thegrandlocus.com/img/cat_in_a_hat.jpg" style="float:left; margin-right:20px; margin-bottom:10px;" title="Cat in a hat, by Rene Magritte (1929)" /&gt;It is summer, Edgard and Sofia are comfortably sitting on the terrace, watching the beautiful light of the end of the day. Edgard starts:
&lt;/p&gt;
&lt;p&gt;“Let’s play a game to see who is the better statistician! Immanuel my cat will give each of us a secret number strictly greater than zero. The other person will have to guess it.”
&lt;/p&gt;
&lt;p&gt;“How are we going to guess?”
&lt;/p&gt;
&lt;p&gt;“Let’s say that the secret numbers are the means of some Poisson variable. We generate samples at random. The one who gets the closest estimate by dinner time wins.”
&lt;/p&gt;
&lt;p&gt;“That sounds easy! Will Immanuel give us the same number?”
&lt;/p&gt;
&lt;p&gt;“What is the fun in that? Let’s ask him to give two different numbers. You know what to do. Just give me your first sample whenever you are ready and I will try to guess your secret number.”
&lt;/p&gt;
&lt;p&gt;Immanuel whispers something in the ear of Sofia and then does the same with Edgard. Sofia opens her laptop and after a few keystrokes she says “The first number I have for you is 1.”
&lt;/p&gt;
&lt;p&gt;“OK, I give up. You win.”
&lt;/p&gt;
&lt;p&gt;Sofia is puzzled at first, but then she notices how Immanuel is rolling...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2021/11/a-gentle-introduction-to-the-cramer-rao-lower-bound-with-a-cat"&gt;A gentle introduction to the Cramér-Rao lower bound (with a cat)&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>A tutorial on t-SNE (3)</title>
        <link href="http://blog.thegrandlocus.com/2021/09/a-tutorial-on-t-sne-3" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2021-09-22:post:e64e3aee75d326841f6ac01eb825bad9330856bb</id>
	<updated>2021-09-22T02:46:54Z</updated>
	<published>2021-09-22T02:46:54Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;This post is the third part of a tutorial on t-SNE. The &lt;a href="http://blog.thegrandlocus.com/2018/08/a-tutorial-on-t-sne-1"&gt;first part&lt;/a&gt; introduces dimensionality reduction and presents the main ideas of t-SNE. The &lt;a href="http://blog.thegrandlocus.com/2019/12/a-tutorial-on-t-sne-2"&gt;second part&lt;/a&gt; introduces the notion of perplexity. The present post covers the details of the nonlinear embedding.
&lt;/p&gt;

&lt;h3&gt;On the origins of t-SNE&lt;/h3&gt;
&lt;p&gt;If you are following the field of artificial intelligence, the name Geoffrey Hinton should sound familiar. As it turns out, the &lt;a href="https://en.wikipedia.org/wiki/Geoffrey_Hinton"&gt;“Godfather of Deep Learning”&lt;/a&gt; is the author of both &lt;a href="https://lvdmaaten.github.io/publications/papers/JMLR_2008.pdf"&gt;t-SNE&lt;/a&gt; and its ancestor &lt;a href="https://www.cs.toronto.edu/~fritz/absps/sne.pdf"&gt;SNE&lt;/a&gt;. This explains why t-SNE has a strong flavor of neural networks. If you already know &lt;a href="https://en.wikipedia.org/wiki/Gradient_descent"&gt;gradient-descent&lt;/a&gt; and &lt;a href="https://en.wikipedia.org/wiki/Variational_Bayesian_methods"&gt;variational learning&lt;/a&gt;, then you should feel at home. Otherwise no worries: we will keep it relatively simple and we will take the time to explain what happens under the hood.
&lt;/p&gt;
&lt;p&gt;We have seen &lt;a href="http://blog.thegrandlocus.com/2018/08/a-tutorial-on-t-sne-1"&gt;previously&lt;/a&gt; that t-SNE aims to preserve a relationship between the points, and that this relationship can be thought of as the probability of hopping from one point to the other in a random walk. The focus of this post is to explain what t-SNE does to preserve this relationship in a space of lower dimension.
&lt;/p&gt;

&lt;h3&gt;The Kullback-Leibler...&lt;/h3&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2021/09/a-tutorial-on-t-sne-3"&gt;A tutorial on t-SNE (3)&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>The most important quality of a scientist</title>
        <link href="http://blog.thegrandlocus.com/2020/09/the-most-important-quality-of-a-scientist" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2020-09-15:post:21f47c7d67f7e0e85918edd37ac61bd59c345a2d</id>
	<updated>2020-09-15T04:32:02Z</updated>
	<published>2020-09-15T04:32:02Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;&lt;img src="http://blog.thegrandlocus.com/img/portrait_of_Gerti_Schiele.png" style="float:left; margin-right:20px; margin-bottom:10px;" title="Portrait of Gerti Schiele, by Egon Schiele, 1909" /&gt;When I established my lab and started to recruit people, I thought that it would be interesting to gather some information about what makes a good or a bad scientist. To this end, I designed a short questionnaire with nine questions.  There was no right or wrong, nor even a preferred answer. Those were just questions to help me know the candidates better.
&lt;/p&gt;
&lt;p&gt;The first question was “What is the most important quality of a scientist?” I had no particular expectation. Actually, I did not even know my own answer to this question. As it turned out, most candidates answered that it was either &lt;em&gt;creativity&lt;/em&gt; or &lt;em&gt;persistence&lt;/em&gt;.
&lt;/p&gt;
&lt;p&gt;If you have been in science for even a short while, you know why this makes  sense. We have complicated problems to solve, so creativity and persistence are important. Yet, I was not convinced that a good scientist is someone who is either very creative or very persistent. The reason is that neither of these qualities &lt;em&gt;defines&lt;/em&gt; a scientist. Artists, politicians, business people, social workers and pretty much everyone else greatly benefits from being creative or persistent.
&lt;/p&gt;
&lt;p&gt;Having spent more time with scientists, I came to find the answer to my...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2020/09/the-most-important-quality-of-a-scientist"&gt;The most important quality of a scientist&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>Scientific models</title>
        <link href="http://blog.thegrandlocus.com/2020/01/scientific-models" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2020-01-01:post:67978a829d465b00dee09b9a049df5964ca83ff0</id>
	<updated>2020-01-01T20:07:24Z</updated>
	<published>2020-01-01T20:07:24Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;Literature discussions were usually very quiet in the laboratory, but somehow, this article had sparked a debate. Linda thought it was very bad. Albert liked it very much. Kate, the PI, was undecided. At some point the discussion stalled, so Kate made a move to wrap up.&lt;/p&gt;
&lt;p&gt;“So, Linda, why do you think the article is bad?”&lt;br/&gt;
“Because they are missing a thousand controls.”&lt;br/&gt;
“OK. Albert, why do you like this article?”&lt;br/&gt;
“I find their model in figure 6 really cool. Actually, if it is true, it…”&lt;br/&gt;
“Precisely my point!” interrupted Linda. “It’s pure speculation!”&lt;br/&gt;
Kate intervened.&lt;br/&gt;
“Albert, you describe figure 6 as a &lt;em&gt;model&lt;/em&gt;. What makes it a model?”&lt;br/&gt;
Albert spoke after a pause.&lt;br/&gt;
“It’s an idealized summary of...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2020/01/scientific-models"&gt;Scientific models&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>A tutorial on t-SNE (2)</title>
        <link href="http://blog.thegrandlocus.com/2019/12/a-tutorial-on-t-sne-2" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2019-12-16:post:cb88d56e8fa617298ede6dcdf004e5345ef4cd5d</id>
	<updated>2019-12-16T01:12:58Z</updated>
	<published>2019-12-16T01:12:58Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;In this post I explain what perplexity is and how it is used to parametrize t-SNE. This post is the second part of a tutorial on t-SNE. The &lt;a href="http://blog.thegrandlocus.com/2018/08/a-tutorial-on-t-sne-1"&gt;first part&lt;/a&gt; introduces dimensionality reduction and presents the main ideas of t-SNE. This is where you should start if you are not already familiar with t-SNE.
&lt;/p&gt;

&lt;h3&gt;What is perplexity?&lt;/h3&gt;
&lt;p&gt;Before you read on, pick a number at random between 1 and 10 and ask yourself whether I can guess it. It looks like my chances are 1 in 10 so you may think “no there is no way”. In fact, there is a &lt;a href="https://www.reddit.com/r/dataisbeautiful/comments/acow6y/asking_over_8500_students_to_pick_a_random_number/"&gt;28% chance&lt;/a&gt; that you chose the number 7, so my chances of guessing are higher than you may have thought initially.
   In this situation, the random variable is &lt;em&gt;somewhat predictable but not completely&lt;/em&gt;. How could we quantify that?
&lt;/p&gt;
&lt;p&gt;To answer the questions, let us count the possible samples from this distribution. We ask N people to choose a number at random between 1 and 10 and we record their answers (x1, x2, ..., xN). The number 1 shows up with probability p1 = 0.034 so the total in the sample is approximately...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2019/12/a-tutorial-on-t-sne-2"&gt;A tutorial on t-SNE (2)&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
    <entry>
        <title>Focus on: the Kullback-Leibler divergence</title>
        <link href="http://blog.thegrandlocus.com/2019/06/focus-on-the-kullback-leibler-divergence" rel="alternate" type="text/html"/>
        <id>tag:blog.thegrandlocus.com,2019-06-23:post:7d65b8316598344ef83f92c74b32988ab8884df4</id>
	<updated>2019-06-23T12:27:00Z</updated>
	<published>2019-06-23T12:27:00Z</published>
        <author>
            <name>Guillaume Filion</name>
            <uri>http://blog.thegrandlocus.com/</uri>
        </author>
        <content type="html">
          &lt;style type="text/css"&gt;
          code { display: inline; padding: 5px 0 0 0; }
          sup { vertical-align: super; }
          sub { vertical-align: sub; }
          &lt;/style&gt;
          &lt;p&gt;The story of the &lt;a href="https://en.wikipedia.org/wiki/Kullback%E2%80%93Leibler_divergence"&gt;Kullback-Leibler divergence&lt;/a&gt; starts in a top secret research facility. In 1951, right after the war, &lt;a href="https://en.wikipedia.org/wiki/Solomon_Kullback"&gt;Solomon Kullback&lt;/a&gt; and &lt;a href="https://en.wikipedia.org/wiki/Richard_Leibler"&gt;Richard Leibler&lt;/a&gt; were working as cryptanalysts for what would soon become the &lt;a href="https://en.wikipedia.org/wiki/National_Security_Agency#Formation"&gt;National Security Agency&lt;/a&gt;. Three years earlier, &lt;a href="https://en.wikipedia.org/wiki/Claude_Shannon"&gt;Claude Shannon&lt;/a&gt; had shaken the academic world by formulating the modern theory of information. Kullback and Leibler immediately saw how this could be useful in statistics and they came up with the concept of &lt;em&gt;information for discrimination&lt;/em&gt;, now known as &lt;em&gt;relative entropy&lt;/em&gt; or &lt;em&gt;Kullback-Leibler divergence&lt;/em&gt;.
&lt;/p&gt;
&lt;p&gt;The concept was introduced in an oringinal &lt;a href="https://projecteuclid.org/download/pdf_1/euclid.aoms/1177729694"&gt;article&lt;/a&gt;, and later expanded by Kullback in the book &lt;a href="https://books.google.es/books/about/Information_Theory_and_Statistics.html?id=05LwShwkhFYC&amp;redir_esc=y"&gt;Information Theory and Statistics&lt;/a&gt;. It has now found applications in most aspects of information technologies, and most prominently artificial neural networks. In this post, I want to give an advanced introduction on this concept, hoping to make it intuitive.
&lt;/p&gt;

&lt;h3&gt;Discriminating information&lt;/h3&gt;
&lt;p&gt;The original motivation given by Kullback and Leibler is still the best way to expose the main idea, so let us follow their rationale. Suppose that we hesitate between two competing hypotheses $(H_1)$ and $(H_2)$. To make things more concrete, say that we have an encrypted message $(x)$ that may come from two possible...&lt;/p&gt;
          &lt;br /&gt;
	  Read more on the blog: &lt;a href="http://blog.thegrandlocus.com/2019/06/focus-on-the-kullback-leibler-divergence"&gt;Focus on: the Kullback-Leibler divergence&lt;/a&gt;
          &lt;br /&gt;
        </content>
    </entry>
    
</feed>