<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://nemenmanlab.org/~ilya/index.php?action=history&amp;feed=atom&amp;title=Entropy_Estimation</id>
	<title>Entropy Estimation - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://nemenmanlab.org/~ilya/index.php?action=history&amp;feed=atom&amp;title=Entropy_Estimation"/>
	<link rel="alternate" type="text/html" href="https://nemenmanlab.org/~ilya/index.php?title=Entropy_Estimation&amp;action=history"/>
	<updated>2026-05-17T08:38:10Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.31.0</generator>
	<entry>
		<id>https://nemenmanlab.org/~ilya/index.php?title=Entropy_Estimation&amp;diff=193&amp;oldid=prev</id>
		<title>Ilya: 1 revision imported</title>
		<link rel="alternate" type="text/html" href="https://nemenmanlab.org/~ilya/index.php?title=Entropy_Estimation&amp;diff=193&amp;oldid=prev"/>
		<updated>2018-07-04T16:28:39Z</updated>

		<summary type="html">&lt;p&gt;1 revision imported&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;1&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;1&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 16:28, 4 July 2018&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-notice&quot; lang=&quot;en&quot;&gt;&lt;div class=&quot;mw-diff-empty&quot;&gt;(No difference)&lt;/div&gt;
&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;</summary>
		<author><name>Ilya</name></author>
		
	</entry>
	<entry>
		<id>https://nemenmanlab.org/~ilya/index.php?title=Entropy_Estimation&amp;diff=192&amp;oldid=prev</id>
		<title>nemenman&gt;Ilya: /* Differences of IT quantities */</title>
		<link rel="alternate" type="text/html" href="https://nemenmanlab.org/~ilya/index.php?title=Entropy_Estimation&amp;diff=192&amp;oldid=prev"/>
		<updated>2012-01-14T22:15:54Z</updated>

		<summary type="html">&lt;p&gt;‎&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Differences of IT quantities&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{PROJECTS}}&lt;br /&gt;
&lt;br /&gt;
Last updated on 14 January 2012.&lt;br /&gt;
&lt;br /&gt;
==What is this about?==&lt;br /&gt;
&lt;br /&gt;
Information theoretic (IT) methods are now routinely used for natural sciences data analysis. For example, in neuroscience one uses IT tools to understand mechanisms by which neurons encode sensory information, while in molecular systems biology the same techniques help to uncover statistical associations among concentrations of various molecules. The strength of the methods comes from the very limited assumptions that go into them. For example, it is appealing to quantify ''all'' statistical dependencies among variables, but not just linear ones. However, the widespread use of the IT methods has been hindered by a major difficulty: it is hard to estimate entropy from experimental datasets of realistic sizes. It turns out that estimators are typically biased. For example, for estimation of entropies of discrete data, most common methods result in a bias of &amp;lt;math&amp;gt;\propto 2^S/N&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is the (unknown) entropy and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; is the data set size (see this [http://www.menem.com/~ilya/pages/NIPS03/nemenman.pdf review] for details). In the recent years, many methods (some described [http://www.menem.com/~ilya/pages/NIPS03/index.html here]) have been developed to solve the problem, but the majority only works when &amp;lt;math&amp;gt;2^S/N\ll 1&amp;lt;/math&amp;gt; (for discrete variables), or requires strong assumptions about the smoothness of the underlying probability density (for continuous variables).&lt;br /&gt;
&lt;br /&gt;
;[[image:entropy.jpg|frameless|360px|left|Completion time|link=Bel et al., 2010]]&lt;br /&gt;
&lt;br /&gt;
Our goal has been to solve this problem, at least under certain weak assumptions. We've made some progress. In particular, for severely undersampled data, our NSB method is probably still the most powerful, ten years after we proposed it.  For example, the plot to the left show the convergence of various entropy estimators to the true entropy value when the data set size grows. The NSB method clearly converges better.&lt;br /&gt;
&lt;br /&gt;
==Results==&lt;br /&gt;
===NSB entropy estimator===&lt;br /&gt;
Our [http://nsb-entropy.sf.net NSB entropy estimator] reduced the entropy estimation bias to &amp;lt;math&amp;gt;2^{S/2}/N\ll 1&amp;lt;/math&amp;gt; for a large class of data.&lt;br /&gt;
The method doesn't work always, but it is possible to diagnose when it fails and, even in these cases, it performs not worse than most traditional estimators.&lt;br /&gt;
&lt;br /&gt;
If you are interested in the NSB method, the following link will be of use:&lt;br /&gt;
*SourceForge-hosted project that implements the NSB method in C++ and Octave/MatLab, http://nsb-entropy.sf.net .&lt;br /&gt;
*Original paper that introduced the method, [[Nemenman et al., 2002]].&lt;br /&gt;
*[[Nemenman, 2011b]] elucidates the coincidence-counting nature of the estimator and analyzes a lot of its technical properties.&lt;br /&gt;
*The following papers have applies the estimator for different natural datasets: [[Nemenman et al., 2004]], [[Nemenman et al., 2008]].&lt;br /&gt;
*You can also see some additional discussion of [[entropy estimation methods]].&lt;br /&gt;
&lt;br /&gt;
===Differences of information quantities===&lt;br /&gt;
In many cases, in particular when IT quantities are [[Reverse engineering cellular networks |used as a measure of statistical interactions]], the precise values of these quantities are less important than their rankings (i.e., the sign of their differences). In the case of mutual information, we showed in [[Margolin et al., 2006a]] that  the latter task is much easier than the former, and it can be completed with almost any estimator with very little prior tuning.&lt;/div&gt;</summary>
		<author><name>nemenman&gt;Ilya</name></author>
		
	</entry>
</feed>