<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://nemenmanlab.org/~ilya/index.php?action=history&amp;feed=atom&amp;title=Physics_380%2C_2011%3A_Lecture_13</id>
	<title>Physics 380, 2011: Lecture 13 - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://nemenmanlab.org/~ilya/index.php?action=history&amp;feed=atom&amp;title=Physics_380%2C_2011%3A_Lecture_13"/>
	<link rel="alternate" type="text/html" href="https://nemenmanlab.org/~ilya/index.php?title=Physics_380,_2011:_Lecture_13&amp;action=history"/>
	<updated>2026-05-17T09:39:58Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.31.0</generator>
	<entry>
		<id>https://nemenmanlab.org/~ilya/index.php?title=Physics_380,_2011:_Lecture_13&amp;diff=329&amp;oldid=prev</id>
		<title>Ilya: 1 revision imported</title>
		<link rel="alternate" type="text/html" href="https://nemenmanlab.org/~ilya/index.php?title=Physics_380,_2011:_Lecture_13&amp;diff=329&amp;oldid=prev"/>
		<updated>2018-07-04T16:28:41Z</updated>

		<summary type="html">&lt;p&gt;1 revision imported&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;1&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;1&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 16:28, 4 July 2018&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-notice&quot; lang=&quot;en&quot;&gt;&lt;div class=&quot;mw-diff-empty&quot;&gt;(No difference)&lt;/div&gt;
&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;</summary>
		<author><name>Ilya</name></author>
		
	</entry>
	<entry>
		<id>https://nemenmanlab.org/~ilya/index.php?title=Physics_380,_2011:_Lecture_13&amp;diff=328&amp;oldid=prev</id>
		<title>nemenman&gt;Ilya: /* Main lecture */</title>
		<link rel="alternate" type="text/html" href="https://nemenmanlab.org/~ilya/index.php?title=Physics_380,_2011:_Lecture_13&amp;diff=328&amp;oldid=prev"/>
		<updated>2011-11-03T13:40:31Z</updated>

		<summary type="html">&lt;p&gt;‎&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Main lecture&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{PHYS380-2011}}&lt;br /&gt;
This is the last lecture in the information theory block.&lt;br /&gt;
&lt;br /&gt;
==In-class presentation==&lt;br /&gt;
Xiang Cheng presents the article by Pedraza and van Oudenaarden, 2005.&lt;br /&gt;
&lt;br /&gt;
==Warmup question==&lt;br /&gt;
I am showing you a diagram of cellular information processing pathways of certain kind (MAPK pathways -- more on them later) courtesy of Jim Faeder (U Pittsburgh). We know how to measure the amount of information that travels along these pathways -- need to estimate mutual information between the quantities of interest. But can information-theoretic ideas also help us understand which of these pathways contribute more to the information processing than the others and, therefore, need first attention?&lt;br /&gt;
&lt;br /&gt;
==Main lecture==&lt;br /&gt;
#Information theory provides a measure for characterization of quality of input output relations. But in addition, due to the data processing inequality, it also provides ways of unambiguously reducing dimensionality of the modeled biological system.&lt;br /&gt;
#*Indeed, say we have a large-dimensional signal &amp;lt;math&amp;gt;\vec{s}&amp;lt;/math&amp;gt; and response &amp;lt;math&amp;gt;\vec{r}&amp;lt;/math&amp;gt;. There's a certain mutual information between these &amp;lt;math&amp;gt;I[\vec{s};\vec{r}]=I_0&amp;lt;/math&amp;gt;. If we propose a reduction of the signal and response to &amp;lt;math&amp;gt;f=f(\vec{s}),g=g(\vec{r})&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;I[f;g]\le I[\vec{s};\vec{r}]&amp;lt;/math&amp;gt; by the data processing inequality.&lt;br /&gt;
#We can, for example, solve the problem like: Which inputs are informative of the outputs (and hence need to be accounted for in a model)? We omit different subsets of the inputs &amp;lt;math&amp;gt;s_i&amp;lt;/math&amp;gt;, calculate &amp;lt;math&amp;gt;I[\vec{s}-s_i;\vec{r}]=I_{\not i}&amp;lt;/math&amp;gt;, and calculate the error due to omitting this signal &amp;lt;math&amp;gt;\Delta_i=\frac{I_0-I_{\not i}}{I_0}&amp;lt;/math&amp;gt;. Those components that have a small &amp;lt;math&amp;gt;\Delta_{\not i}&amp;lt;/math&amp;gt; can be safely neglected. This type of analysis can be used, for example, to understand which features of the neural code are important. In the subsequent presentation by Farhan, we will hear about using this trick to understand if high precision of neural spikes is important or not.&lt;br /&gt;
#We have discussed the problem of lossless coding earlier in the class. What if one is willing to transmit the message with errors, but still wants to reconstruct the message with a small loss. This is Shannon's rate distortion or lossy coding theorem. A good place to look this up is the book by Cover and Thomas.&lt;br /&gt;
#*Suppose there's a loss function &amp;lt;math&amp;gt;d(x,x')\ge0&amp;lt;/math&amp;gt; for recovering a value of signal &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; when, in fact, it should have been &amp;lt;/math&amp;gt;x&amp;lt;/math&amp;gt;. &lt;br /&gt;
#*One can encode the signal as &amp;lt;math&amp;gt;P(x'|x)&amp;lt;/math&amp;gt;, and the amount of bits one would need to store this encoding would be &amp;lt;math&amp;gt;I[x;x']&amp;lt;/math&amp;gt;.&lt;br /&gt;
#*The average loss experienced would be &amp;lt;math&amp;gt;&amp;lt;d(x,x')&amp;gt;=\int dx\, dx' P(x)P(x'|x) d(x,x')&amp;lt;/math&amp;gt;.&lt;br /&gt;
#* We are interested in the shortest, the most compressed encoding, &amp;lt;math&amp;gt;I[x;x']\to0&amp;lt;/math&amp;gt;, and yet we want to have the smallest loss as well &amp;lt;math&amp;gt;&amp;lt;d(x,x')&amp;gt;\to0&amp;lt;/math&amp;gt;. We can choose then to minimize &amp;lt;math&amp;gt;L=I[x;x']+\lambda &amp;lt;d(x,x')&amp;gt;&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt; is an arbitrary constant that controls how much we value compression over quality (think different bitrates in the mp3 coding).&lt;br /&gt;
#*We minimize &amp;lt;math&amp;gt;L&amp;lt;/math&amp;gt; over all &amp;lt;math&amp;gt;P(x'|x)&amp;lt;/math&amp;gt;. This can only be done numerically, but there are effective algorithms (namely, Blahur-Arimoto algorithm described in the Cover and Thomas textbook) to do this.&lt;br /&gt;
#In some cases, it is unclear how to define &amp;lt;math&amp;gt;d&amp;lt;/math&amp;gt;. But maybe instead there's the following setup. &lt;br /&gt;
#*We observe &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, and compress it to &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;P(x'|x)&amp;lt;/math&amp;gt;. However, we really care not about &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; but about some other ''relevant'' variable &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, given by &amp;lt;math&amp;gt;P(y|x)&amp;lt;/math&amp;gt;.&lt;br /&gt;
#*With compression, &amp;lt;math&amp;gt;P(y|x')=\sum_x P(y|x)P(x|x')=\sum_x P(y|x)\frac{P(x,x')}{P(x')}=\sum_x P(y|x)\frac{P(x'|x)P(x)}{\sum_x P(x'|x)P(x)}&amp;lt;/math&amp;gt;.&lt;br /&gt;
#*We can now maximize the information the compressed variable has about the relevant variable, while maximizing the compression itself. That is, we want to maximize &amp;lt;math&amp;gt;L=I[x';y]-\lambda I[x;x']&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt; is again a control parameter.&lt;br /&gt;
#*This should also be done numerically with the same Blahut-Arimoto algorithm. This approach is know as the Information Bottleneck method (see Tishby et al., 2000).&lt;/div&gt;</summary>
		<author><name>nemenman&gt;Ilya</name></author>
		
	</entry>
</feed>