<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Some stuff &#187; mit</title>
	<atom:link href="http://blog.yhuang.org/?feed=rss2&#038;tag=mit" rel="self" type="application/rss+xml" />
	<link>https://blog.yhuang.org</link>
	<description>here.</description>
	<lastBuildDate>Wed, 27 Aug 2025 08:50:58 +0000</lastBuildDate>
	<language>en</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=3.1.1</generator>
		<item>
		<title>language context, ambiguity and inference</title>
		<link>https://blog.yhuang.org/?p=788</link>
		<comments>https://blog.yhuang.org/?p=788#comments</comments>
		<pubDate>Thu, 19 Jan 2012 23:49:49 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[ambiguity]]></category>
		<category><![CDATA[communication]]></category>
		<category><![CDATA[constraint]]></category>
		<category><![CDATA[inference]]></category>
		<category><![CDATA[language context]]></category>
		<category><![CDATA[mit]]></category>
		<category><![CDATA[natural language processing]]></category>
		<category><![CDATA[today]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=788</guid>
		<description><![CDATA[This article on today&#8217;s MIT front page sketches an argument that I&#8217;ve been thinking about for a while, ever since the IBM Watson Jeopardy contest &#8212; that natural language processing is hopeless to the extent that there is additional context (side information) not visible to the machine. Many prominent linguists &#8230; have argued that language [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://web.mit.edu/newsoffice/2012/ambiguity-in-language-0119.html">This article</a> on today&#8217;s MIT front page sketches an argument that I&#8217;ve been thinking about for a while, ever since the <a href="?p=306">IBM Watson</a> Jeopardy contest &#8212; that natural language processing is hopeless to the extent that there is additional context (side information) not visible to the machine.</p>
<blockquote><p>Many prominent linguists &#8230; have argued that language is &#8230; poorly designed for communication [and] &#8230; point to the existence of ambiguity: In a system optimized for conveying information&#8230;, they argue, each word would have just one meaning. &#8230; Now, a group of MIT cognitive scientists has turned this idea on its head &#8230; [T]hey claim that ambiguity actually makes language more efficient, by allowing for the reuse of short, efficient sounds that listeners can easily disambiguate with the help of context.</p></blockquote>
<p>Although this is just talking about homophonic ambiguity at the word level, the same applies to all levels of language, including full messages whose decoding requires potentially deep context.<br />
<span id="more-788"></span><br />
It isn&#8217;t surprising that human speakers use context &#8212; ad-hoc or shared understanding &#8212; to save effort. It is annoying to have to spell out everything and delightful when the listener understands anyway. It is like a communication problem, but with a power constraint on the encoder and a complexity constraint on the decoder. The listener (decoder) is really trying to choose the most probable meaning from all the possibilities, and the speaker (encoder) attempts to construct &#8212; specifically for the decoder&#8217;s context &#8212; the most modal distribution of meanings where the intended meaning takes the largest probability.</p>
<p>A few thoughts on this. One, successful encoder-decoder pairs synchronize context. Two, viewed in this way, double entendres and certain jokes are not mere oddities of language, but <strong>sophisticated communication schemes</strong>!</p>
<p>One could go further and argue that there is even a counterpart to steganography, because there is a certain amusement in hiding most of the intended information within the context, while making the direct message without context empty or obfuscating for any other decoder. An encoder that successfully does this knows the decoder well, and a decoder that successfully processes this information possesses either a large amount of shared context with the encoder or a sophisticated inference capability.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=788</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>watson v. mit</title>
		<link>https://blog.yhuang.org/?p=306</link>
		<comments>https://blog.yhuang.org/?p=306#comments</comments>
		<pubDate>Thu, 17 Feb 2011 02:52:42 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[common sense]]></category>
		<category><![CDATA[condition]]></category>
		<category><![CDATA[experiential knowledge]]></category>
		<category><![CDATA[general purpose]]></category>
		<category><![CDATA[image]]></category>
		<category><![CDATA[knowledge]]></category>
		<category><![CDATA[mit]]></category>
		<category><![CDATA[sense knowledge]]></category>
		<category><![CDATA[watson]]></category>

		<guid isPermaLink="false">http://allegro.mit.edu/~zong/wpress/?p=306</guid>
		<description><![CDATA[So being at the event captured in the image, I got to ask a question toward the end. Actually I asked two questions. The first was whether Watson would ring in and use the remaining 3 seconds or whatever to continue to compute. Gondek said it would if it helped. In actual competition it doesn&#8217;t [...]]]></description>
			<content:encoded><![CDATA[<p><img src="wp-content/uploads/images/539w.jpg" alt="http://cache.boston.com/resize/bonzai-fba/Globe_Photo/2011/02/14/1297740468_0202/539w.jpg" /></p>
<p>So being at the event <a href="http://www.boston.com/business/technology/articles/2011/02/15/computer_holds_its_own_in_1st_jeopardy_contest/">captured in the image</a>, I got to ask a question toward the end. Actually I asked two questions. The first was whether Watson would ring in and use the remaining 3 seconds or whatever to continue to compute. Gondek said it would if it helped. In actual competition it doesn&#8217;t appear to be the case, as the buzz-in thresholding condition ensured that further computation would not have been helpful. The second question was a follow-up on the identified weakness of Watson &#8212; learning on &#8220;common sense&#8221; knowledge. I asked what path AI research would take to tackle such knowledge, which are by its very definition, &#8220;not in the books.&#8221; Gondek said that IBM is building up semantic information (e.g. a &#8220;report&#8221; is something that can be &#8220;turned in&#8221; and &#8220;assessed,&#8221; etc.) from corpus. That wasn&#8217;t exactly what I was asking, however.</p>
<p>My point was whether all &#8220;knowledge&#8221; is written down. There is such a thing as experiential &#8220;knowledge,&#8221; and humans take years to learn it/be trained in it through parenting (i.e., to &#8220;mature&#8221;). If only there were a handbook on life, or that life could be learned through reading a series of textbooks, then perhaps I&#8217;d believe that the kind of general-purpose AI that most people are probably imagining (rather than an expert/Q&#038;A system) can be achieved along the lines of current methods.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=306</wfw:commentRss>
		<slash:comments>1</slash:comments>
		</item>
		<item>
		<title>fail</title>
		<link>https://blog.yhuang.org/?p=239</link>
		<comments>https://blog.yhuang.org/?p=239#comments</comments>
		<pubDate>Mon, 15 Feb 2010 06:48:24 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Ave]]></category>
		<category><![CDATA[mass ave]]></category>
		<category><![CDATA[mit]]></category>
		<category><![CDATA[piece]]></category>
		<category><![CDATA[windsor street]]></category>
		<category><![CDATA[work]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=239</guid>
		<description><![CDATA[Must have been a piece of work by MIT students&#8230; Windsor Street near Mass. Ave.]]></description>
			<content:encoded><![CDATA[<p>Must have been a piece of work by MIT students&#8230; Windsor Street near Mass. Ave.</p>
<p><img src="wp-content/uploads/images/fail.jpg" /></p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=239</wfw:commentRss>
		<slash:comments>1</slash:comments>
		</item>
	</channel>
</rss>
