<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Some stuff &#187; problem</title>
	<atom:link href="http://blog.yhuang.org/?feed=rss2&#038;tag=problem" rel="self" type="application/rss+xml" />
	<link>https://blog.yhuang.org</link>
	<description>here.</description>
	<lastBuildDate>Wed, 27 Aug 2025 08:50:58 +0000</lastBuildDate>
	<language>en</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=3.1.1</generator>
		<item>
		<title>die throwing problem</title>
		<link>https://blog.yhuang.org/?p=1796</link>
		<comments>https://blog.yhuang.org/?p=1796#comments</comments>
		<pubDate>Thu, 14 Sep 2017 03:21:41 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[dice]]></category>
		<category><![CDATA[math]]></category>
		<category><![CDATA[Probability]]></category>
		<category><![CDATA[problem]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=1796</guid>
		<description><![CDATA[Here&#8217;s a link to a subtle probability problem. You throw a die until you get 6. What is the expected number of throws (including the throw giving 6) conditioned on the event that all throws gave even numbers? The &#8220;obvious&#8221; answer is incorrect. The correct answer can be brute-forced by computing probabilities of this sort: [...]]]></description>
			<content:encoded><![CDATA[<p>Here&#8217;s a <a href="https://gilkalai.wordpress.com/2017/09/08/elchanan-mossels-amazing-dice-paradox-answers-to-tyi-30/">link</a> to a subtle probability problem.</p>
<blockquote><p>You throw a die until you get 6. What is the expected number of throws (including the throw giving 6) conditioned on the event that all throws gave even numbers?</p></blockquote>
<p>The &#8220;obvious&#8221; answer is incorrect.<br />
<span id="more-1796"></span><br />
The correct answer can be brute-forced by computing probabilities of this sort:<br />
If \(s\) is a sequence of length \(n\), then<br />
\(P_n\triangleq\) = P(\(s\) ends in 6 and \(s\) all even) = \((2/6)^{n-1} (1/6)\)<br />
(The total probability over all \(n\) is therefore \(Z=(1/6)/(4/6)=1/4\).)<br />
The expected length of \(s\) is \(\sum_{n=1}^\infty n (P_n / Z)\)<br />
\(= \sum_{n=1}^\infty n (2/6)^{n-1} (1/6) / (1/4) = \sum_{n=1}^\infty n (2/6)^{n-1} (2/3) = 3/2\)</p>
<p>This is interesting since premature conditioning on the 3 even sides of a die in every roll produces an (incorrect) expected length of 3.</p>
<p>There is already an elegant and convincing derivation of the correct answer attributed to Paul Cuff if you follow the links from the original article, but here&#8217;s an intuitive explanation for why the <em>incorrect</em> answer overestimates the expected length in such a way &#8211;</p>
<p>Notice that in \(P_n\) the probability \((1/6)\) of obtaining the final 6 has no effect on the expected length. The expected length depends entirely on the shape of \(P_n\) over \(n\). The calculation for the &#8220;incorrect&#8221; reasoning would have made the sum \(\sum_{n=1}^\infty n (2/3)^{n-1} (1/3) = 3\), and the difference is in the \((2/3)^{n-1}\) vs. \((2/6)^{n-1}\) term. So the existence of 1, 3, and 5 matter; they make longer prefixes of 2 and 4 even less likely, as more of the longer sequences that would have been valid ends up containing a 1, 3, or 5 instead and getting thrown away.</p>
<p>Note: There is also a good discussion <a href="https://www.quora.com/What-is-the-expected-number-of-rolls-of-a-fair-six-sided-die-to-get-a-6-given-that-all-rolls-leading-up-to-that-6-are-even-numbered">on Quora</a>.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=1796</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>capital markets</title>
		<link>https://blog.yhuang.org/?p=922</link>
		<comments>https://blog.yhuang.org/?p=922#comments</comments>
		<pubDate>Thu, 13 Sep 2012 21:43:38 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[assumption]]></category>
		<category><![CDATA[Capital markets]]></category>
		<category><![CDATA[consumption]]></category>
		<category><![CDATA[contract]]></category>
		<category><![CDATA[expectation]]></category>
		<category><![CDATA[problem]]></category>
		<category><![CDATA[return]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=922</guid>
		<description><![CDATA[QE3 was announced today and reactions have been relatively muted. There are some complaints that money is again being redistributed from asset holders to debtors via the mechanism of negative real rates. It seems like a good occasion to put forth two oddities that I&#8217;ve always seen as embedded in capital markets as they&#8217;re currently [...]]]></description>
			<content:encoded><![CDATA[<p>QE3 <a href="http://www.bloomberg.com/news/2012-09-13/fed-plans-to-buy-40-billion-in-mortgage-securities-each-month.html">was announced today</a> and reactions have been relatively muted. There are some complaints that money is again being redistributed from asset holders to debtors via the mechanism of negative real rates. It seems like a good occasion to put forth two oddities that I&#8217;ve always seen as embedded in capital markets as they&#8217;re currently constructed. They are: the assumption that money doesn&#8217;t spoil, and the assumption of market optimality.<br />
<span id="more-922"></span><br />
The first assumption. Most people expect money to be a store of value, of an almost contractual nature. This is why they keep it in the bank or invest in capital markets in the expectation that a non-negative real return is deserved. When this does not happen, they are understandably unsettled. However, whence comes the notion that money doesn&#8217;t spoil like food? In the days of barter, non-perishable goods were stored and exchanged as money is, but their non-spoilage was predicated on the fact that they were goods that were to be directly used during the consumption period. In separating the good from the money function, now, excess production can only be &#8220;stored&#8221; (it is of course actually immediately consumed) to the extent that the social contract will be honored to exchange them for goods during the consumption period. This is where non-perishability is presumed lost. The social contract does not specify the exchange rate of money to goods. In expecting a non-negative real return on money, we are essentially expecting that during the consumption period, we will get at least as much value as we &#8220;stored,&#8221; as measured by the valuation of goods at the time of &#8220;storage.&#8221; This is exactly the non-perishability assumption. This is very strong. As mentioned, everything in nature is perishable to some extent. We are already expecting more than is natural right off the bat. The only reason such a social contract is even possible is due to non-decreasing output, i.e. the assumption of growth.</p>
<p>Then there are those who expect not only to have a non-negative real return, but at least a market return. This is the expectation that during the consumption period, we will get at least as much value as we &#8220;stored&#8221; <em>proportionally</em> to the fraction of output we produced in the economy at the time of &#8220;storage.&#8221; For example, if technology improved, we would expect that our &#8220;stored&#8221; widget version 1, would be returned to us as the much improved widget version 10. This seems greedy yet fair. In  contrast, the expectation for non-negative real return seems rather tame now, doesn&#8217;t it? Until you consider that the economy might actually contract. What if, due to whatever circumstances, the excess producers during the consumption period do not produce enough value to proportionally divide among those who &#8220;stored&#8221; value an amount that is more than what they stored? This could be by will or by necessity. What if all we had were widget version 0.5? Would it still be plausible to insist on non-negative returns? It would not.</p>
<p>In fact, following a period of great excess in production, especially one that prematurely fulfilled useful work for some time into the future (until new technology and demand develop), the right view is that the nominal amount of stored value is irrelevant &#8212; it is not exchangeable.</p>
<p>Now the second assumption. Capital markets are a great invention. They solve the centralized allocation problem in a distributed way. We assume that rational actors freely acting in their own self interests will produce a solution that allocates capital in the &#8220;optimal&#8221; way. Perhaps it isn&#8217;t a global optimal (requires cooperative strategies), but a more basic question is what planning problem do they solve in the first place? Let us give individual actors the benefit of the doubt that they can do planning and are perfectly rational. Nevertheless, it seems difficult to imagine that any would do planning much beyond their own lifespans. That would not be rational. Thus, capital markets are essentially solving an allocation problem based on each individual&#8217;s preferences for maximizing their own utilities <em>during their lifespans</em>. This has great implications. First, it puts an upper bound on the planning period for the allocation problem at the longest remaining lifespan among individuals. This is like 80 years or so. That is pretty myopic. It seems like a long time, but for many decisions, especially concerning resources, it is very myopic.</p>
<p>It gets worse. That was just an upper bound. Inputs into capital markets are weighted by the amount of money committed to a trade. It is a given that older (working) people have greater wealth, and as a consequence their inputs as expressed by capital commitment are weighted more in the allocation decisions. That is exactly what you don&#8217;t want, because a shorter remaining lifespan is now correlated with larger weight in decisions. The net result is that the capital allocation decisions are being carried out for optimality over very short planning periods, possibly only on the order of 10 or 20 years at most. It isn&#8217;t even optimal for most people alive any more.</p>
<p>So this is a fundamental problem in capital markets. Under the law of one price for open markets, only one view can be ultimately carried out, and it is one that is too myopic, and even myopic for most people. The only way to correct for this is to weight the inputs properly, and that would require external intervention to enforce a political planning decision.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=922</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>two problems with win7</title>
		<link>https://blog.yhuang.org/?p=907</link>
		<comments>https://blog.yhuang.org/?p=907#comments</comments>
		<pubDate>Sun, 27 May 2012 20:38:30 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[CPU]]></category>
		<category><![CDATA[drain]]></category>
		<category><![CDATA[fan]]></category>
		<category><![CDATA[problem]]></category>
		<category><![CDATA[win]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=907</guid>
		<description><![CDATA[I finally had enough of two long-standing problems that bugged me in Windows 7, so I fixed them. 1. The &#8216;System&#8217; process uses a low level of CPU (~6%) when nothing is going on. 2. Chinese characters in the system (e.g. file names) suddenly become boxes. The first problem caused excessive battery drain along with [...]]]></description>
			<content:encoded><![CDATA[<p>I finally had enough of two long-standing problems that bugged me in Windows 7, so I fixed them.</p>
<p>1. The &#8216;System&#8217; process uses a low level of CPU (~6%) when nothing is going on.<br />
2. Chinese characters in the system (e.g. file names) suddenly become boxes.</p>
<p>The first problem caused excessive battery drain along with the CPU fan going on and off all the time. Because Task Manager doesn&#8217;t tell you what is under the &#8216;System&#8217; process, I had to use <a href="http://technet.microsoft.com/en-us/sysinternals/bb896653.aspx">Process Explorer</a>, which promptly identified the culprit: a revolving series of threads labeled &#8216;ntoskrnl.exe!KeReleaseInStackQueuedSpinLock+0x1e0&#8242;. This would appear to be some issue with a driver stuck on something and aimlessly retrying. It turns out turning off Wifi cured it. An updated driver for the Broadcom 4313 802.11n card fixed the problem permanently. The old version was 5.100.82.82 from 5/2011. The new version that does not have the problem is 5.100.82.112 from 10/2011.</p>
<p>The second problem is caused by a little-known new setting in Windows 7. Under Control Panel -> Fonts -> Font settings, there is an option &#8220;Hide fonts based on language settings.&#8221; It&#8217;s simply broken, so I just turned it off.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=907</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>phone vs. tablet vs. laptop vs. desktop vs. server</title>
		<link>https://blog.yhuang.org/?p=644</link>
		<comments>https://blog.yhuang.org/?p=644#comments</comments>
		<pubDate>Fri, 02 Sep 2011 00:39:28 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[computer]]></category>
		<category><![CDATA[computing history]]></category>
		<category><![CDATA[desktop]]></category>
		<category><![CDATA[dual interface]]></category>
		<category><![CDATA[functional view]]></category>
		<category><![CDATA[hardware]]></category>
		<category><![CDATA[hardware distribution]]></category>
		<category><![CDATA[problem]]></category>
		<category><![CDATA[support]]></category>
		<category><![CDATA[windows media center]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=644</guid>
		<description><![CDATA[It seems that Microsoft&#8217;s all-in-one strategy on support for different devices is still progressing. Windows 8 will have interfaces for both the desktop and touchscreen devices. This is akin to how Windows Media Center works. This model must have an unusual level of attraction to Microsoft due to the large base of existing applications, but [...]]]></description>
			<content:encoded><![CDATA[<p>It seems that Microsoft&#8217;s all-in-one strategy on support for different devices is still progressing. Windows 8 <a href="http://news.cnet.com/8301-10805_3-20100365-75/windows-8-to-offer-both-metro-and-desktop-interface/">will have interfaces for both the desktop and touchscreen devices</a>. This is akin to how Windows Media Center works. This model must have an unusual level of attraction to Microsoft due to the large base of existing applications, but it makes assumption that you&#8217;d want to use all the applications on all the devices, if only you could &#8212; that may turn out not to be right.</p>
<p>Microsoft has for years tried to get into mobile devices. <a href="http://www.youtube.com/watch?v=oLUSnPB08kc">Here</a> you see Bill Gates really uncomfortable with the notion that Apple has succeeded more than Microsoft in this space. He is not wrong, since for a time Windows phones and tablets were the only ones out there, while Apple&#8217;s Newton was forgotten memory. Those devices either used a slightly modified Windows OS or one that copied all of its metaphors. The latest Windows phones are an exception, but with Windows 8, it will no longer be. It cannot be disputed that there are important applications that do not exist on mobile devices (currently), and therefore mobile devices are not complete (currently). So people argue that mobile devices will be full-fledged computers or desktops will not die. The idea of a dual interface seems to be aimed in this direction. However, a third possibility exists. Applications, after all, merely solve real life problems. They are not themselves holy. If there were a different way of accomplishing the same things, the applications could be replaced. One could argue that data is the rather more holy object. Back to this later.<br />
<span id="more-644"></span><br />
While devices are converging, it becomes a question of what the hardware distribution of the future will look like, and how functions will be partitioned among them. In <a href="http://www.youtube.com/watch?v=5UeLk6vmbtM">this video</a>, Steve Jobs posits a rather linear, functional view of computing history, where things moved from desktops to more mobile devices as usage functions evolved from scientific and office work to entertainment and socializing. Bill Gates posits a more encompassing and parasitic view, seeing computing power spreading to colonize all niches, a niche for each device, none really going away. To him, the kind of space or niche makes a difference, like whether a device fits in a pocket or not.</p>
<p>What they are both getting at is that there are constraints &#8212; some hardware, some social convention &#8212; that limit what functions can be used where. Because if it were at all possible, why wouldn&#8217;t one want all functions on all devices? But there are power, weight, screen-size, and input device constraints that are fundamental. Given that, you can&#8217;t possibly have all applications run in all devices. </p>
<p>To address this, one way is to have all devices become one device &#8212; a hardware solution along with its companion software like Windows 8. This &#8220;classic&#8221; solution has existed quite a while now, e.g. convertible laptops, some better than others. The problem is in both hardware and software. The equivalent tethered power and heavy-case computing power cannot be had with mobility at any given time, even though mobile devices are more powerful than computers of even a few years ago. And the software interface is also different &#8212; requiring a stylus for mouse-like precision (although I like the stylus, it&#8217;s one more thing to hold). With Windows 8, the interface problem <em>maybe</em> is solved, but the hardware problem remains. There is talk of some dual-part computer where you can remove a light (both weight and CPU power) piece of it. The non-mobile base of such a computer would have additional processing capabilities as well as keyboard and mouse like a standard docking station. The hardware design for this though, would be enormously complicated if it were to be efficient. For example, two processors separated a great physical distance, does not make for good communication speed. Either that, or when the light piece is docked, its own capabilities are totally disabled for its trivial contribution to total computing power. This would be a waste of hardware and the cost would be even greater than a tablet and a separate non-mobile computer combined.</p>
<p>So what about another way. Forget combining all devices into one device &#8212; in hardware. Why not have all these devices, and even let them run all their vastly different applications and interfaces at vastly different processing capabilities, but combine them at the level of <em>data</em>? Given the constraints of the devices, people will write any and all applications that support functions natural to them &#8212; we need not worry about that. We only care that these applications can access a common set of data and have seamless sync&#8217;ing between them. This also has a buzzword already, it&#8217;s called cloud computing. Yet I don&#8217;t think it&#8217;s about migrating applications to the cloud &#8212; not so, although some of that will take place, for &#8220;light&#8221; applications (light on bandwidth and computation). The full power of each device though, is going to harnessed, I am sure of that. So the best gains from the cloud is data sync&#8217;ing. This is a problem not merely of sync&#8217;ing, but of a method to record data in a way that is universally available regardless of software <em>or</em> hardware platform. It&#8217;s not just document data, but things like preferences, and program states. And I&#8217;m not talking about applications that are simple and entertainment-like or applications already on the web for which devices are only terminals. Furthermore, this &#8220;cloud&#8221; doesn&#8217;t even need to be an internet company, it can be managed among the devices themselves or by any mostly-on device that is at a common locus of interaction, like a &#8220;cloud server&#8221; or some such in the home. I think this is the more likely future, because it makes more sense.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=644</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>risk matching in gambling</title>
		<link>https://blog.yhuang.org/?p=520</link>
		<comments>https://blog.yhuang.org/?p=520#comments</comments>
		<pubDate>Thu, 23 Jun 2011 06:57:06 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[chip unit]]></category>
		<category><![CDATA[fruitful path]]></category>
		<category><![CDATA[game]]></category>
		<category><![CDATA[group scheme]]></category>
		<category><![CDATA[player]]></category>
		<category><![CDATA[problem]]></category>
		<category><![CDATA[risk profiles]]></category>
		<category><![CDATA[risk reward]]></category>
		<category><![CDATA[Scheme]]></category>
		<category><![CDATA[size]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=520</guid>
		<description><![CDATA[An argument for playing a game such as poker with &#8220;real&#8221; money is that it forces people to play with true risk-reward calculations. While this is certainly better than playing without risk, there exists the question of how to match risk profiles among players. With enough players (large liquid market), they can self-sort by stake [...]]]></description>
			<content:encoded><![CDATA[<p>An argument for playing a game such as poker with &#8220;real&#8221; money is that it forces people to play with true risk-reward calculations. While this is certainly better than playing without risk, there exists the question of how to match risk profiles among players. With enough players (large liquid market), they can self-sort by stake size, and this seems fair. With only few people though, the situation is turned around, where a stake size has to be <em>agreed upon</em> at some clearing size (so that enough people agree to play the game) rather than chosen individually, and that same amount of money may be considered as very different values by different people. A pauper and a millionaire do not see $100 as the same value, and will adjust their utilities accordingly, and this will materially affect wagering. Since risk is measured in utility units, it is desirable to match utilities rather than dollar amounts. But there isn&#8217;t an agreed-upon utility currency. Or is there?<br />
<span id="more-520"></span><br />
There is: the chips, they are perfect representations. So utility matching could be a fruitful path. Start with this: suppose prior to playing, each player declares an &#8220;exchange rate&#8221; for chips, say 1 chip-unit = $5, and pays however much it takes to get the amount of chips needed for his stake in the game. Without a loss of generality, all games can be played with say 1000 chip units. So this player would pay $5000 to get his chips. Another player who declares 1 chip-unit = $0.05 would pay $50 to get her chips, and so on. At the end of the game, the winning player takes all of the chips at the table and exchanges them back for money. For example, If four people played, then there are 4000 chips, so the person who declared 1 chip-unit = $5 would get $20K, and the person who declared 1 chip-unit = $0.05 would get $200. If we call the usual way of stake-setting with a small group Scheme A, then we call the one described in this paragraph Scheme B.</p>
<p>Two problems arise with Scheme B. One, winning players whose dollar-per-chip exchange rate happens to be &#8220;higher&#8221; than average can&#8217;t be paid back fully from that game alone, or ever if he keeps winning. Two, the players may not declare truthfully. Is there a clean way out of this?</p>
<p>The second problem is complicated, so ignore it now. The first problem occurs because the group of players does not, in aggregate, support the payoff profile demanded by the players with above-average dollar-per-chip exchange rates, if they also win more than average. In this context, we see that Scheme A (a group playing with agreed-upon equal-dollar stakes) gets around this problem by essentially setting a common exchange rate for everybody at the minimum of the group (that which enticed the last person to join). This works, but can we combine Scheme A&#8217;s payoff fairness with Scheme B&#8217;s utility matching fairness?</p>
<p>To some extent, yes. A marginal improvement can be made by noting that, as long as the players do not win 100% of the games, some exchange rate flexibility can be allowed. This is an interesting fundamental trade-off: the less spread out the winning percentages of players, the more spread out they can bid their exchange rates, and vice versa. In the one extreme, players set the same exchange rates (Scheme A) and any game outcome can be supported. At the other extreme, if the players are equally likely to win, then on average no chips change hands, so it wouldn&#8217;t matter what exchange rates were set (Scheme B). I should probably quantify what happens in the middle but can&#8217;t be bothered to do it today.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=520</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>learning in social networks</title>
		<link>https://blog.yhuang.org/?p=307</link>
		<comments>https://blog.yhuang.org/?p=307#comments</comments>
		<pubDate>Mon, 07 Mar 2011 18:31:09 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[belief propagation]]></category>
		<category><![CDATA[binary decision]]></category>
		<category><![CDATA[connected graph]]></category>
		<category><![CDATA[decision]]></category>
		<category><![CDATA[decision-making]]></category>
		<category><![CDATA[error propagation]]></category>
		<category><![CDATA[folk belief]]></category>
		<category><![CDATA[learning]]></category>
		<category><![CDATA[model]]></category>
		<category><![CDATA[problem]]></category>

		<guid isPermaLink="false">http://allegro.mit.edu/~zong/wpress/?p=307</guid>
		<description><![CDATA[There was this talk (by M. Dahleh) on modeling whether distributed learning occurs in a social network, i.e., is the crowd always right? The problem model was like this: there is a &#8220;truth&#8221; which is either 0 or 1, representing some binary preference. Then in a connected graph representing a learning network, each node makes [...]]]></description>
			<content:encoded><![CDATA[<p>There was <a href="http://www.sysid2009.org/plenaries/dahleh.pdf">this talk (by M. Dahleh)</a> on modeling whether distributed learning occurs in a social network, i.e., is the crowd always right? The problem model was like this: there is a &#8220;truth&#8221; which is either 0 or 1, representing some binary preference. Then in a connected graph representing a learning network, each node makes a binary decision (0 or 1 again) based on an independent noisy read on the &#8220;truth,&#8221; as well as the decisions made by some or all of its neighbors who have already made a decision. (Each nodal decision is made once and binding, so there is a predefined decision-making order among the nodes.)</p>
<p>This is an interesting question because at first thought, one would think that in a large enough network, a sufficient number of independent reads on the truth will occur in the aggregate to allow at least the later-deciding nodes to get a really good estimate of the truth. This is the basis of the folk belief in &#8220;wisdom of the crowd.&#8221; However, this is not what happens all the time.</p>
<p><span id="more-307"></span><br />
The problem really lies in the fact that a social network is (inadvertently) running a rather constrained version of belief-propagation, not the full-fledged belief-propagation that we would like, in which the nodes are more sharing. For example, in belief-propagation, full distributional beliefs are passed instead of these binary decisions, where estimation information is abridged and suppressed locally. If the observed decisions embody estimates so compressed as to be heavily distorted and not very indicative of the truth, then given certain network topologies, error propagation may not decay away. Then there is also something about single-pass decision-making that makes it impossible for nodes with the wrong answers to correct themselves and for the network overall to always converge on the right answer. Of course, some cooperative protocol could solve the problem, but we can&#8217;t really assume all nodes in a social network are cooperative, or even not adversarial.</p>
<p>So the crowd is not always right. In fact, it can be manipulated by &#8220;excessively influential&#8221; nodes in bad network topologies, as some other results in the talk indicated. Learning occurs a little more robustly in network topologies where some small (but non-negligible) fraction of nodes are &#8220;independently observing&#8221; ones that don&#8217;t listen to other nodes, but are listened to.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=307</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>double-sided usb</title>
		<link>https://blog.yhuang.org/?p=303</link>
		<comments>https://blog.yhuang.org/?p=303#comments</comments>
		<pubDate>Thu, 27 Jan 2011 22:46:02 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[moving parts]]></category>
		<category><![CDATA[pins]]></category>
		<category><![CDATA[point]]></category>
		<category><![CDATA[problem]]></category>
		<category><![CDATA[receptacle]]></category>
		<category><![CDATA[Stuff]]></category>
		<category><![CDATA[two rows]]></category>
		<category><![CDATA[two sides]]></category>
		<category><![CDATA[usb]]></category>

		<guid isPermaLink="false">http://allegro.mit.edu/~zong/wpress/?p=303</guid>
		<description><![CDATA[This design is pretty good, though it would probably break prematurely. Stuff that needs to be plugged in and out repeatedly maybe should not have moving parts. More to the point, why was USB specified to be one-sided in the first place? There are no two rows of pins like in some connectors, so there [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://www.yankodesign.com/2011/01/25/this-usb-plugs-in-both-ways/">This design</a> is pretty good, though it would probably break prematurely. Stuff that needs to be plugged in and out repeatedly maybe should not have moving parts. More to the point, why was USB specified to be one-sided in the first place? There are no two rows of pins like in some connectors, so there is really no problem with a receptacle having contacts on two sides.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=303</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>different kind of coupon collector&#8217;s problem</title>
		<link>https://blog.yhuang.org/?p=283</link>
		<comments>https://blog.yhuang.org/?p=283#comments</comments>
		<pubDate>Mon, 11 Oct 2010 09:18:39 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[call]]></category>
		<category><![CDATA[column sums]]></category>
		<category><![CDATA[math 1]]></category>
		<category><![CDATA[math types]]></category>
		<category><![CDATA[number]]></category>
		<category><![CDATA[partition functions]]></category>
		<category><![CDATA[pigeonhole principle]]></category>
		<category><![CDATA[problem]]></category>
		<category><![CDATA[sum]]></category>
		<category><![CDATA[theory]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=283</guid>
		<description><![CDATA[The classic coupon collector&#8217;s problem asks for the number of samples it takes to collect all coupon types from a sequence of coupons in which each of the types of coupons has an unlimited number of copies. Here is a different kind of problem: if there are limited copies of each of the coupon types [...]]]></description>
			<content:encoded><![CDATA[<p>The classic coupon collector&#8217;s problem asks for the number of samples it takes to collect all coupon types from a sequence of coupons in which each of the \(k\) types of coupons has an unlimited number of copies.</p>
<p>Here is a different kind of problem: if there are limited copies of each of the \(k\) coupon types (say, \(d\) copies), how many samples does it take to collect \(t\) coupon types?</p>
<p><span id="more-283"></span><br />
By the pigeonhole principle, we have two extremes: from \(n=0\) to \(n=t-1\) samples, it is impossible to collect \(t\) coupon types; and for \(n>(t-1)d\) samples, it is impossible not to. For number of samples in between, there is some monotonically decreasing probability that \(t\) or more types are <strong>not</strong> collected. Let&#8217;s call the latter probability \(P(n)\). The expected number of samples it takes to collect \(t\) types is then</p>
\(\sum_{n=1}^{(t-1)d} P(n)\)
<p>Finding \(P(n)\) for \(n\ge t\) doesn&#8217;t appear easy, though the combinatorics could be written down (in theory), maybe involving partition functions, Stirling numbers, or whatever. But the question is reduced to the following, which somebody must have solved: given a \(d \times k\) 0-1 matrix, in which \(n\) 1&#8242;s are placed randomly, what is the probability that exactly \(z\) columns are 0-weight (have 0 column sums)?</p>
<p>These papers give some approximations:</p>
<p>http://arxiv.org/PS_cache/arxiv/pdf/0806/0806.1480v3.pdf</p>
<p>http://cs.anu.edu.au/~bdm/papers/irregular.pdf</p>
<p>Does anybody know the actual answer?</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=283</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>stupid powerpoint</title>
		<link>https://blog.yhuang.org/?p=249</link>
		<comments>https://blog.yhuang.org/?p=249#comments</comments>
		<pubDate>Fri, 12 Mar 2010 10:33:31 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[default master]]></category>
		<category><![CDATA[degree symbol]]></category>
		<category><![CDATA[different fonts]]></category>
		<category><![CDATA[font]]></category>
		<category><![CDATA[initial drawings]]></category>
		<category><![CDATA[master template]]></category>
		<category><![CDATA[problem]]></category>
		<category><![CDATA[symbol]]></category>
		<category><![CDATA[template]]></category>
		<category><![CDATA[TeX]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=249</guid>
		<description><![CDATA[I ran into a problem with TeX4PPT the other day. * As an aside, TeX4PPT inside PowerPoint is probably the fastest way to make scalable figures and math. There is just no way that any other drawing tool even comes close to the speed with which simple diagrams can be made using PowerPoint. Since the [...]]]></description>
			<content:encoded><![CDATA[<p>I ran into a problem with <a href="http://users.ecs.soton.ac.uk/srg/softwaretools/presentation/TeX4PPT/">TeX4PPT</a> the other day.</p>
<p>* As an aside, TeX4PPT inside PowerPoint is probably the fastest way to make scalable figures and math. There is just no way that any other drawing tool even comes close to the speed with which <em>simple</em> diagrams can be made using PowerPoint. Since the LaTeX people seem completely uninterested in the end user, there is basically no choice in the matter. My work flow these days involves LyX initial drafts with LaTeX final drafts for papers and PowerPoint with TeX4PPT for presentations, using pen input for the initial drawings. Not ideal, but getting there.</p>
<p>But I ran into a problem where every time I saved any LaTeX expression with the symbol \(\gamma\), it turns into a degree symbol, that little circle °. It turns out \(\gamma\) is codepoint <tt>U+00B0</tt> in the font <tt>cmmi10</tt>. But PowerPoint&#8217;s default master template has gone stupid, and never wants to change the font type of <u>just that one codepoint</u>, so it goes with the Arial glyph there, which is the degree symbol. It doesn&#8217;t help that LaTeX is stuck in the stone age and doesn&#8217;t know about Unicode, instead using about 20 different fonts with overlapping codepoints to make the expressions.</p>
<p>So in the end I had to carefully recreate a clean master template in PowerPoint. For future use, it is saved <a href="wp-content/uploads/blank.pot">here</a>.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=249</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>google wave lacks structure</title>
		<link>https://blog.yhuang.org/?p=226</link>
		<comments>https://blog.yhuang.org/?p=226#comments</comments>
		<pubDate>Tue, 01 Dec 2009 22:03:43 +0000</pubDate>
		<dc:creator>admin</dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[communication]]></category>
		<category><![CDATA[conversation mode]]></category>
		<category><![CDATA[endeavor]]></category>
		<category><![CDATA[linear structure]]></category>
		<category><![CDATA[metaphor]]></category>
		<category><![CDATA[playback]]></category>
		<category><![CDATA[problem]]></category>
		<category><![CDATA[structure]]></category>
		<category><![CDATA[today]]></category>
		<category><![CDATA[topic]]></category>

		<guid isPermaLink="false">http://scripts.mit.edu/~zong/wpress/?p=226</guid>
		<description><![CDATA[Got an invitation to Google Wave today. The problem I find immediately is the lack of structure. Say what you will about the restrictions of email or IM, but the same restrictions of those ways of communication, namely time-flow or thread-flow, are also well enforced structures to keep things sane. Wave takes away these and [...]]]></description>
			<content:encoded><![CDATA[<p>Got an invitation to Google Wave today. The problem I find immediately is the lack of structure. Say what you will about the restrictions of email or IM, but the same restrictions of those ways of communication, namely time-flow or thread-flow, are also well enforced structures to keep things sane. Wave takes away these and substitutes &#8220;playback.&#8221; Unfortunately, playback is not natural. (The other way is to fall back on social convention to keep order, but that rarely works with more than 2 peers.)</p>
<p>I think there are two options here. Either structure needs to be explicitly enforced or presentation needs to be refined.</p>
<p>In the former, perhaps it is better to only allow replies in certain places. Perhaps it is better to only allow edits in certain places. Perhaps it is better to separate the two and keep the distinction between edit mode, thread mode, and conversation mode, and only allow mixing in very restricted settings (or require some extra steps to discourage its use). After all, in preparing a shared endeavor, the purpose should be defined and known ahead of time.</p>
<p>In the latter, perhaps a lot of hiding and collapsing should be used. Perhaps hyperlinks should be used for in-place edits that often hijack a topic. And now that subthreads can sprout like a tree, it makes little sense to retain the linear structure of conversations. Perhaps a topic based graph, or a conversation stack would be the more appropriate presentation metaphor.</p>
<p>Wave is a good idea, but not well thought out. In its attempt to differentiate, it has forsaken useability for chaotic flexibility, which would have had redeeming value, were it matched by equally ambitious presentation/visualization.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.yhuang.org/?feed=rss2&#038;p=226</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
	</channel>
</rss>
