My Cuisinart ice cream maker had arrived only a few days before the e-mail did. I did a double-take when I saw the subject line:

trying to track down Nathan Kurz

Nathan Kurz? *The* Nathan Kurz? The Nathan Kurz of *Scream sorbet*? Yes, *that* one. The e-mail was from Charlotte Druckman, the reporter who had written about Scream for the *New York Times*. The article told the story of Nathan Kurz (yes, that one!), the computer programmer turned ice cream genius who had brought the creaminess of ice cream to the dairy-free world using nothing but the fat content of nuts. His Scream sorbet had built a devoted following in the Bay Area only to close its doors a few years later.

I was one of their devoted, and last Thanksgiving, I’d written a blog post about reviving their signature pistachio sorbet recipe. Since then, I’d been making new flavors, recreating old ones, and sharing the results at parties. It was all thanks to a recipe that Nathan Kurz had shared with Joy of Blending. I actually hadn’t seen or spoken with him since Scream had closed its doors, and this newbie to the ice cream world had a bunch of questions for him.

Charlotte Druckman had questions for him, too, and when she discovered my blog, figured she’d check if I had his contact info. I wanted the answer to be yes, so before responding, I got in touch with a friend in the food industry who might. While waiting for a response, I searched the Internet and came across a GitHub account by someone named Nathan Kurz.

I compared the account details with what I knew from the *New York Times* article. *This* Nathan Kurz had started the account in 2009, shortly before the birth of Scream sorbet but only started checking in code in late 2013, a few months after Scream had closed its doors. This Nathan Kurz was interested in CPU caches, and another Google search turned up a blog post he’d written about it. This Nathan Kurz’s blog included an e-mail address, so I sent him an e-mail asking if he was also the Nathan Kurz of Scream sorbet.

A response came within an hour:

]]>Hi Krish,

Yes, you’ve got the right me.

When Scream Sorbet opened a store front in the Temescal a few months later, I looked for any excuse to go, evangelizing it to friends and visitors alike. Like a Teddy Ruxpin that had just been turned on, my eyes would open wide as I described the magic of nut-based sorbets and attempted to win over converts. I would even share an article in the *New York Times* about Scream to entice skeptics to give it a try.

I wasn’t expecting such an abrupt end when Scream closed its doors in 2013, and despite a one-off pop-up at Bittersweet Cafe a few months later, the sorbet never found a home elsewhere or returned to farmers’ markets. It was reduced to just a memory. While nut-based sorbets found their way to grocery stores, I never found a pistachio sorbet that matched the taste of Scream’s.

On the face of it, it felt like it should be relatively simple to make. I had access to its four ingredients, and the *Times* article went into some detail about getting the proportions of fat to sugar to solid to liquid correct, as well as the type of blender Kurz used in his process, but without the numbers. The question was what those proportions were, so I searched online. It turned out that Kurz had shared his pistachio sorbet recipe, so I set out to revive the sorbet as a dessert following Thanksgiving dinner.

I was at my cousin’s for Thanksgiving, whose husband incidentally had first shown me the *Times* article. He became my partner in crime, and we combined the ingredients using a blender that featured none of the bells and whistles of a Pacojet or a Vitamix. I tried a spoonful of the blend, and the flavor that I thought was lost to history was in the spoon. I gave a sample to my cousin-in-law, whose eyes lit up like Teddy Ruxpin.

Suffice it to say that the pistachio sorbet was a hit, and we wanted to try it again the next day. We were out of pistachios, but my cousin had walnuts at home, which contain a greater percentage of fat than pistachios, so we combined it with some papayas, running the numbers to maintain the proportions of solid to liquid (papaya’s are roughly 91% water) to fat to sugar so that they roughly matched up with with the pistachio recipe. The result was a papya-walnut sorbet that was as creamy as ice cream. We had cracked the code!

For our batch, the cost of making the pistachio sorbet was roughly $1.50 a scoop, and I think Scream used to charge $3 or $4 a scoop, but the per scoop costs don’t tell the entire story. Based on some back-of-the-envelope calculations and the conservative estimate that retail space in the Temescal is $5k a month, the business needed at least one employee, e.g. the owner himself (in practice there were more) paid for a 40 hour work-week at $15/hour (i.e. minimum wage), Scream would have needed to sell nearly 20 scoops an hour to break even charging $4 a scoop or over 30 an hour charging $3 a scoop. During many of the times I was there, Scream was not pulling in anywhere near that kind of traffic.

On the other hand, if it were a dessert item in a restaurant, the expenses could be offset by other items on the menu. In the meantime, I’ve found a workaround.

]]>Then on a whim yesterday, I decided to try playing the chords of “Wagon Wheel” on my keyboard. I wasn’t setting out to learn how to play the piano and figured I could just follow a rhythm that roughly matched up with the strum pattern for the song on the guitar. In fact, the chord progression is a slight variation on the Axis of Awesome’s 4 chords, and my strumming pattern is just a series of quarter notes, so it’s completely straightforward.

My initial strategy was simply to voice the chords from their root notes, but it felt awkward to readjust all my fingers every time I changed chords. I revised the strategy accordingly: shift from chord to chord by readjusting as few fingers as possible. While I didn’t pen out a closed-form solution to this optimization problem, the heuristic I did find led to new chord voicings that made changing chords significantly more fluid.

Then I noticed that I was only using one hand to play, and the “Wagon Wheel” melody comprises only four notes. I started playing the chords as whole notes with my left hand and focused on the melody with my right, and pretty soon I was playing “Wagon Wheel” on the piano! The simple act of translating the song from guitar to piano taught me something about playing the piano, albeit with some habits I might need to unlearn were I to take lessons in the future.

So I repeated the process for “Dink’s Song” and then “Hey There, Delilah” while discovering how much fun it is to learn the piano this way.

]]>Now that I knew about the bug, the next thing to do was tell my advisor. It had been just a few weeks since he said to a luminary visiting campus, “One of my students generalized the entropy power inequality,” to which the luminary replied, “That’s impressive!” What had led me to smile then felt embarrassing now. Had the news spread elsewhere? Would this affect my advisor’s reputation in addition to my own if we ultimately had to retract the result? Why hadn’t I noticed the issue earlier?

To reach a plausible explanation for that last question, it helps to understand how my biases played a role in checking the proof. The incorrect proof had been following a technique similar to the one in Blackman’s “The convolution inequality for entropy powers”, and while Blackman confidently swapped things like the order of derivatives and expectations, I was less confident about these steps in my proof. To justify these steps required applying some measure-theoretic results, and while I’d been exposed to these in a couple probability theory courses, this was the first time I needed to apply them in my research. As a consequence of this insecurity, I focused my attention on making sure that these parts of the argument were watertight and failed to notice that I ultimately wouldn’t be able to apply these watertight arguments to prove the result.

The actual bug came from somewhere that didn’t require any measure-theoretic sophistication. Both Blackman’s proof and mine applied Gaussian perturbations to random variables, with the difference that I had introduced an auxiliary random variable and required a Markov chain to hold. The problem for my proof was that the way I was applying the Gaussian perturbations broke the Markov structure.

I swallowed my pride, found my advisor, and explained the technical issue to him. Our discussion immediately shifted to practical matters. Could we repair the bug in time for the camera-ready deadlines of the different conferences? Part of this would depend on whether we could use the same proof technique or not. We still had a couple weeks before the first camera-ready deadline (a paper that depended on the result) and three months before the camera-ready deadline for the last one (the result itself), so the decision was to work on a patch first and retract later if necessary. We started brainstorming some attacks and looking up possible references that could help.

]]>By February of 2006, the thesis had been filed, and paper acceptances had started coming in, so my focus had shifted to trying to find alternate ways to show the same result, and with that I had pulled up Dembo et al.’s “Information theoretic inequalities” among other papers of interest like Amir Dembo’s “Information inequalities and uncertainty principles”; the latter came back from Stanford with notes scrawled in blue ink on its margins, likely from Dembo himself, that read, “Ignore this part.”

My eyes had fixed upon a result in the paper showing that for real numbers where , the following two statements are equivalent:

- For all ,

The first statement falls out of the second by setting , and the second falls from the first by showing that this choice of maximizes the right side of the inequality in the second statement. The equivalence above let me rewrite my claim in a way that reduced the act of computing derivatives and showing inequalities from a multipage undertaking to a few lines, and I would be able to simplify the “walking downhill” technique I was using to show the result.

I started typing up the new proof, and when I went to apply one of the lemmas I had proven to show the result, I caught myself. Because of an operation that I was performing, one of the conditions needed to apply the lemma wouldn’t hold. Then I went back to the original proof and noticed that it suffered from the same problem. I had found a bug in the proof, and if I couldn’t resolve it quickly, I would have to retract papers and withdraw from conferences. If I couldn’t resolve it at all, a lot of the results that I had proven over the past several months would no longer hold, including one of the key ones from my Master’s thesis.

My Master’s thesis had focused on a class of problems known as CEO problems, which try to characterize fundamental limits on compressing noisy data from multiple sensors when those sensors communicate to a central estimation unit via rate-constrained links. The abstract of the paper introducing the CEO problem contains a more business-focused exposition:

A firm’s Chief Executive Officer (CEO) is interested in the data sequence which cannot be observed directly, perhaps because it represents tactical decisions by a competing firm. The CEO deploys a team of agents who observe independently corrupted versions of . Because is only one among many pressing matters to which the CEO must attend, the combined data rate at which the agents may communicate information about their observations to the CEO is limited to, say, bits per second. If the agents were permitted to confer and pool their data, then in the limit as they usually would be able to smooth out their independent observation noises entirely. … In particular, with such data pooling can be made arbitrarily small if exceeds the entropy rate of . Suppose, however, that the agents are not permitted to convene, Agent having to send data based solely on his own noisy observations . We show that then there does not exist a finite value of for which even infinitely many agents can make arbitrarily small.

The inspiration for my work had started from a paper by Yasutada Oohama, who had found the sum-rate-distortion function of the quadratic Gaussian CEO Problem.

My goal had been to extend his results to non-Gaussian sources, and while I could show that his sum-rate-distortion function was an upper bound in those cases, I had struggled to find a corresponding lower bound. My hope had been to apply his techniques to derive such a lower bound, but Oohama’s work relied on a specific property regarding the geometry of Gaussian random variables to derive the bound. Specifically, his lower bound followed by tying the orthogonality principle of conditional expectation to statistical independence, which is true for Gaussian random variables, which allowed him to use the entropy power inequality to derive a lower bound.

The proof in which I had just found a bug was an attempt to generalize the entropy power inequality to get around the fact that for non-Gaussian sources, this statistical independence condition would not hold. My attempts to fix the proof or find a counterexample would take me down the rabbit hole of related results and techniques, from Young’s inequality to the Brascamp-Lieb inequality, all because I had just realized the claim I had based part of my identity on was merely a conjecture.

]]>Suppose I were to say, “Red is better than blue.” Alternatively, suppose I were to say, “I prefer red to blue.” In what sense are these two sentences different from one another?

Both of the statements are making a statement from my perspective, so one could argue that if I say, “Red is better than blue,” then I must prefer red to blue. On the other hand, based on the handful of people to whom I’ve posed these sentences, the first one is more likely to invite an argument; to argue the first, one might be able to appeal to some objective knowledge about the world that contradicts the statement, but to argue against the second, one has to know something about my preferences, and one could argue there is no greater authority on my preferences than I.

My friend’s intuitions appeared to match up with what others had said about that example, so I tried to construct an analogue that was related to a statement that isn’t concerned with the opinions of the speaker to see if a larger pattern might emerge.

Suppose I were to say, “The sky is blue.” Alternatively, suppose I were to say, “I read in an encyclopedia that the sky is blue.”

Again, the first sentence is more likely to invite an argument than the second. Taking this along with with the previous example, I wonder if this all comes down to authority. What makes me an authority on the color of the sky? On the other hand, it could be argued that I am an authority on my memories, and therefore to say, “I read in an encyclopedia that…” suggests that I did read the information. Furthermore, it could also be argued that the encyclopedia is a more reliable authority on the color of the sky than I am, and therefore the second sentence is more authoritative than the first.

I’d be curious to find out to what extent these types of sentences have been studied and whether the thoughts others have in any way line up with the intuitions I’ve detailed.

]]>The first step in Puzzled Pint is to figure out where the event will be held. To do this, one solves a location puzzle. In our case, we needed to traverse a maze, and the puzzle was sensitive in the sense that early mistakes propagate to steps later on. It took a while to realize that we had missed a clue, but once it was resolved, we found ourselves on our way.

What’s interesting about a puzzle with this type of propagating error is that it becomes fairly simple to detect that an error was made since the solution is typically an English word or phrase, but not necessarily obvious what the error was, akin to losing an inter frame in video and seeing an error propagate across subsequent frames until the start of the next GOP. The difference between the puzzle and video coding is that given a codec, there is a well understood way to decode the video frames, whereas in the puzzle, the decoding mechanism is intentionally obfuscated and to be deduced along with the solution.

This makes a mistake in a puzzle all the more challenging to debug. For instance, in an early attempt at the February location puzzle, I decoded ROTATERETAMETOYONE, so I was confident that my method was on the right track because the prefix ROTATE is a word, but it wasn’t necessarily clear whether I was applying a correct method and had simply overlooked a piece in the process, or whether the method I was applying was too simplistic to solve the puzzle, and there was something more complex occurring that needed to be applied before moving to the next word.

Ultimately, someone figured out what was missing, and we had a mechanism to validate that the solution we had arrived at was correct. That leads to a second difference between video coding and puzzle solving; namely, it’s sometimes possible to arrive at and validate a solution by sidestepping the expected method the puzzle wants one to apply. For instance, if one gets enough clues that the entropy of the solution is sufficiently low, taking advantage of one’s knowledge of English can provide an alternate mechanism to arrive at the solution instead of figuring out the intended method. In fact, in some cases, one might reverse-engineer the remaining pieces of the method after discovering the solution.

]]>The thing I like about this puzzle is that there’s information encoded in the wording of the problem, and changing one sentence changes the puzzle significantly. “Each carton was mislabeled” indicates that no box contains the correct label, which eliminates 4 out of 6 permutations of labels. If we replace the phrase “each carton was” with “some cartons were” in that sentence, the sentence only eliminates 1 out of 6, and it uncovers a potential ambiguity in how the final sentence of the puzzle is worded, but regardless of how one interprets that sentence, its solution is different from the “each carton was” case. If “each carton was” changes to “the cartons may have been” then the sentence doesn’t eliminate any of the permutations, and depending on how one interprets the wording of the final sentence of the puzzle, it may or may not have the same solution as the “some cartons were” case.

]]>During a workshop about academic careers at Allerton in 2008, Roy Yates responded to the above question in one of my favorite responses to a question of this type: “It was the most socially acceptable way to procrastinate what I wanted to do with the rest of my life.”

Recently, a friend posted on Facebook that she decided to spend time working on a jigsaw puzzle with her daughter when the ISIT 2015 submission deadline was extended. Roy responded to the thread:

I’ve been trying to understand why a puzzle with several hundred pieces is easier to put together than an ISIT paper with a few simple ideas. Each puzzle piece has several possible rotations. The ordering of the pieces would seem to explode combinatorially. You can argue that the picture provides a lot of clues, but I suspect you could do that same 200 piece puzzle upside down, without looking at the picture in a maybe one extra hour. I suppose google could tell me why, but I’d rather tag Lalitha Sankar, Chris Rose, Anand Sarwate, Krish Eswaran and Bobak Nazer.

The question immediately reminded me of a paper I’d encountered during my summer at Xerox PARC back in 2002: David Goldberg, Christopher Malon, and Marshall Bern’s “A Global Approach to Automatic Solution of Jigsaw Puzzles”. The algorithm is for the apictorial case, which corresponds to flipping over all the jigsaw pieces face-down so that the picture doesn’t help.

If one tailors an algorithm based on the picture, the discrepancy between the face-up and face-down cases could be quite severe when compared with the abstraction of Goldberg, Malon, and Bern. For instance, imagine a picture whose pixels follow an even gradient that goes from light to dark vertically and from green to red horizontally. Then by splitting the puzzle pieces by greenness and luminance, one could subdivide the initial set of puzzle pieces into four groups of one fourth the size. In fact, one could continue this subdivision until the number of pieces are small, solve them brute force, creating larger puzzle pieces that can then be solved one level above, similar to a divide-and-conquer approach like the merge sort algorithm, taking roughly steps for -piece puzzles.

The problem with the tailor-made algorithm above is that it doesn’t necessarily generalize to all pictures. One could have a picture that is symmetric, and then any sort of division would result in smaller groups that don’t necessarily merge into a single piece. Similar problems may occur if the puzzle is of a painting by Ad Reinhardt, which effectively reduces the problem to the apictorial case.

While allowances could be made, it’s worth noting that any divide-and-conquer approach cannot automatically take advantage of an algorithm like the one in Goldberg, Malon, and Bern, which attempts to solve the corners first. On the other hand and as noted in the paper, solving the corners, if abstracted in the way that the authors suggest, turns out to be equivalent to a traveling salesman problem and is NP-hard.

It’s unclear to me, however, that the traveling salesman problem is the most useful abstraction. For instance, if one assumes that puzzle pieces only fit together if they fit together in the solution, and there is a unique way in which they do, in the apictorial case, for -piece puzzles, one could test all pieces against each other pairwise in steps to create a graph based on connecting pieces (vertices) that fit together. After that, one could prune the resulting graph one vertex at a time by combining vertices that share an edge at each step, terminating in steps, thereby solving the puzzle. I suspect there is an even more efficient algorithm one could employ in this case, but this is more to illustrate how an alternate abstraction could make the problem simpler.

Ultimately, it appears there is both a sensitivity to how the jigsaw puzzle problem is abstracted as well as how useful the image is. As to whether a jigsaw puzzle is easier to solve than an ISIT paper is to write, it might be worth revisiting that in a week.

]]>We found a strategy that would work, but calculating the probability seemed a little tricky. When I asked Paul, he mentioned that the probability was about 1/3 but that he hadn’t gleaned any special insights from it. Thoughts about the puzzle were quickly replaced with the more common preoccupations of a final-year graduate student: finishing a dissertation and finding a job. The “about 1/3” result entered my head as a folk theorem.

Recently, I posed the puzzle to Jim and Chrysteena, two friends in the San Francisco improv community. Jim designs puzzles and seemed dissatisfied when he evaluated the probability for the candidate solution purported to be the best. I quoted the folk theorem, but I was dissatisfied, as well, and I thought it was time to revisit the puzzle.

There are 100 people and a room with 100 doors. Behind each door is the name of exactly one of the 100 people such that every name is behind exactly one door. The location of the names are unknown to the 100 people. The 100 people are then asked to play the following game:

- Each person is given a chance to go into the room and open 50 doors.
- After a person opens 50 doors, the room is reset to the state it was in before entry, and the person who entered may not communicate with any of the other people for the remainder of the game.
- The game ends in victory if and only if each of the 100 people opens the door with his or her name. In other words, if any one person fails, everyone loses.

If the location of the names behind the doors is chosen at random, find a strategy that maximizes the probability of victory.

It helps to look at cases with a smaller number of doors to build intuition. Let’s first consider two doors and two people, in which each person is allowed to open one door. Note that if both people open the same door, then they are guaranteed to lose since we require that they both find their names. Thus, the best strategy is for them to choose separate doors, at which point the probability of them winning is . Note that this matches the probability that one of them finds his or her name, so it is optimal.

Now let’s move to four doors, four people, in which each person is allowed to open two doors. After some thought, one strategy that would generalize to 100 doors is to arrange the names in alphabetical order and assign a number to each name based on where it is in the ordering, e.g. the first name in alphabetical order is assigned the number 1, and the last is assigned 100. Then, each person starts by opening the door with the number corresponding to his or her name. Upon opening the door, that person will see a new name, which has its own number associated with it in the alphabetical ordering, and will then proceed to open the door associated with this new number. The process repeats until all tries are exhausted.

If one follows this approach, the cases leading to victory are those in which the process of going through the subsequent doors will cycle through to each person’s name within 50 tries in the case of 100 doors, or within 2 tries in the case of 4 doors. For instance, with 4 doors, victory occurs in a case like the following:

Door 1 (Bob), Door 2 (Alice), Door 3 (Dennis), Door 4 (Cindy)

Bob would start at door 2, see Alice’s name and then go to door 1, where he would find his name. Following the approach for the others results in something similar. On the other hand, we fail if we have a case like this:

Door 1 (Bob), Door 2 (Cindy), Door 3 (Alice), Door 4 (Dennis)

Again, Bob would start at door 2, which would give him Cindy’s name and lead him to door 3, which would give him Alice’s name, by which point he has exhausted his two tries without finding his name. Since one person failed, everyone loses.

It turns out that for four doors, there are 10 cases among the 24 permutations that have these smaller cyles, resulting in a victory with probability , which is close to . For 100 doors, there’s the folk theorem.

One thing that has bothered me in the formulation of the puzzle is that the focus is to maximize the probability, but it’s unclear to me whether the candidate strategy above does this. If “about 1/3” continues to hold as the probability of victory as the number of doors increases say even beyond 100, then at least there is a sense that this is a scaleable strategy if not the best since no strategy can exceed 1/2. On the other hand, if the probability decays to 0, it raises a new question to understand why that’s the case, if all strategies decay to a victory probability of 0 as the number of doors increase, or if there exists a strategy that is within some bound of .

We can start asking these questions if we have an expression for the probability of victory. Let’s consider doors, where represents the number of tries, and in the original problem. We have characterized the victory cases as those in which the permutation of doors can be represented as cycles of at most . To calculate the probability, we can use the cycle index of the symmetric group as

,

where denotes cycles of length .

It turns out the cycle index of the symmetric group can be expressed as the recurrence

Let’s define to highlight the dependence on the cycles. Thus, we can rewrite the above equation as

.

Let’s define the cumulative distribution function (CDF) as evaluated at and . In words, is the probability that all cycles are of length or less.

Thus, if we can find an expression for , we can evaluate it at to get our result. The results are contained in Propositions 1 and 2 below, where Proposition 2 is reminiscent of our folk theorem, which yields a probability of , which is slightly less than 1/3. Furthermore, Corollary 1.1 indicates that we decrease monotonically to this probability, so regardless of the number of doors, this strategy is within of , the upper bound on how well one could possibly do.

**Proposition 1**. .

*Proof*: The first equality is by definition, so we simply need to show the second equality. For , we can evaluate the probabilities as

.

Note that if , then , so we can apply the same result to get that

and substituting this into our expression for yields

.

Likewise, one can similarly derive

.

Starting with the base case that and the above inductive steps, one can conclude that

.

However, , so we have our result.

**Corollary 1.1**. .

**Proposition 2**.

*Proof*: From Proposition 1, we can write

. Note that we can rewrite the summation

.

The key is to observe that this is simply a Riemann sum approximating the integral

,

which converges as we let .

]]>