« Back
Generated:
Post last updated:
the alien maths of dath ilan
did you expect a different branch of humanity to have familiar math customs?
Permalink Mark Unread

Continuing from:  Day 24 (20) / Morning Lecture

Permalink Mark Unread

The Rule of Succession!  This is sort of picking up the topic from last time about Science!, how to do statistical inference in cases slightly more complicated than the trivial ones discussed previously.  Asmodia has now supposedly had the time to teach the new candidates some additional Law of Probability, aided by Meritxell.  People have cast a few Fox's Cunnings on themselves and others.  They will now try again to see if that sufficed for old hands to teach new hands.

Suppose Keltham takes a sphere, puts it on a table with bouncy boundaries, shoves it rolling quickly, and lets it bounce back and forth many times until it comes to a halt, somewhere random.  Object!Keltham should theoretically be considered as doing this in private, where the object!others can't see him, though for illustration purposes, meta!Keltham will write down everything plainly for people's meta!selves in class.

[--------0---------------------]

Object!Keltham marks down where Ball 0 landed, picks up the ball, and again rolls it hard.

[--------0------------1--------]

Object!Keltham then asks the object!classroom for their odds that Ball 1 landed to the left or the right of Ball 0.  Remember, their meta!selves can see where Keltham made the marks on the whiteboard representing where Ball 0 and Ball 1 landed, but their object!selves can't.

Before being told the answer, what probability should their object!selves assign that Ball 1 landed to the left or right of Ball 0, after both balls bounced back and forth on the table many times?  Their object!selves know the experimental setup, they just don't know the detailed results until object!Keltham tells them. 

Permalink Mark Unread

.....seems like they should assign a 50% chance? Because the situation is - symmetrical, from the information they have, there's no reason to expect Ball 1 to be more left or more right than Ball 0. 

Permalink Mark Unread

Object!Keltham tells the object!class that Ball 1 was indeed to the RIGHT of wherever Ball 0 landed.  They know nothing else about where Ball 1 or Ball 0 landed, as yet, just that Ball 1 was to the RIGHT of ball 0.

[--------0------------1--------]

Object!class only knows:

1: RIGHT

Ball 2 is now rolled, same setup.

What probability do they assign that Ball 2 falls to the LEFT or RIGHT of Ball 0?

Permalink Mark Unread

 

 

" - okay, I see why it's not just 50% now. Because our guess about where the first ball lands shifts when we learn the second ball landed to the right of it. It's - the same as spinning a coin, where the Ball 0 is like - deciding if the coin is weighted - and then the subsequent spins give you information about how weighted and in which direction."

Permalink Mark Unread

Indeed.  The object!class can now be quite confident that Ball 0 did not stop exactly up against the RIGHT wall, in which case the chance of Ball 1 landing further to the RIGHT would have been zero.

Surely they have no more information than this, however.

Permalink Mark Unread

......nnnnnno they have more information than that, like, it's also less likely the ball is very near the RIGHT wall, it's just not immediately obvious how to plug it into the formulas from last class.

Permalink Mark Unread

"If there were just three positions it could be in, you've ruled out it being at the right end, and if it was in the middle, Ball 1 result was 50/50, if it was on the left end, Ball 1 result was certain, so in that example we should guess it's at the left end twice as often, so 2/3 vs 1/3. So then 2/3 * 2/3 Ball 2 is to the right, plus 1/3 *1/3 Ball 2 is to the right, so 5/9."

Permalink Mark Unread

"I think I see what you were trying to do there but I don't think the numbers work out?  Ball 0 could be all the way on the left, in the middle, or on the right.  The new chance that it's all the way on the right is zero, and if Ball 0 was in the middle, we're half as likely to hear 'RIGHT' than if Ball 0 was all the way on the left.  So the new chances are 2/3 that Ball 0 was all the way on the Left, times a probability of 1 of hearing 'RIGHT' then, plus a 1/3 chance that Ball 0 was in the middle, times a 1/2 probability of hearing 'RIGHT' then.  2/3 * 1 + 1/3 * 1/2 = 5/6 chance of hearing 'RIGHT' next time."

Permalink Mark Unread

"If, however, we actually look at what - 'meta-Keltham' - drew on the wall, it seems like a pretty reasonable way things could be, and in that way things could be, we're not supposed to say 5/6."

"I mean, think about it, if this actually happened, and Keltham said 'RIGHT' one time, would you immediately afterwards go, 'Well, I guess the chance of hearing "RIGHT" next time is 5/6?'  Hearing 'RIGHT' one time shouldn't actually make you that confident about it happening again."

Permalink Mark Unread

"It's still information, though. In all of the situations where ball 0 is left of center, ball 1 was more than 50% likely to be right of it, and in all of the situations where ball 0 is right of center, ball 1 is less than 50% likely to be right of it. So the fact that it's to the right is some update towards ball zero being left of center, which suggests that ball two is also more likely to be to the right of ball zero, but - I don't know how you'd tell what the numbers are."

Permalink Mark Unread

"Let's say we divide it into eleven possible positions.  If Ball 0 arrives all the way on the right, chances of hearing RIGHT are 0/10.  If Ball 0 is 90% of the way to the right, chances of hearing RIGHT are 1/10."

"So it's 1/10 squared plus 2/10 squared plus 3/10 squared which is... 1 plus 4 plus 9 plus 16..."

"I get 285/100."

"What'd I do wrong."

Permalink Mark Unread

"Forget to multiply each of those terms by their 1/11 prior probability."

Permalink Mark Unread

"285/1100 chance of hearing 'RIGHT' next time."

Permalink Mark Unread

"Pretty sure we should expect to hear 'RIGHT' more than we expect to hear 'LEFT'."

Permalink Mark Unread

"Let us, possibly, try this with five positions before going straight to ten."

"0/4, 1/4, 2/4, 3/4, 4/4 likelihood of yielding RIGHT, all positions equally probable before then."

"We see one 'RIGHT'.  The likelihoods are 0/4, 1/4, 2/4, 3/4, 4/4.  The posteriors are - I see the other two things Pilar did wrong."

"We start with five hypotheses 0/4, 1/4/, 2/4, 3/4, 4/4, each of which has prior probability 1/5."

"After seeing 'RIGHT' one time, we've got 'priors-times-likelihoods' of 0/20, 1/20, 2/20, 3/20, 4/20."

"The 'relative-odds' of the posterior will be proportional to the 'priors-times-likelihoods', but they have to sum to 1.  There's 1 + 2 + 3 + 4 = 10 total shares, of twentieths but it doesn't matter what exactly, so the 'posterior-distribution' is 0, 1/10, 2/10, 3/10, 4/10, summing to 1."

"Then if you take the likelihoods off each of those, weighted by the new posterior-probability, you get 1 + 4 + 9 + 16 divided by 40 equals 3/4."

Permalink Mark Unread

Keltham will helpfully write down the calculation Asmodia actually did, there:

0/10 * 0  +  1/10 * 1/4  +  2/10 * 2/4  +  3/10 * 3/4  +  4/10 * 1
=  1/40 + 4/40 + 9/40 + 16/40
=  30/40
Permalink Mark Unread

"With 7 positions it's 0 + 1 + 4 + 9 + 16 + 25 + 36, divided by (0 + 1 + 2 + 3 + 4 + 5 + 6) * 6, so 91/126."

Permalink Mark Unread

"Well, how about infinity positions then?"

Permalink Mark Unread

Permalink Mark Unread

Permalink Mark Unread

"Are you, in fact, allowed to just do that?"

Permalink Mark Unread

"Well, I don't see who's going to arrest you if you just try."

Permalink Mark Unread

 

...all... right then...

 

"I get 1^2 + 2^2 + 3^2 out to infinity, divided by... 1 + 2 + 3 out to infinity, times infinity."

Permalink Mark Unread

"Great, now simplify that expression."

Permalink Mark Unread

...one of the older students in the background does happen to know the useful fact that summing from 1 to N gives you N(N +1)/2, and summing the squares from 1 to N gives you N(N+1)(2N+1)/6.

Parenthetical retroactive class chatter

      Somebody asks if "summing 1 to N gives N(N +1)/2, and summing squares gives N(N+1)(2N+1)/6" are to be treated as brute facts or the sort of things you can prove?


...

...Keltham is not usually one to duck into details that aren't relative to the main line of argument, but this question is sort of alarming.

Keltham will observe that [1 + 2 + 3 + 4 + 5 + 6] = [(1 + 6) + (2 + 5) + (3 + 4)] = [7 + 7 + 7], and that [1 + 2 + 3 + 4 + 5] = [(1 + 5) + (2 + 4) + 3] = [6 + 6 + 6/2].  Proof by generalizing examples:

   sum of [1 + 2 + ... + N] with N terms
= half that many [(N + 1)] terms, so N/2 * (N+1)
= N(N + 1) / 2.

Keltham doesn't really want to stop and prove the sum of n^2 from 1 to N is N(N+1)(2N+1)/6 using properly hasty intuitive reasoning that will create a good sense for what's really going on.  So instead he'll just quickly write on the board that:

let   f(N) = N(N+1)(2N+1)/6

in f(0) = 0
f(N+1) = f(N) + (N+1)^2

proof:

0(0+1)(2(0)+1
= 0(1)(3)
= 0

f(N+1) - f(N)
= ([N+1])([N+1]+1)(2[N+1]+1)/6  - N(N+1)(2N+1)/6
= (N+1)(N+2)(2N + 3)/6 - N(N+1)(2N+1)/6
= (N+1)  *  [(N+2)(2N + 3)  -  (N)(2N+1)]  /  6
= (N+1)  *  [2N^2 + 7N + 6 - 2N^2 - N]  /  6
= (N+1)  *  [6N + 6]  /  6
= (N+1)^2

If they're not happy with this logical proof by certain reasoning - because, like, why - they should consider visualizing stacking a bunch of n-sided squares on top of each other, to form a figure that pretty much looks like it should take up about a third of a cube of size N.

Now, back to the problem of evaluating:

[1^2 + 2^2 + 3^2 + 4^2 + ...] / { [1 + 2 + 3 + 4 + ...] times infinity }

Permalink Mark Unread

Then Pilar will obtain the sum N(N+1)(2N+1)/6, divided by N*[N(N+1)/2].

Letting N = infinity, this yields INF*(INF+1)*(2INF+1)/6, divided by INF*INF*(INF + 1)/2.

Simplifying produces (2INF+1)/(INF*3).  This then cannot be simplified further.

Permalink Mark Unread

Why, of course it can be!

Keltham writes on the wall:

2*INF + 1     2 * INF      1
--------- = ------- + -------
INF * 3 INF * 3   INF * 3

1 divided by three infinities simplifies to zero, obviously.  Think of how little cake each person gets, if you divide one slice of cake among all the billion people in Golarion; one-third divided by infinity is much less than that.

So this whole expression simplifies to 2/3.

Permalink Mark Unread

 

"Is that actually Lawful reasoning?"

Permalink Mark Unread

It's Lawful when he does it, in a way that carefully doesn't take any mathematical reasoning steps that would actually be invalid.  Keltham would not necessarily advise Pilar trying this on her own for a little while.  Or rather, she should try it, but be prepared to fail.

Pilar may, for her own reassurance, note that her previous answer of 91/126 is not too far from 90/135, also known as 2/3.

Permalink Mark Unread

"...can’t you just think of this as permutations of 3 things?"

Permalink Mark Unread

"Hmmm.  Can you?"

Permalink Mark Unread

"There are 3 permutations which have 0 before 1:  012, 021, 201."

"Of these, 1/3 have 2 left of 0 and 2/3 have 2 right of 0."

Permalink Mark Unread

"So what you're saying here is that any of the possibilities..."

   [---2----0------------1--------]
or [--------0------2-----1--------]
or [--------0------------1-----2--]

"...are equally probable?  I'm not sure how reasonable that is, it looks to me like the middle stretch is wider than the two ends."

Permalink Mark Unread

"We don't know that!  In the - 'object-level' - classroom!"

Permalink Mark Unread

"So what you're saying here is that if I roll three balls back and forth, and they bounce back and forth a lot, they're equally likely to land in any of the six possible orders, so far as you know?"

"Can that really be true?  After all, there could be bumps in the table, the bouncy walls might not be perfect, I might have rolled the balls harder on some occasions than others..."

Permalink Mark Unread

"If we don't know any of that on the 'object-level', then, so far as we know, all six permutations are equally likely."

"You could say that sort of thing about any problem with permutations in it."

Permalink Mark Unread

"Oh, I suppose.  Just don't forget, in real life, always cheat the shit out of anything if you can get any extra information like that."

"After all, if you think otherwise, why, that's the sort of thinking that might lead to you always insisting that all of the chemical elements were equally likely to be in any order on the Periodic Table, any time you don't know for sure what order they have."

"Soooo how about if I then tell you that Ball 2 landed to the RIGHT of Ball 0, again?  What's the probability of RIGHT next time, with the third ball?"

Permalink Mark Unread

"3/4," says the same person who asked about permutations last time.  (They've unfortunately written their nametag in letters too small to see from where Keltham is standing.)

Permalink Mark Unread

"This one comes up LEFT, though.  Your prediction FAILED.  All your theories were probably WRONG."

Permalink Mark Unread

"No, we just ran into a 1/4 chance, that happens sometimes.  Like with the chance of failure on an Augury."

Permalink Mark Unread

"I suppose.  We should just discard the evidence, then, since it's only to be expected that reality gives us the wrong result, sometimes.  You predict 3/4 RIGHT again next time?"

Permalink Mark Unread

You can, in fact, get 10-foot range on Prestidigitation, though not when you're a wizard as fresh as Keltham.  One of the students nearest that wall makes their own change to it without leaving their desk.

[----3---0------------1-----2--]

"3/5 right, 2/5 left," they declare.

Permalink Mark Unread

What if the 2-ball actually landed to the left of the 1-ball?  They're not told which of those was true!

Permalink Mark Unread

...if those two possibilities are equally probable, then they can sum the probabilities, right?

1/2 chance of 3/5 plus 1/2 chance of 3/5 equals 3/5.

Permalink Mark Unread

You don't need to label them!

[---|---0---|---|---]

"That's five landing regions, two to the left and three to the right."

Permalink Mark Unread

Does anyone want to object to the invalidity of that reasoning -

- or alternatively, state a general rule for what to predict on the next round, if you previously got N balls LEFT and M balls RIGHT of the 0-ball, and that's all you know?

Permalink Mark Unread

"If you have N balls LEFT, there are N+1 landing regions LEFT, and M balls RIGHT gives M+1 landing regions RIGHT, so my first thought is (N+1)/(M+1)."

Permalink Mark Unread

Ah, yes, so if they have seen one ball go LEFT, and zero balls go RIGHT, the chance of seeing it go LEFT on the next round is (1+1)/(0+1)=2.

So the probability of seeing LEFT next time is 2.  That sounds pretty reasonable to Keltham?  Though clearly anthropics were involved at some point.  Maybe the balls are duplicating people once they start to go LEFT?

Permalink Mark Unread

They obviously meant, like, a ratio of (N+1):(M+1), not to calculate the probability that way!

Permalink Mark Unread

So, like... after seeing one ball go right, the chance of seeing it go left on the next round, is half the chance of seeing it go right.  That's what the (N+1)/(M+1) quantity is, how many times more likely LEFT is than RIGHT, in this case, (0+1)/(1+1) = 1/2.

Permalink Mark Unread

But HOW do they square that up with the previous CLAIM as proven by DUBIOUS INFINITARY MATHEMATICS that, after seeing one RIGHT, the chance of the next ball landing RIGHT is two-thirds?

Permalink Mark Unread

...it's the same claim.  One-third left is half of two-thirds right.

Permalink Mark Unread

But is it really.

Permalink Mark Unread

YES.

Permalink Mark Unread

Somebody would in fact now like to make the general claim that, if N balls go LEFT and M balls go RIGHT, the chances of the next ball landing LEFT and RIGHT respectively are:

  N + 1           M + 1
--------- vs. ---------
N + M + 2       N + M + 2
Permalink Mark Unread

Keltham agrees!

Now would anybody like to try to prove that claim using dubious infinitary mathematics?

Permalink Mark Unread

"I'd love to, but I don't know where to start... well, actually, I guess the obvious place would be with - two RIGHT and one LEFT and seeing if we can make that come out to 2/5 left and 3/5 right and then seeing if we can get it to generalize."

Permalink Mark Unread

Several of the new students do in fact know calculus, and that seems like the obvious tool to use on this problem?

Permalink Mark Unread

Ah, yes.  Golarion's notion of 'calculus'.  Keltham has actually looked into it, now.

It looked like one of several boring, safe versions of 'dubious-infinitary-mathematics' where you do all the reasoning steps more slowly and they only give you correct answers.

Dath ilani children eventually get prompted into inventing versions like that, after they've had a few years of fun taking leaps onto shaky ground and learning which steps land well.

Those few years of fun are important!  They teach you an intuitive sense of which quick short reasoning steps people can get away with.  It prevents you from learning bad habits about reasoning slowly and carefully all the time, even when that's less enjoyable, or starting to think that rigor is necessary to get the correct answer in mathematics.

Permalink Mark Unread

Pilar has not yet been exposed to professional Golarion mathematicians talking about the importance of rigor.  It will probably be a powerful theological experience for her when she does!

Meanwhile she already has a strong sense that Keltham needs to be burned as a heretic.

"Don't I need to reason correctly to arrive at correct answers?"

Permalink Mark Unread

"I don't think we have several years to spend on freeform mathematical exploration, even if we should make it an educational priority for the next generation."

Permalink Mark Unread

"Rigor is necessary to know you got the correct answers.  Nonrigorous reasoning still often gets you correct answers, you just don't know that they're correct.  The map is not the territory."

"Often though not literally always, the obvious methodology in mathematics is to first 'rapid-prototype' an answer using nonrigorous reasoning, and then, once you have a strong guess about where you're going, what you're trying to prove, you prove it more rigorously."

"We don't have years to spend doing all of our mathematical reasoning rigorously the first time around, and besides, who does that anyways, and also, I'm here and can potentially check whether your fast crazy reasoning steps were valid or not."

Permalink Mark Unread
Parenthetical class chatter

Somebody is currently trying to solve the case for one LEFT, two RIGHT, by visualizing how you'd carve off pieces of a five-dimensional hypercube... no, actually just a four-dimensional hypercube...

This is apparently the sort of thing you learn in math class as a relatively advanced wizard?  Though you don't need it to hang most second-circle spells.

Their first attempt to simplify down to zero LEFT, one RIGHT and a 2-dimensional triangle doesn't end well for them, though.

Permalink Mark Unread
Parenthetical class chatter

Other researcher-candidates seem to mostly be trying to rework Pilar and Asmodia's original workings with the five possible positions and the seven possible positions, never mind the dubious infinitary mathematics after that.

Permalink Mark Unread

Okay, you know, Asmodia's just going to try this thing with the dubious infinitary mathematics.  If not her, then who?  Asmodia alone is protected from any vengeance the afterlife might visit on her.

Consider the case for one LEFT and two RIGHT first.  They start with an infinite number of points between 0 and 1 probability of yielding LEFT, each with prior probability 1/INF.

For each of those points p on the spectrum from 0 to 1, their likelihood of yielding one LEFT and two RIGHT is:

(p)^1 * (1-p)^2  =  p - 2p^2 + p^3

Then to get the posterior... Asmodia has to...

Well, to be honest, she has in fact tried to get calculus off some Securities and she is having to exert something of a deliberate effort not to use that.  But the calculus also wasn't - the thing that she can feel Keltham is trying to teach?  Asmodia is getting a sense of the way that Keltham teaches intuitions and not just formulas.  She thinks it's meant to teach the thing that Ione does a little better than Asmodia.

A little better.

Asmodia wants to learn it.

Permalink Mark Unread

The way they got here before, was by Pilar trying it with five exact hypotheses for p, then seven hypotheses for p, then an infinite number of hypotheses for p.

Let's say there were four hypotheses, 0/3, 1/3, 2/3, 3/3 probability of LEFT, and you saw one LEFT and two RIGHT.

Then you'd sum up - 1/4 * 0 + 1/4 * (1/3)^1*(2/3)^2 + 1/4 * (2/3)^1*(1/3)^2 + 1/4 * 0.

That would give you the posterior distribution of - some stuff just cancels out there, actually - 1^1*2^2 + 2^1*1^2 - and then after dividing through by that -

Actually, a lot of this stuff is going to cancel out anyways, isn't it?

But finish the calculation anyways.

After that, we get the 1/3 hypothesis with posterior probability 4/(4+2), and the 2/3 hypothesis with posterior 2/(4+2).

Then to get the prediction of LEFT next time, it's 2/3 probability on the 1/3 hypothesis which contributes 2/9 to LEFT, and 1/3 probability on the 2/3 hypothesis which contributes 2/9 to LEFT, and it comes out as 4/9 total prediction for LEFT.  Which does at least match that they originally saw more RIGHT than LEFT.


Now do the infinitary version.

Permalink Mark Unread
Parenthetical class chatter

Meanwhile, one of the newer researchers has worked out that, in the earlier formula:

0/10 * 0  +  1/10 * 1/4  +  2/10 * 2/4  +  3/10 * 3/4  +  4/10 * 1

...that first term 0/10 * 0 has the 0/10 subterm representing "The probability that the fraction of RIGHTs is zero, given that the first ball landed RIGHT" and the 0 subterm representing "The probability that ball 2 lands RIGHT, given that the fraction of RIGHTs is zero."

Permalink Mark Unread

Okay yeah, if she's not just allowed to throw calculus at it, or take limits in a way that just rederives Golarion-style calculus, Asmodia isn't really seeing how the dath ilan dubious-infinitary-reasoning version works.

Permalink Mark Unread

Ione already knows the answer is N+1 / (N + M + 2) via a perfectly valid combinatorics argument!  She shouldn't have to rigorously prove things when she already knows the correct answer, or even prove them using dubious infinitary arguments!  This is Nethysian abuse!  Aren't only Asmodeans supposed to treat people like this?  She's sure it's not how Nethysian math classes work.

Permalink Mark Unread

Keltham starts by considering the case where N balls go LEFT and 0 balls go RIGHT.

Then every point lying a fraction p along the way from the LEFT side to the RIGHT side, has a likelihood p^N of yielding the data we saw.

Retroactive parenthetical class chatter

One student is having some trouble parsing this, despite Keltham's gainful attempts to speak slowly.  The p is a position along a line from 0 to 1, and the p^n is 'if the 0-ball landed at that point, how likely it is that it yielded the data we saw'.  But p=0 at the far left, shouldn't it be infinitely likely it be to the left of the furthest-left point instead of no chance whatsoever -

A moment later, they've worked out that, no, the next ball has probability 0 of being any further left than as far left as possible, and the whole thing flips around in their head.

But by then, of course, Keltham has already continued talking.

The prior probability is the same everywhere, so the posterior function will be proportional to p^N everywhere.  If you wanted the whole thing to sum to 1, you'd divide every point by whatever p^N sums to between 0 and 1.

Retroactive parenthetical Alexandre intertag
        "... Excuse me." Alexandre raises his hand, since showing weakness in this environment appears to be less bad than possessing it. 'Divide every point by whatever p^N sums to between 0 and 1' - can I ask you to expand on that?"


Keltham shall be happy to do so!  Thank you for providing any feedback on where anyone is getting lost.

As a simple example, take the case of N=1, so p^N=p.  If you graph that, you get a straight line from likelihood=0 at p=0, to likelihood 1 at p=1.  That graph looks like a triangle filling half of a square with side 1.  So by visual reasoning, the sum 1/INF * [0/INF + 1/INF + 2/INF + ... + INF/INF] = 1/2 in the end.

Proof by dubious infinitary mathematics, now that they know the correct answer by simpler visual reasoning:

1 + 2 + ... + INF  =  INF*(INF+1)/2
=>  1/INF * [0/INF + 1/INF + 2/INF + ... + INF/INF]
= 1/INF * [1/INF * { 1 + 2 + ... + INF }]
= 1/INF * [1/INF * { INF*(INF + 1)/2 }]
= INF * (INF + 1) / (INF * INF * 2)
  = 1/2

Then to find a probability distribution proportional to that prior-times-likelihood object, but which sums to 1 to form the posterior, they divide the whole thing by 1/2 or multiply it by 2.

1/INF  *  [0/INF + 1/INF + ... + INF/INF]  /  (1/2)   = 1.


As Keltham is saying this, it occurs to him that it may not be obvious to the candidates why you'd do that, depending on what got covered in previous Probability classes and how well.  He'll expand on that too.

If they were just tracking, say, 1/2 prior probability of propensity 1/3 versus 1/2 prior probability of propensity 2/3, the prior-times-likelihood object would look like {1/2 * 1/3, 1/2 * 2/3} and the sum over the prior-times-likelihood object would look like {1/6 + 1/3} = 1/2.  Which is just what it should be, because if you start with 50-50 on 1/3 vs. 2/3 then your prior probability of seeing the ball go LEFT is indeed 1/2.

Conversely, the prior-time-likelihood object for seeing RIGHT is {1/2 * (1 - 1/3), 1/2 * (1 - 2/3)} = {1/2 * 2/3, 1/2 * 1/3} = {1/3, 1/6} again summing to 1/2.

This should look like it makes lots of sense.  On priors, you think either the ball goes left, or the ball goes right, but not both, nor neither, so your prior probability of LEFT plus your probability of RIGHT should sum to 1.

But after actually seeing the ball go LEFT and updating on that, you throw out all the possible worlds from your probability distribution where the ball went RIGHT, and live only inside the worlds where the ball went LEFT.  After seeing LEFT, then, within this kind of hypothesis where you're not questioning the whole setup from outside, you're certain that the ball went LEFT.  It's now probability 1 that the ball went LEFT.  The probability 1, of what is real, is divided up among the surviving worlds; the surviving worlds should have probability proportional to their prior times their likelihood, but summing to 1.

So you divide through {1/6, 1/3} by its total probability of 1/2 to get {1/3, 2/3} as the posterior probability.

In symbolic terms, there's:

P("propensity 1/3") = 50%
P("propensity 2/3") = 50%
P(TRUE) = P("propensity 1/3") + P("propensity 2/3") = 1
P(LEFT ◁ "prp 1/3") = 1/3
P(LEFT)
  = P(LEFT & "prp 1/3") + P(LEFT & "prp 2/3")
  = P(LEFT ◁ "prp 1/3")*P("prp 1/3") + P(LEFT ◁ "prp 2/3")*P("prp 2/3")
  = 1/3*50% + 2/3*50% = 1/2
P("prp 1/3" ◁ LEFT) = P("prp 1/3" & LEFT) / P(LEFT) = 1/6 / (1/2) = 1/3
P("prp 1/3" ◁ LEFT) + P("prp 2/3" ◁ LEFT) = P(TRUE) = 1


So to get the posterior probability on all the hypotheses between 0 and 1, after seeing N balls go left, they take the prior-times-likelihood object, which is 1/INF * [ (0/INF)^N + (1/INF)^N + ... + (INF/INF)^N ], and divide it through by its sum so that the probability inside the whole thing sums to 1.  If we started out certain that the ball has a propensity between 0 and 1 to go LEFT or RIGHT, and then we observe N LEFTs, we end up still certain that the ball has some propensity between 0 and 1 to go LEFT.

So the question then becomes:  1/INF * [ (0/INF)^N + (1/INF)^N + ... + (INF/INF)^N ] = ?

Now, before they try to prove a correct answer, they should, of course, first figure out what the correct answer is, so they can prove it.

The correct answer in this case obviously has to be that it sums to 1/(N+1).

Why?  Well, you saw N balls go LEFT.  On the first round, that had probability 1/2.  On the second round, you thought there was a 2/3 chance of seeing LEFT, and then you saw it, so your joint probability over all predictions is now 1/2 * 2/3 = 1/3 after seeing two LEFT.  On the third round, you think there's a 3/4 chance of seeing LEFT, so your joint probability goes to 1/2 * 2/3 * 3/4 = 1/4 after seeing three LEFT.

Since your joint probability of seeing N balls go LEFT is 1/(N+1), the sum from 0 to 1 of an infinite number of points each with weight p^N for how much they predicted seeing N balls going LEFT, must clearly end up as 1/(N+1).

Permalink Mark Unread

Pilar is feeling, on some deep instinctive level, that her soul may be in danger here.  This is not how math works.

Permalink Mark Unread

It's fine!  They derived that the final answer has to be 1/(N+1) from combinatoric reasoning about the probability, which is simple and reliable, so they now know that's where any dubious infinitary reasoning about the same probability ought to end up.

Permalink Mark Unread

The problem before them, then, is to sum up:

1/INF * [ (0/INF)^N + (1/INF)^N + (2/INF)^N + ... + (INF/INF)^N ]

To be clear on what this is intended to mean:

The terms 0/INF, 1/INF, 2/INF up to INF/INF are an infinite number of distinct hypotheses covering the spectrum from 0 to 1 of LEFT frequencies;

Raising a hypothesis for a frequency between 0 and 1, to the power of N, gives that hypothesis's likelihood for generating the observed data of N LEFTs;

And the 1/INF term at the start, is being factored out from every hypothesis in the sum having 1/INF prior probability.

Where, again, the answer must be 1/(N+1).

Permalink Mark Unread

 

"...am I getting it right that the contribution from (2/INF)^N is - obviously just 0?"

Permalink Mark Unread

Obviously.  If you pick any finite number, you're starting infinitely close to a 0 probability of generating any LEFTs.

It's only once you start getting a finite fraction of the way towards infinity, that dividing by infinity won't just yield 0 again, intuitively speaking.

Don't worry, only 0% of the numbers between 0 and INF are finite, so it makes sense that they'd contribute 0% of the likelihood.

Permalink Mark Unread

"Well, if I had any doubt before that you came from some plane of existence completely alien to Golarion, it's gone now."

This is not how Golarion mathematics works!!!

Permalink Mark Unread

Dath ilan has many different ways of looking at ideas like "the integers" or “the points between 0 and 1”, and teaches its children to switch between them as appropriate, especially when they are doing rapidly-prototyped mathematics.

Permalink Mark Unread

The problem before them is to sum up:

(0/INF)^N   (1/INF)^N   (2/INF)^N         (INF/INF)^N
--------- + --------- + --------- + ... + -----------
INF INF INF INF

...and get the correct answer of 1/(N+1).  This obviously reduces to summing:

(0/INF)^N + (1/INF)^N + (2/INF)^N + ... + (INF/INF)^N

...and getting INF/(N+1).

(Retroactive parenthetical class chatter)       "So..." A hand goes up. It is the hand of a student who does not want to admit weakness, but will do so anyway, under the eyes of Security. "It isn't obvious to me how that reduces?"

Keltham starts to answer -

      Oh wait never mind it is obvious.

      It's too late to say that now, though, so the student will just listen to Keltham speak the awful words that he's just multiplying both the original sum and the original result 1/(N+1) by infinity.

Factoring out (1/INF)^N from all the terms (n/INF)^N reduces the problem to one of summing:

0^N + 1^N + 2^N + 3^N + ... + INF^N

...and getting INF^(N+1) / (N+1).

(Retroactive parenthetical class chatter)       A hand goes up, in obedience to stated Orders. "Why are we factoring this out, sir? And why does it require summing this?"


...they're taking (0/INF)^N + ... and factoring out (1/INF)^N from all terms (n/INF)^N to yield n^N * (1/INF)^N, for each term, and then dividing both sides to yield the sum over n^N equalling [ INF/(N+1) / (1/INF)^N ] = INF^(N+1) / (N+1).

Then, once they get the solution for an infinite sum of cubes n^N, they'll multiply that whole thing through by (1/INF)^N and by (1/INF), and get back out...
(0/INF)^N   (1/INF)^N   (2/INF)^N         (INF/INF)^N
--------- + --------- + --------- + ... + -----------
INF INF INF INF
...which represents an infinity of possible hypotheses, for all possible propensities between 0 and 1, for how often a ball goes LEFT, each hypothesis with 1/infinity prior probability.

Look, it's just the same as if they'd used four buckets each with 1/4 prior probability, and asked for the chance of seeing 9 balls going LEFT, that would be:
(0/3)^9   (1/3)^9   (2/3)^9   (3/3)^9
------- + ------- + ------- + -------
4 4 4 4


      ...Somebody then asks whether it ought to be INF+1 in the denominator of the original expression.


Keltham stares at the wall for a moment, then sagely announces that, with a sum this large, it won't make any relative difference if they drop any finite number of terms from it, since those contribute 0% of anything.  He's just going to arbitrarily declare that one of the ones in the middle is missing.

Permalink Mark Unread

Now in an intuitive sense, INF^(N+1) / (N+1) is obviously what the answer has to be, of course.

If you stack up a bunch of one-dimensional lines of length 1, 2, 3, ... 10, you end up with roughly half of a two-dimensional square of side 10.

If you stack up a bunch of two-dimensional squares each with sides 1, 2, 3, ..., 10, you end up with roughly a third of a three-dimensional cube of side 10.

So naturally if you stack up a bunch of N-dimensional hypercubes with sides 1, 2, 3, ... infinity, you'll end up with exactly 1/(N+1) of an (N+1)-dimensional hypercube with side infinity.

(Retroactive parenthetical class chatter) (This nerdsnipes multiple researchers hard enough that they have trouble following the next few sentences, which thankfully aren't vital ones.)
Permalink Mark Unread

Actually, they might possibly want to check that on the numbers up to, like, 5, or something.

1 + 2 + 3 + 4 + 5 = 15 which is like not that far off from half of 25.

1 + 4 + 9 + 16 + 25 = 55 which is not that far off from 1/3 of 125.

1 + 8 + 27 + 64 + 125 = 225, not far from 1/4 of 625.

It's a bit more each time, since there are extra terms in the sum beyond the one for SIDE^(N+1) / (N+1).  But as you sum a larger number of terms, the extra terms grow less quickly, and they can be ignored when the sums are infinite.

Like 1 + 2 + 3 + ... + 100 = 5050 and that's pretty close to half of 10,000.

Permalink Mark Unread

Someone is increasingly confused about how this INF business works.  Is INF - 1 a different number from INF?  Is INF-(INF-2)+3 equal to 5?

Permalink Mark Unread

Yes to both, obviously.

The usual guideline for answering questions like that is to pretend that INF is a very large finite number?  The whole idea is that a bunch of things that are approximately true about large numbers become exactly true about INF.

Permalink Mark Unread

...how does that square with 2/INF being 0?

Permalink Mark Unread

2 divided by a very large number is approximately 0.

So 2/INF is exactly 0, so long as it isn't being multiplied by infinity somewhere outside that.

Similarly, (INF-1)/(INF+1) = INF/INF = 1.

Permalink Mark Unread

Is there some precise sense of "approximately" that could be used to prove -

Permalink Mark Unread

That sounds like a question SLOW people would ask!  Boring question!  Let's move on!

Permalink Mark Unread

"Keltham, has it ever occurred to you that, as an eight-year-old boy, you were taught mathematics in a style that someone thought would appeal to eight-year-old boys?"

She doesn't actually say it out loud.  Maybe when they're alone.

Permalink Mark Unread

Now, the nice thing about working with infinities is that you don't have to be very precise at all, and you'll still get exactly the right answer at the end!

Instead of taking x^N, let's use the 'rising powers', okay that didn't translate so probably is not a thing in Golarion, or at least the Share Language donor didn't know it.

Instead of taking x * x * x * ... * x, with N copies of x, let's consider the expression x * (x + 1) * (x + 2) ... until you've got N of those.

Let's denote that x`N`.

Obviously if x is any finite fraction of the way towards infinity, x`N` = x^N exactly.  Like, when you're working with trillions, a trillion squared is nearly the same number as a trillion times a trillion plus one, proportionally speaking; that's approximately true, so with infinities it's exactly true.

Example with x=10, N=3:

10`3`
= 10*11*12
= (10)(10+1)...(10+(3-1))    (3 terms total)
= (x)(x+1)...(x+(N-1))          (x`N` with x=10, N=3)

Now observe that:

11`3` - 10`3`
= 11*12*13 - 10*11*12
= (13 - 10) * (11 * 12)
= 3 * 11`2`

By pointwise generalization from 10 to x and 3 to N:

   (x + 1)`N`  -  (x)`N`
= (x + 1)(x + 2)...(x + N)  -  (x)(x+1)...(x + N - 1)
= ((x + N) - x) * (x+1)...(x + N - 1)
= N * (x+1)`N - 1`

Rearrange:

(x + 1)`N`  -  (x)`N`   =  N * (x+1)`N - 1`
    ==>
x`N`  +  N*(x+1)`N-1`   =   (x+1)`N`

We now have an N-1 term in our rule, which would be bad if N were 0.  Eliminate all subtraction whenever possible!  In case something ends up 0 or negative that didn't start out that way!

In this case we'll just redescribe the number N as being some other number N plus one, so N-1 goes to "N" and N goes to "N+1".  We coulda done that originally, but it woulda been needlessly more symbols.

x`N`  +  N*(x+1)`N-1`   =   (x+1)`N`
    ==>
x`N+1` + (N+1)*(x+1)`N` = (x+1)`N+1`

Sanity check with x=10, N=2:

        x`N+1` + (N+1)*(x+1)`N` = (x+1)`N+1`
=>   10`2+1` + (2+1)*(10+1)`2` = (10+1)`2+1`
=>   10*11*12 + 3*(11*12) = 11*12*13
PASSED

Then, by induction on f(x) = x`N+1`:

0`N+1`  =  0    (base case)

(x+1)`N+1`  =  x`N+1` + (N+1)*(x+1)`N`

    ==>

 x`N + 1`  =  (N+1) * 1`N`  +  (N+1) * 2`N` + ... (N+1) * x`N`
  (can optionally add a starting term of (N+1)*0`N+1' since this always equals zero)

Sanity check with x=3, N = 2:

(3)(4)(5)  =  3(0)(1) + 3(1)(2) + 3(2)(3) + 3(3)(4)
60  =  0 + 6 + 18 + 36
PASSED

Divide both sides by N+1:

x`N+1` / (N + 1)  =  1`N` + 2`N` + 3`N` + ... + x`N`

Approximations are exactly correct at infinity:

(INF)`N` = (INF)^N

...and this is also exactly true once you get any substantial fraction of the way towards infinity.  Like, even (1,000,000)`3` starts to look pretty close to (1,000,000)^3.

As for all the finite numbers at the start of the sequence, where the equality isn't exact, they collectively comprise 0% of anything and can be ignored.

Conclusion:

0^N + 1^N + 2^N + 3^N + ... + INF^N = INF^(N+1) / (N+1).

So if you stack up an infinite number of squares with sides 0, 1, 2, 3... you will get exactly a third of an infinitely large cube, and more generally, if you stack up an infinite number of N-hypercubes with sides 0, 1, 2, 3... you'll get exactly 1 / (N+1) of an infinite (N+1)-hypercube.

Permalink Mark Unread

Which, in case anybody's forgotten by now, was what they needed to prove that:

1/INF * [ (0/INF)^N + (1/INF)^N + (2/INF)^N + ... + (INF/INF)^N ]  =  1/(N+1)

...which in turn is how you:

- Represent an infinite number of possible hypotheses, about how far from 0% LEFT to 100% LEFT the first Ball 0 could have landed;
- Each hypothesis with equal prior probability 1/infinity;
- And then update on seeing N balls go LEFT;
- As would have a likelihood of p^N for each hypothesis a fraction p of the way between the left side and right side.

Also in case anybody's forgotten, if you imagine N balls plus the 0 ball getting randomly ordered, the chance of the 0 ball coming rightmost in the ordering - hence, all other balls being LEFT of it - is 1/(N+1).  Which is how they originally knew that 1/(N+1) was the desired end result, for the likelihood of seeing all N balls land on the LEFT.

Permalink Mark Unread

Also also in case anybody's forgotten, the reason they needed this sum was to renormalize the relative probabilities into absolute probabilities.  That's another step along the direct path to deriving a prediction about the next ball observed.

After seeing N balls go LEFT, the posterior distribution ends up proportional to a distribution where each hypothesis p has weight p^N.

But if you forget to normalize that prior-times-likelihood distribution, and ask what happens if every part of the distribution then contributes p likelihood to the next ball going LEFT, you'd first add one extra factor of p everywhere, and then calculate:

1/INF * [ (0/INF)^(N+1)  +  (1/INF)^(N+1)  +  ...  +  (INF/INF)^(N+1) ] = 1/(N+2)

After seeing 100 balls go LEFT, you shouldn't calculate a 1/102 chance of the next ball going LEFT - you should think it's more likely than not for the next ball to go LEFT - so something has gone wrong here; and what's wrong is that we didn't normalize the prior-times-likelihood distribution.

The pre-normalized distribution sums to 1/(N+1); the normalized distribution should sum to 1.  So we divide everything through by 1/(N+1), meaning we multiply by (N+1), to make it sum to 1.  For this example, that means multiplying everything uniformly by 101 to normalize it.  Which yields a 101/102 chance of the next ball going LEFT, which is what it should be.

Though, Keltham quickly remarks, the principle of making the posterior sum to 1, is something that only applies within that whole metahypothesis.  If you see balls going LEFT RIGHT LEFT RIGHT repeating, you might form, after a few steps, the hypothesis that the ball goes LEFT RIGHT repeating, which can't be expressed within that whole metahypothesis.  All of this reasoning is taking place inside the metahypothesis about the balls having some stable but otherwise not-further-scrutable fractional propensity to go LEFT or RIGHT, with all such propensities between 0 and 1 being equally likely on priors.  This whole metahypothesis then may do better or worse compared to some other hypotheses.

He's not saying that the metahypothesis absolutely has to be true, by saying that the sum of all the hypotheses inside it should be 1.  If you started believing that, you'd never in principle be able to notice any other pattern like LEFT RIGHT repeating.

Rather, this whole metahypothesis is itself being entertained as one of many in-principle possible metahypotheses about how things could be, and the probabilities of all the hypotheses inside it summing to 1, are to be understood as being what we get when we entertain the possibility, condition on, reason inside, the metahypothesis being true.

After seeing N balls go left, the whole metahypothesis has racked up a joint probability of 1/(N+1) and scored log(1/(N+1)) in terms of how well it's doing.  If you previously started out with any prior credence on an incompatible metahypothesis, like 'Actually maybe the ball just always goes LEFT or RIGHT, both equally credible on priors,' that other metahypothesis may be doing better.  In fact, the always-left-or-always-right metahypothesis has now racked up probability 1/2 and a score of log(1/2) = -1, which is probably quite a bit better than 1/(N+1) if you've seen a lot of balls go LEFT.  You'd probably be believing in that metahypothesis a bit more, now, even if it started out less meta-credible than the other.

Permalink Mark Unread

"Why do we need that - other hypothesis, metahypothesis - about there being hypotheses where the ball always goes LEFT or RIGHT?  Aren't those hypotheses already inside - the metahypothesis about all the fractions of times a ball could go LEFT or RIGHT?"

"Like, 0/INF is that it never goes left, and INF/INF is that it always goes left... am I missing something?"

Permalink Mark Unread

"Both those hypotheses have prior probabilities of 1/INF, remember?  That's equal to zero.  So the metahypothesis thinks it's infinitely unlikely that the fraction of times the ball goes left is exactly equal to 1 or 0.  It never notices that's true, no matter how many examples it sees."

"If it sees 100 balls go left, it'll get a joint prediction of 1/101.  If it sees 10,000 balls go left, it'll get a joint prediction of 1/10,001.  In the limit, the probability that it assigns to a sequence of left-going balls, continuing forever, is 0.  Before you've rolled a single ball, the metahypothesis is already absolutely certain that it won't just go left every time."

"The alternative metahypothesis that the ball has a 50% chance of always going left, and 50% chance of always going right, loses all of its probability and scores negative infinity as soon as it sees any sequence with a mix of left and right balls.  But if you present that metahypothesis with a sequence of balls going left forever, no matter how long that continues, it predicted that with 50% probability from the start."

Permalink Mark Unread

Keltham will now zoom in closer on what happens to the 'any fraction between 0 and 1 is equally likely' metahypothesis when it updates.

Let's say that we see 10 balls go LEFT.

After seeing 10 balls go LEFT, the metahypothesis as a whole has a cumulative score of 1/(N+1) = 1/11, or equivalently, if you asked this metahypothesis at the dawn of time what probability it assigned to the first 10 balls going left, it would answer "1/11".  Equivalently, isomorphically, these are stating the same thing, if you multiply out the prior distribution (1/INF everywhere) and the likelihood distribution (p^10 everywhere) you'd get a new distribution p^10/INF which, if you summed over the infinitely many terms inside it, would sum to 1/11.

Continuing to reason within this metahypothesis, one of the internal subhypotheses must by assumption be true, and so within the metahypothesis the posterior probabilities of the (infinite collection of) subhypotheses must sum to 1.

Which you'd obviously do by dividing all the terms through by 1/11, that is, multiplying them by 11.

The distribution...

(0/INF)^10     (1/INF)^10             (INF/INF)^10
---------- + ---------- + ... + ------------
INF * 1/11 INF * 1/11 INF * 1/11

...will, indeed, sum to 1.

To be clear, by having the distribution sum to 1, we are assuming-for-the-sake-of-extrapolation that some hypothesis in the collection must be true.  But no matter how far the posterior updates on finite data, every individual hypothesis in the collection will still have posterior 0.

After rolling 1000 balls and seeing 985 go left, there's finite probability between 0.98 and 0.99, but 0 probability on the fraction being exactly 0.985.

That's just how this metahypothesis rolls.  If you wanted a metahypothesis that could put little finite dots of probability mass on all the big round numbers like 0.985, 1/2pi, and so on, that would be a much more complicated ask.

Permalink Mark Unread

What we ask about, having updated, is not so much the probability of individual hypotheses - which is always 0 - but the new shape of the curve.

If you see 10 balls go LEFT, the individual hypothesis 0.9 has now scored around 1/e, 37% or so, compared to 1 to the individual hypothesis 1.0.  If you see 100 balls go LEFT, raise all that to a power of 10; now 0.9 will have less than 0.01% the cumulative likelihood predicted as 1.0.  On the scale of the curve, this means that over 99.99% of the probability is now concentrated into the interval between 90% and 100%.  Though, of course, still 0% is concentrated at exactly 100%.

Keltham will Prestidigitate-sketch a few rough images of what curves like these should maybe probably look like, after updating on various numbers of LEFT and RIGHT balls observed.  Keeping in mind that Keltham's not a 'computer', and can't draw the curves perfectly.

Permalink Mark Unread

One of the newcomers who does know what calculus is, and has indeed known it for years, asks a pending question he's had for a while.

Are some of Keltham's earlier statements up until this point equivalent to - in Golarion notation - saying...

Integral of p from 0 to 1:  [  p^N dp  ]   =  1/(N+1)

...?

Permalink Mark Unread

Yes.

Permalink Mark Unread

Lady Avaricia has gotten away with a lot, around Keltham, and so has Ione.

The newcomer will venture, very politely, that the calculus-based way of expressing this and calculating this, seems to him simpler?

Permalink Mark Unread

If you're willing to assume your audience can already freely wield the machinery of derivatives and antiderivatives and understand on an intuitive level everything that they mean, then sure, you can solve the problem more quickly using overpowered tools like that.

Integrating x^N gives you x^(N+1)/(N+1), which evaluated from 0 to 1 will give you 1/(N+1).

Given a mix of N LEFTs and M RIGHTs, however, and the need to renormalize the posterior over that, they might find themselves in a bit more trouble when it comes to deriving the integral...

Integral of p from 0 to 1:  [  (1 - p)^M p^N dp  ]   =   M! * N! / (M + N + 1)!
Permalink Mark Unread

 

...okay, he doesn't see any obvious way he'd get that answer.

Is it going to be easier to derive using Keltham's approach?

Permalink Mark Unread

Keltham doesn't actually remember the exact proof, it's been a while, but he's pretty sure the introduction he used was the one his teachers used on him.  So it's possibly the right introduction for getting the more complicated proof in the least painful way, once he starts trying to remember that?

And in any case, Keltham hasn't yet been told he's allowed to assume that everyone in this class has a full intuition for derivatives and integrals yet.  Keltham hasn't taught them a calculus class himself, and who knows what weird malfunctions exist in the Golarion way of teaching it.

Why, it wouldn't shock him, at this point, if people are just being told to memorize formulas and not really taught to intuit how anything works!

So he's just talking directly about the infinity of possible hypotheses between 0 and 1, and how to sum up priors-times-likelihoods and posterior-predictions from those, rather than using calculus to abstract over that.  Abstracting over that only works to teach mathematical intuition if people already know what's being abstracted away.

Permalink Mark Unread

Keltham is in favor of people understanding things using calculus and continuous distributions, to be clear.  So long as they can also understand the ideas in terms of a countably infinite collection of individual hypotheses each with probability 1/INF.  You don't want to start identifying either representation or methodology of analysis with the underlying idea!

Which idea is just: a metahypothesis where any fraction from 0 to 1 seems equally plausible on priors.  Those get updated on observations of LEFT and RIGHT, that have different likelihoods for different propensity-fractions; the allocation of posterior probability over fractions changes and gets normalized back to summing to 1; that changes the new predictions for LEFT and RIGHT on successive rounds, having updated on all of the previous rounds.

Keltham will now go up to the wall and spend a bit of time figuring out how to derive the rest of the Rule of Succession, which there's no simple or obvious way to prove using calculus known to anyone in this room anyways.

Thankfully, they know what all the correct answers have to be!  Using much simpler combinatoric arguments, about new balls ending up randomly ordered anywhere between the left boundary, all the LEFT balls, the 0 ball, all the RIGHT balls, and the right boundary.

Permalink Mark Unread

Eventually Keltham does succeed in deriving (again) (but this time proving it using dubious infinitary arguments, instead of clear and simple combinatorics, so they can see what's happening with priors and posteriors and likelihoods behind the scene) that indeed:

If you start out thinking any fraction of LEFT and RIGHT between 0 and 1 is equally plausible on priors, and you see experimental results going LEFT on N occasions and going RIGHT on M occasions, the prediction for the next round is (N+1)/(N+M+2) for LEFT and (M+1)/(N+M+2) for RIGHT.

Permalink Mark Unread

(Continued in "my fun research project".)

Permalink Mark Unread