Student-b
Dear Sir/Madam,
In this letter, I derive and tabulate the maximum entropy values for the probabilities of each side of biased -sided dice, for . These probabilities for each of the n options (sides), are those which have the least input information beyond what we know, which is nothing more than the bias or average score on the -sided die. I generalise the “Brandeis dice” problem from E T Jaynes’ 1963 lectures, to an -sided die, from the 6-sided case. To calculate these probabilities, I obtain the solution of an -order polynomial equation, derived using a power series identity, for the value of the Lagrange multiplier, . The resulting maximally-equivocated prior probabilities at the 5th, 17th, 34th, 50th (fair), 66th, 83rd, and 95th percentiles of the range from 1 up to will aid in decision-making, where the options are the conditions we cannot influence, but across which we may have a non-linear payoff.
We use the standard variational principle in order to maximise the entropy in the system.
where the index is not summed over in the first equation, and where the are the probabilities of the options, e.g. sides of an n-dice. are the numbers given in the problem statement (constraints or biases), and are functions of the Lagrange multipliers . The second equation is just the probability axiom requiring the probabilities to sum to one. This set of constraints is solved by using Lagrange multipliers. The formal solution is
where is the partition function and are the set of multipliers, of which for a solution to the problem there need to be fewer than n, though in our current problem as we shall see, there is only one. The constraints are satisfied if:
for k ranging from 1 to m.
Our measure of entropy is given by and in terms of our constraints, i.e. the data, this function is:
The solution for the maximum of S is:
For k in same range up to m. For our set of n-sided dice, and so I can simplify to F. The are simply the set of the values on the n sides of our die.
For the problem at hand of the biased die, I introduce the quantity q which I define as the tested, trusted average score on the given n-sided die in hand. That is, I set here, our bias constraint number, which can range from the lowest die value 1 through to the highest value, n.
i.e. 3.5 on a 6-sided die, then the die is fair, otherwise, it has a bias and therefore an additional constraint. I assume this is all I know and believe about the die, other than the number of sides, n.
We see that becomes just and the equation for reduces to
and the equation for reduces to
and its solution is
After a little algebra, I found that the partition function Z is given by
and after some further algebra, I found that in order to determine the value of , where , corresponding to the maximum entropy (least input information) set of probabilities, we must find the positive, real root of the following equation, which is not unity:
By inspection, this equation is always satisfied by the real solution x = 1, which corresponds to the fair or unbiased die, with all probabilities equal to 1/n for n sides. We need the other real root, and we obtain this by simple numerical calculation. From the solution for the given value of bias q, the set of probabilities corresponding to maximum entropy for each side of the relevant n-sided die are easily generated.
The following tables may be of use in decisionmaking in business and other contexts, especially where the agent (the organisation or individual making a decision) has a non-linear desirability or utility function over the outcomes (i.e. the values of the discrete set of possible options), does not have perfect intuition and does not wish to put any more information into the decision that is not within the agent’s state of knowledge.
I present tables for n = 3, 4, 6, 8, 10, 12, 15, and 20 here, each at 7 bias values of q for each n, corresponding to percentages of the range from 1 to n of 5%, 17, 34, 50, 66, 83 and 95%. There is transformation group symmetry in this problem. If represents the side with spots up, then when we reflect from and transform we obtain the same probability, e.g. the probability of a 1 on a six sided die at 5th percentile bias is the same as a 6 at 95th percentile bias. This is why in our tables we can observe the corresponding symmetry in the values of the probabilities and in the entropy, which is maximal of all biases when there is no bias and thus no constraint. Readers may wish arbitrarily to adjust any of the probabilities in the tables in the appendix and recalculate the entropy , which will be lower than the maximum entropy value in the table.
APPENDIX Student-b Maximum Entropy Probability Tables
n=3 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
q05 | q17 | q34 | q0 | q66 | q83 | q95 | |||||
q-vals | 1.1 | 1.34 | 1.68 | 2 | 2.32 | 2.66 | 2.9 | ||||
Score | Probabilities | ||||||||||
1 | 0.9078 | 0.7232 | 0.5064 | 0.3333 | 0.1864 | 0.0632 | 0.0078 | ||||
2 | 0.0843 | 0.2137 | 0.3072 | 0.3333 | 0.3072 | 0.2137 | 0.0843 | ||||
3 | 0.0078 | 0.0632 | 0.1864 | 0.3333 | 0.5064 | 0.7232 | 0.9078 | ||||
entropy | 0.3343 | 0.7386 | 1.0203 | 1.0986 | 1.0203 | 0.7386 | 0.3343 |
n=4 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
q05 | q17 | q34 | q0 | q66 | q83 | q95 | |||||
q-vals | 1.15 | 1.51 | 2.02 | 2.5 | 2.98 | 3.49 | 3.85 | ||||
Score | Probabilities | ||||||||||
1 | 0.8689 | 0.6425 | 0.4136 | 0.25 | 0.1241 | 0.0324 | 0.002 | ||||
2 | 0.1141 | 0.2374 | 0.2769 | 0.25 | 0.1854 | 0.0877 | 0.015 | ||||
3 | 0.015 | 0.0877 | 0.1854 | 0.25 | 0.2769 | 0.2374 | 0.1141 | ||||
4 | 0.002 | 0.0324 | 0.1241 | 0.25 | 0.4136 | 0.6425 | 0.8689 | ||||
Entropy | 0.445 | 0.9502 | 1.2921 | 1.3863 | 1.2921 | 0.9502 | 0.445 |
n=6 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
q05 | q17 | q34 | q0 | q66 | q83 | q95 | |||||
q-vals | 1.250 | 1.85 | 2.70 | 3.5 | 4.3 | 5.15 | 5.75 | ||||
Score | Probabilities | ||||||||||
1 | 0.7998 | 0.5260 | 0.3043 | 0.1666 | 0.072 | 0.0134 | 0.0003 | ||||
2 | 0.1602 | 0.2527 | 0.2282 | 0.1667 | 0.0961 | 0.028 | 0.0013 | ||||
3 | 0.0321 | 0.1214 | 0.1711 | 0.1667 | 0.1282 | 0.0583 | 0.0064 | ||||
4 | 0.0064 | 0.0583 | 0.1282 | 0.1667 | 0.1711 | 0.1214 | 0.0321 | ||||
5 | 0.0013 | 0.028 | 0.0961 | 0.1667 | 0.2282 | 0.2527 | 0.1602 | ||||
6 | 0.0003 | 0.0135 | 0.0721 | 0.1667 | 0.3043 | 0.526 | 0.7998 | ||||
Entropy | 0.6254 | 1.2655 | 1.6794 | 1.7918 | 1.6794 | 1.2655 | 0.6254 |
APPENDIX Student-b Maximum Entropy Probability Tables (ctd)
n=8 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
q05 | q17 | q34 | q0 | q66 | q83 | q95 | |||||
q-vals | 1.35 | 2.19 | 3.38 | 4.5 | 5.62 | 6.81 | 7.65 | ||||
Score | Probabilities | ||||||||||
1 | 0.7407 | 0.4454 | 0.2412 | 0.125 | 0.05 | 0.0076 | 0.0001 | ||||
2 | 0.1921 | 0.2489 | 0.1927 | 0.125 | 0.0626 | 0.0136 | 0.0002 | ||||
3 | 0.0498 | 0.1391 | 0.1539 | 0.125 | 0.0784 | 0.0243 | 0.0009 | ||||
4 | 0.0129 | 0.0777 | 0.1229 | 0.125 | 0.0982 | 0.0434 | 0.0034 | ||||
5 | 0.0034 | 0.0434 | 0.0982 | 0.125 | 0.1229 | 0.0777 | 0.0129 | ||||
6 | 0.0009 | 0.0243 | 0.0784 | 0.125 | 0.1539 | 0.1391 | 0.0498 | ||||
7 | 0.0002 | 0.0136 | 0.0626 | 0.125 | 0.1927 | 0.2489 | 0.1921 | ||||
8 | 0.0001 | 0.0076 | 0.05 | 0.125 | 0.2412 | 0.4454 | 0.7407 | ||||
Entropy | 0.7726 | 1.5012 | 1.9569 | 2.0794 | 1.9569 | 1.5012 | 0.7726 |
n=10 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
q05 | q17 | q34 | q0 | q66 | q83 | q95 | |||||
q-vals | 1.45 | 2.53 | 4.06 | 5.5 | 6.94 | 8.47 | 9.55 | ||||
Score | Probabilities | ||||||||||
1 | 0.6896 | 0.3862 | 0.2 | 0.1 | 0.0381 | 0.005 | 0.0000 | ||||
2 | 0.214 | 0.2382 | 0.1663 | 0.1 | 0.0458 | 0.0081 | 0.0001 | ||||
3 | 0.0664 | 0.147 | 0.1383 | 0.1 | 0.055 | 0.0131 | 0.0002 | ||||
4 | 0.0206 | 0.0907 | 0.115 | 0.1 | 0.0662 | 0.0213 | 0.0006 | ||||
5 | 0.0064 | 0.0559 | 0.0957 | 0.1 | 0.0796 | 0.0345 | 0.002 | ||||
6 | 0.002 | 0.0345 | 0.0796 | 0.1 | 0.0957 | 0.0559 | 0.0064 | ||||
7 | 0.0006 | 0.0213 | 0.0662 | 0.1 | 0.115 | 0.0907 | 0.0206 | ||||
8 | 0.0002 | 0.0131 | 0.055 | 0.1 | 0.1383 | 0.147 | 0.0664 | ||||
9 | 0.0001 | 0.0081 | 0.0458 | 0.1 | 0.1663 | 0.2382 | 0.214 | ||||
10 | 0.0000 | 0.005 | 0.0381 | 0.1 | 0.2 | 0.3862 | 0.6896 | ||||
Entropy | 0.8981 | 1.6905 | 2.1735 | 2.3026 | 2.1735 | 1.6905 | 0.8981 |
APPENDIX Student-b Maximum Entropy Probability Tables (ctd)
n=12 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
q05 | q17 | q34 | q0 | q66 | q83 | q95 | |||||
q-vals | 1.55 | 2.87 | 4.74 | 6.5 | 8.26 | 10.13 | 11.45 | ||||
Score | Probabilities | ||||||||||
1 | 0.6451 | 0.3408 | 0.1708 | 0.0833 | 0.0306 | 0.0036 | 0.0000 | ||||
2 | 0.2289 | 0.2255 | 0.1461 | 0.0833 | 0.0358 | 0.0055 | 0.0000 | ||||
3 | 0.0812 | 0.1492 | 0.125 | 0.0833 | 0.0419 | 0.0083 | 0.0001 | ||||
4 | 0.0288 | 0.0987 | 0.1069 | 0.0833 | 0.049 | 0.0125 | 0.0002 | ||||
5 | 0.0102 | 0.0653 | 0.0915 | 0.0833 | 0.0572 | 0.0189 | 0.0005 | ||||
6 | 0.0036 | 0.0432 | 0.0782 | 0.0833 | 0.0669 | 0.0286 | 0.0013 | ||||
7 | 0.0013 | 0.0286 | 0.0669 | 0.0833 | 0.0782 | 0.0432 | 0.0036 | ||||
8 | 0.0005 | 0.0189 | 0.0572 | 0.0833 | 0.0915 | 0.0653 | 0.0102 | ||||
9 | 0.0002 | 0.0125 | 0.049 | 0.0833 | 0.1069 | 0.0987 | 0.0288 | ||||
10 | 0.0001 | 0.0083 | 0.0419 | 0.0833 | 0.125 | 0.1492 | 0.0812 | ||||
11 | 0.0000 | 0.0055 | 0.0358 | 0.0833 | 0.1461 | 0.2255 | 0.2289 | ||||
12 | 0.0000 | 0.0036 | 0.0306 | 0.0833 | 0.1708 | 0.3408 | 0.6451 | ||||
Entropy | 1.0078 | 1.8001 | 2.1252 | 2.4849 | 1.7684 | 1.1462 | 1.0081 |
n=15 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
q05 | q17 | q34 | q0 | q66 | q83 | q95 | |||||
q-vals | 1.7 | 3.38 | 5.76 | 8 | 10.24 | 12.62 | 14.3 | ||||
Score | Probabilities | ||||||||||
1 | 0.5882 | 0.2898 | 0.1402 | 0.0667 | 0.0236 | 0.0025 | 0.0000 | ||||
2 | 0.2422 | 0.2063 | 0.1235 | 0.0667 | 0.0269 | 0.0035 | 0.0000 | ||||
3 | 0.0997 | 0.1469 | 0.1087 | 0.0667 | 0.0305 | 0.0049 | 0.0000 | ||||
4 | 0.0411 | 0.1046 | 0.0957 | 0.0667 | 0.0346 | 0.0069 | 0.0000 | ||||
5 | 0.0169 | 0.0745 | 0.0843 | 0.0667 | 0.0393 | 0.0097 | 0.0001 | ||||
6 | 0.007 | 0.053 | 0.0743 | 0.0667 | 0.0447 | 0.0136 | 0.0002 | ||||
7 | 0.0029 | 0.0378 | 0.0654 | 0.0667 | 0.0507 | 0.0191 | 0.0005 | ||||
8 | 0.0012 | 0.0269 | 0.0576 | 0.0667 | 0.0576 | 0.0269 | 0.0012 | ||||
9 | 0.0005 | 0.0191 | 0.0507 | 0.0667 | 0.0654 | 0.0378 | 0.0029 | ||||
10 | 0.0002 | 0.0136 | 0.0447 | 0.0667 | 0.0743 | 0.053 | 0.007 | ||||
11 | 0.0001 | 0.0097 | 0.0393 | 0.0667 | 0.0843 | 0.0745 | 0.0169 | ||||
12 | 0.0000 | 0.0069 | 0.0346 | 0.0667 | 0.0957 | 0.1046 | 0.0411 | ||||
13 | 0.0000 | 0.0049 | 0.0305 | 0.0667 | 0.1087 | 0.1469 | 0.0997 | ||||
14 | 0.0000 | 0.0035 | 0.0269 | 0.0667 | 0.1235 | 0.2063 | 0.2422 | ||||
15 | 0.0000 | 0.0025 | 0.0236 | 0.0667 | 0.1402 | 0.2898 | 0.5882 | ||||
Entropy | 1.1517 | 2.0471 | 2.5698 | 2.7081 | 2.5698 | 2.0471 | 1.1517 |
APPENDIX Student-b Maximum Entropy Probability Tables (ctd)
n=20 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
q05 | q17 | q34 | q0 | q66 | q83 | q95 | |||||
q-vals | 1.95 | 4.23 | 7.46 | 10.5 | 13.54 | 16.77 | 19.05 | ||||
Score | Probabilities | ||||||||||
1 | 0.5128 | 0.2318 | 0.108 | 0.05 | 0.0171 | 0.0016 | 0.0000 | ||||
2 | 0.2498 | 0.1784 | 0.098 | 0.05 | 0.0188 | 0.0021 | 0.0000 | ||||
3 | 0.1217 | 0.1372 | 0.0889 | 0.05 | 0.0207 | 0.0027 | 0.0000 | ||||
4 | 0.0593 | 0.1056 | 0.0807 | 0.05 | 0.0229 | 0.0035 | 0.0000 | ||||
5 | 0.0289 | 0.0812 | 0.0732 | 0.05 | 0.0252 | 0.0045 | 0.0000 | ||||
6 | 0.0141 | 0.0625 | 0.0665 | 0.05 | 0.0278 | 0.0059 | 0.0000 | ||||
7 | 0.0069 | 0.0481 | 0.0603 | 0.05 | 0.0306 | 0.0077 | 0.0000 | ||||
8 | 0.0033 | 0.037 | 0.0547 | 0.05 | 0.0337 | 0.01 | 0.0001 | ||||
9 | 0.0016 | 0.0285 | 0.0497 | 0.05 | 0.0371 | 0.013 | 0.0002 | ||||
10 | 0.0008 | 0.0219 | 0.0451 | 0.05 | 0.0409 | 0.0169 | 0.0004 | ||||
11 | 0.0004 | 0.0169 | 0.0409 | 0.05 | 0.0451 | 0.0219 | 0.0008 | ||||
12 | 0.0002 | 0.013 | 0.0371 | 0.05 | 0.0497 | 0.0285 | 0.0016 | ||||
13 | 0.0001 | 0.01 | 0.0337 | 0.05 | 0.0547 | 0.037 | 0.0033 | ||||
14 | 0.0000 | 0.0077 | 0.0306 | 0.05 | 0.0603 | 0.0481 | 0.0069 | ||||
15 | 0.0000 | 0.0059 | 0.0278 | 0.05 | 0.0665 | 0.0625 | 0.0141 | ||||
16 | 0.0000 | 0.0045 | 0.0252 | 0.05 | 0.0732 | 0.0812 | 0.0289 | ||||
17 | 0.0000 | 0.0035 | 0.0229 | 0.05 | 0.0807 | 0.1056 | 0.0593 | ||||
18 | 0.0000 | 0.0027 | 0.0207 | 0.05 | 0.0889 | 0.1372 | 0.1217 | ||||
19 | 0.0000 | 0.0021 | 0.0188 | 0.05 | 0.098 | 0.1784 | 0.2498 | ||||
20 | 0.0000 | 0.0016 | 0.0171 | 0.05 | 0.108 | 0.2318 | 0.5128 | ||||
Entropy | 1.351 | 2.3085 | 2.8526 | 2.9957 | 2.8526 | 2.3085 | 1.351 |