Why It Works
The audio recording isn’t great with functions yet, and there will be a lot of those. listener’s discretion is advised.
Also, the time estimate is a little off, I’d estimate it takes around 17 minutes to read, including breaks & pauses. Not including breaks & pauses, I’d say about 9 minutes.
Also, this makes little to no sense on its own, so read This one first, if you haven’t already. It’ll take around 8 minutes.
Also, it has come to my attention that this does not include the discrete case, and the probability density functions are a bit wonky, as the method currently provided is optimized purely for comparing abstract functions. At the moment, the function treats “PDF”*s as though there is a uniformly randomly distributed input (x), and a non-uniform output (). The “PDF”* of is equal to for the continuous function , ( being the inverse function of .). I am working to fix this.
WARNING: Only read this if you want to, or are skeptical about whether it works, and need to be sure.
Conceptualizing the Problem:
The problem at hand is to understand the expected value when we “roll” a number of functions a certain number of times. A “roll” in this context refers to generating an input in a function’s domain. For each function, 
, confined to specific input ranges, denoted by
|Minimum input||Maximum input||#times we “roll” the input||Function|
. The coefficients represent the number of times we evaluate each respective function.
The sgn Function:
Before diving deep, let’s clarify the sgn(x) function. The sgn(x) or “signum” function is simple but powerful. It returns:
−1 if x<0
0 if x=0
1 if x>0.
So, checks if x is larger than our function’s output. By adding 1 and dividing by 2, the original function’s outcomes of −1, 0, and 1 transform to 0, 0.5, and 1 respectively. Essentially, this is turning the signum function into a binary (0 or 1) indicator for whether
These functions integrate the modified sgn function over the domain of their respective functions . Essentially, they measure the proportion (and, by extension, probability) of the function()’s domain where its value is less than x.
Congratulations if you’ve made it this far! As a reward, here’s a Monty Python video that I found funny.
This is optional and works best if you try not to worry about what you’ve read so far. If you think you would, I suggest you re-read it, or read the recap.
we randomly generate inputs for functions and use the largest value.
To get the expected value of this, we multiply each output by the probability that it’s the highest using the function above, denoted as , and then “add” all of those outputs to get the expected value.
Bringing it Together in
Understanding the Mechanism of :
calculates the expected value for each function individually and then multiplies them by how many times we “roll” them. ( )
The full function is
Let’s focus on the first term:
This integral gives the expected value of weighted by the probability that has the highest output among the three functions for all valid inputs x.
Role of the Exponents:
The exponents in the functions serve a crucial role. Given that each of these functions essentially returns a probability, raising them to their respective exponents is the equivalent of multiplying the probabilities.
: Represents the probability that is more than the value of , times. ( is the number of times we “roll” an input for the function ). The reason for the −1 is because we don’t need to calculate the chances of being higher than The reason for the is because it outputs for all values where is more than or equal to , and for all values less than 1. We do this because we don’t want to get a negative exponent, as this may result in dividing by , which is undefined.
and Represent the probabilities that surpasses and respectively, the necessary number of times.
Enhancing your understanding
Probability of Being the Highest:
By multiplying these terms, we calculate the joint probability that is the highest output among the three functions a specific number of times as denoted by the coefficients, the same as you would if were the same as a hypothetical , , …
Weighting by the Function Value:
Finally, we multiply this probability by the function’s value to weigh it by the actual outcome of the function, the same way you would always multiply probability by value.
Summing Over the Domain:
By integrating over the domain of , we capture the aggregated effect across all possible inputs.
Applying the Coefficients :
After obtaining the expected contribution of each function, we multiply each result by their respective coefficient , , or . These coefficients account for the number of times we evaluate each respective function in our simulation.
We can then use the same logic for and .
If you have any questions, I suggest you ask Chat-GPT or me. If you have any suggestions as to how I can make this clearer, or a better way of finding the expected value of the best option, or any wording that could be done differently, tell me. (ONLY if you want). No pressure.
If there’s anything incorrect, please tell me.
Congratulations! You made it to the end!
as a special reward, have another funny video. You deserve it.
3 Functions is an arbitrary quantity. you can use as many as you want. (Here’s how)