Hide table of contents

Quick Sidenote

The audio recording isn't great with functions yet, and there will be a lot of those.  listener's discretion is advised.

Also, the time estimate is a little off, I'd estimate it takes around 17 minutes to read, including breaks & pauses. Not including breaks & pauses, I'd say about 9 minutes.

Also, this makes little to no sense on its own, so read This one first, if you haven't already. It'll take around 8 minutes.

Also, it has come to my attention that this does not include the discrete case, and the probability distribution functions are a bit wonky, as the method currently provided is optimized purely for comparing abstract functions. At the moment, the function treats probability distributions* as though there is a uniformly randomly distributed input (x), and a non-uniform output (), and the probability distribution* of  is treated as * for the continuous function , ( being the inverse function of .). I am working to fix this. (I am currently undergoing more ambitious and time-effective projects, and this article will likely not be outdated until many months from now, if ever.)

WARNING: Only read this if you want to, or are skeptical about whether it works, and need to be sure. It doesn't really work.

 

Conceptualizing the Problem:

The problem at hand is to understand the expected value when we "roll" a number of functions a certain number of times. A "roll" in this context refers to generating an input in a function's domain. For each function, [1]

 , confined to specific input ranges, denoted by

Minimum inputMaximum input#times we "roll" the inputFunction


​. The coefficients  represent the number of times we evaluate each respective function.

The sgn Function:

Before diving deep, let's clarify the sgn⁡(x) function. The sgn(x) or "signum" function is simple but powerful. It returns:

  • −1 if x<0
  • 0 if x=0
  • 1 if x>0.

So,  checks if x is larger than our function's output. By adding 1 and dividing by 2, the original function's outcomes of -1, 0, and 1 transform to 0, 0.5, and 1 respectively. Essentially, this is turning the signum function into a binary (0 or 1) indicator for whether 

Functions 

These functions integrate the modified ⁡sgn function over the domain of their respective functions . Essentially, they measure the proportion (and, by extension, probability) of the function()'s domain where its value is less than x.

Small break

Congratulations if you've made it this far! As a reward, here's a Monty Python video that  I found funny. 

This is optional and works best if you try not to worry about what you've read so far. If you think you would, I suggest you re-read it, or read the recap.

Quick recap

we randomly generate inputs for functions and use the largest value.

To get the expected value of this, we multiply each output by the probability that it's the highest using the function above, denoted as , and then "add" all of those outputs to get the expected value.

Bringing it Together in 

Understanding the Mechanism of :

 calculates the expected value for each function  ​ individually and then multiplies them by how many times we "roll" them. ( )

Component Analysis:

The full function is 

Let's focus on the first term:

This integral gives the expected value of  weighted by the probability that  has the highest output among the three functions for all valid inputs x.

Role of the Exponents: 

The exponents in the functions  serve a crucial role. Given that each of these functions essentially returns a probability, raising them to their respective exponents is the equivalent of multiplying the probabilities.

  • : Represents the probability that  is more than the value of  times. ( is the number of times we "roll" an input for the function ). The reason for the -1 is because we don't need to calculate the chances of  being higher than  The reason for the is because it outputs  for all values where  is more than or equal to , and  for all values less than 1. We do this because we don't want to get a negative exponent, as this may result in dividing by , which is undefined.
  •  and ​ Represent the probabilities that  surpasses  and  respectively, the necessary number of times.

Enhancing your understanding

Probability of Being the Highest: 

By multiplying these terms, we calculate the joint probability that  is the highest output among the three functions a specific number of times as denoted by the coefficients, the same as you would if were the same as a hypothetical , ...[1]

Weighting by the Function Value: 

Finally, we multiply this probability by the function's value  to weigh it by the actual outcome of the function, the same way you would always multiply probability by value.

Summing Over the Domain: 

By integrating over the domain of , we capture the aggregated effect across all possible inputs.

Applying the Coefficients :

After obtaining the expected contribution of each function, we multiply each result by their respective coefficient ​, or . These coefficients account for the number of times we evaluate each respective function in our simulation. 

We can then use the same logic for  and .

If you have any questions, I suggest you ask Chat-GPT or me. If you have any suggestions as to how I can make this clearer, or a better way of finding the expected value of the best option, or any wording that could be done differently, tell me. (ONLY if you want). No pressure.

If there's anything incorrect, please tell me.

Congratulations! You made it to the end!

as a special reward, have another funny video. You deserve it.

  1. ^

    3 Functions is an arbitrary quantity. you can use as many as you want. (Here's how) 

2

0
0

Reactions

0
0

More posts like this

Comments
No comments on this post yet.
Be the first to respond.
More from wes R
Curated and popular this week
Relevant opportunities