Determining the Final Ranking
For calculating the distribution of a voters’ voting power we originally used the formula in the Budget Box Whitepaper. However, for this application, to improve UX we want to avoid requiring the extra pairwise comparisons that would be required for this method and instead create a simple scoring algorithm so that each rank within each category gets a specific amount of points and the categories’ rank also gets a specific amount of points. This will speed up the voting experience and remove cognitive overload.
- Each category will have 15-30 projects.
- Voters will filter projects before ranking.
- Projects that do not pass the filter stage will receive 0 points from that voter.
- At least 21% of the projects in a category must pass the filter and be ranked.
- Delegates and OP Token Holders have different voting power weights whereas Recipients and Badgeholders are 1 address - 1 vote.
Pairwise Voting Overview
- The voter will be ranking the projects within each category, and after they have ranked two categories, they will also create a category ranking for the categories they have ranked.
- There are four stakeholder groups and each will have its own completely independent voting result.
- A voter that is part of multiple stakeholder groups only has to vote once, and their vote will be used in each group that they qualify for.
- Each voter has a voting weight in their stakeholder group. Badgeholders and Recipients all have a weight of 1 vote, OP Holders and Delegates have different voting weights depending on OP amounts held/delegated (see post above).
- If someone has a voting weight of 10 as a delegate, their ballot will be counted 10 times.
- Every voter is submitting a ballot for each category that they rank. The projects they rank within a category each receive a score and those scores are combined with the other voters’ ballots for that category to give a weighted average score for each project within each category.
- Every voter is also submitting a ballot for ranking the categories that they reviewed. Their overall category rank ballots are also calculated by a weighted average score just like projects.
- The weighted average category score for each category is then multiplied by the weighted average score for each project within each category to get a final score for each project.
- The final outcome will greatly depend on the voting algorithm that is used to give scores for each rank within each category.
- Each category will have a set score for each rank, this scoring algorithm will depend on how many projects are in the category. The scores for the ranking of categories will follow the same pattern as the scores for the ranking of projects within the category (with the exception of the handicap modification). This scoring algorithm is discussed in detail in the next sections.
Scoring formulas we are considering
We reviewed several different scoring formulas and have narrowed it down to 3. Gaussian, Poisson our own custom algorithm. We have not chosen the final scoring formula yet and would love your feedback! The first decision to make is which baseline distribution formula we should use.
1. Normal (Gaussian) Distribution
- f(x): The score for a project (or category) with rank 𝑥 (first, sixth, eighteenth, etc)
- 𝑥: The rank number of a project inside a category (for scoring projects) OR the rank number of a category among all categories (for scoring categories).
- 𝜎: The desired standard deviation of the distribution. The smaller this value is set, the more weight goes to top ranked projects. We expect most configurations will use a value between 3 & 7 however any positive number can be used.
The Gaussian distribution is the middle way of the distributions, it gives the top 7 to 14 projects a score above 1% when comparing the extremes of the expected configuration space (ignoring the handicap).
2. Poisson Distribution
- λ: The average number of events. The smaller this value is set, the more weight goes to top ranked projects. We expect most configurations will use a value between 1.7 & 1.98 however any number between 1 & 2 can be used.
The Poisson distribution is very top heavy, the winners of each category get much larger scores, only the top 5 to 6 projects get a score above 1% of the points for the entire configuration space (ignoring the handicap). This means that the projects not in the top 5 of someone’s ballot for each category get almost none of that voter’s voting power.
3. Our Custom Distribution
Our custom distribution is the most egalitarian, it gives the top 8 to 23 projects a score above 1% when comparing the extremes of the expected configuration space (ignoring the handicap). In this distribution the preferences expressed lower in the rankings have more of an impact, but being the winner of a category of someone’s ballot has the least impact of the 3 distributions.
- r: The total number of projects in the category or when ranking categories, the total number of categories. Determined by round data.
- δ: The angle of the slope. The larger this value is set, the more weight goes to top ranked projects. We expect most configurations will use a value between 0.5 & 10 however any number between 0 & 180 can be used.
To better adjust these scoring formulas to the task at hand we made several modifications, although we are also considering using the normal distribution without modifications.
Modifications on top of the Algos
While the normal distribution could be used without modifications, we wanted to further customize the scoring algorithms to our use case using the following modifications.
Modification #1: Add a minimum to bump up the long tail of projects.
In these distributions, for a high r, the long tail of projects were given a very small amount. Because these projects being ranked passed the first filter, we assume the user wants to give the bottom projects more votes than the distributions were providing. This simple modification allows us to do that.
- f(x): The score for rank x given the chosen distribution method
- m: The minimum percentage every project should get from the voting algorithm. The higher this number the more votes are allocated to the long tail of lower ranked projects. We expect most configurations will use a value between 0.5% & 1.5% however any number between 0 & 3% can be used.
Modification #2: Add a variance to give rankings more importance
In these distributions, for a low r the difference between the ranks was smaller than expected (e.g. if r = 2 ,the top rank gets 52% of the votes and the bottom rank gets 48% votes, this is not different enough) and with modification #1, the long tail was also getting almost the same amount for the last several ranks. To create a larger difference between the ranks, we created this simple modification that brings more of a power law shape to the distribution.
- f’(x): The score for rank x given the chosen distribution method with Mod1
- v: The variance. The higher this number the more votes are allocated to the top ranked projects and the bigger the difference is between top ranked projects. We expect most proposed configurations will use a value between 0.5 & 8 however any number between 0 & 1000 can be used.
Modification #3: Normalize to a percentage & add a handicap for projects in larger categories
Without this modification, a project that gets first place out of 15 projects actually scores the same or more points than a project that gets first place out of 30 projects! When the actual result should be the exact opposite. This is a simple modification that gives a linear handicap to categories with more projects. If a category has 30 projects each project will receive a boost of h percent multiplied by what their final score would be without the handicap, if it has 25 projects it will get a boost of 2h/3, if it has 20 projects it will get a boost of h/3, and if it has 15 projects, it will not receive a handicap at all. Before the handicap is applied, the scores after mod 1 and 2 are applied are normalized to a percentage. The ranking of categories does not need to apply this mod.
- f’’(x): The score for rank x given the chosen distribution method with Mod 1 & Mod 2
- ∑ f’'(x_i): Sum of scores: All the scores for each project from the previous step added together so that we can make the project scores normalized to a percentage.
- h: The handicap. The higher this number, the larger the bonus that goes to projects from large categories. We expect most proposed configurations will use a value between 15% & 175% however any positive number can be used.
Play with our Dashboard!
I know this sounds crazy, but we want to open it up to the Optimism community to play with, and even propose, your own Pairwise voting algorithm! Our team will pick the best distribution on the 22nd of June.
We have created this dashboard for everyone to copy and play with to propose a configuration.
When playing with the dashboard, the best strategy is to focus on one distribution at a time and tune the parameters for that single scoring formula (either Normal, Modified Normal, Modified Poisson or Custom). The parameters impact each distribution in very different ways. For example, 110% handicap for our custom distribution is totally reasonable, whereas for Modified Poisson, it doesn’t make any sense at all.
Here is a video walk-thru to explain how to use the spreadsheet
We are excited to see what other data analysis enthusiasts, like @StanfordBlockchain and others come up with! Your thoughts and insights will absolutely help us choose our final distribution.
We will also post our proposed algorithms here as well. Once some proposals come through, feel free to review them here on the forum AND FORK THEM. It might be easier to iteratively improve other people’s algo than starting with our dashboard (which is a little raw).