Mission Request made by LauNaMu and originally shared here.
Delegate Mission Request Summary
Update and expand the Impact Evaluation Framework based on the new intents set for Season 5 and diversify access to the Metrics Garden through new interfaces. This will increase the adoption of existing Impact Metrics and leverage existing and new tools to demonstrate impact, enhancing the collective’s ability to identify and support impactful contributions.
S5 Intent
This Request aligns with Intent 4:
Educate the broader community about Optimism governance and RetroPGF
Increase the resiliency of core governance infrastructure
Create user interfaces to interact with governance programs.
Create transparency and accountability for governance participants
30k OP per team, with additional rewards to be received via RetroPGF based on impact/outperformance.
Should this Foundation Mission be Fulfilled by One or Multiple Applicants
Up to 2 teams.
Submit by
To be set by Grants Council.
Selection by
To be set by Grants Council.
Start Date
March or as soon as the Council selects teams.
Completion Date
Within 2 months after the start date.
Specification
This Delegate Mission Request will update and expand the Impact Evaluation Framework in line with Season 5’s intents, making Optimism’s governance more accessible and understandable. It includes creating a new interactive interface for the Metrics Garden and hosting bi-weekly live sessions to guide projects. This will result in more impactful contributions, improved RetroPGF applications, and a better voting experience for Badgeholders.
Key Responsibilities and Deliverables
Deliverable 1: Updated Impact Evaluation Framework based on Season 5’s intents.
Deliverable 2: Stand-alone, interactive, and open-source website hosting the updated Framework.
Deliverable 3: A set of 8 evaluating criteria with defined scales per intent, composable across intents.
Deliverable 4: Awareness Roadmap and feedback tracking from the live sessions.
Deliverable 5: Live interactive interface for the Metrics Garden.
Deliverable 6: Report on learnings and improvements in the execution of this Mission and User Needs identified from both Applicant projects and Badgeholders.
How Should the Token House Measure Progress Towards this Mission
Weekly updates in the governance forum, with the following table tracking milestones, titles, deliverables, due dates, and evidence of completion:
Milestones
Title
Deliverable
Due Date
Evidence of Completion
Milestone 1
Framework Update
Updated Impact Evaluation Framework based on Season 5’s intents
[Specify Due Date]
[Link to updated framework]
Milestone 2
Interface Development
Stand-alone, interactive, and open-source website hosting the updated Framework
[Specify Due Date]
[Link to the live website]
Milestone 3
Criteria Formulation
Set of 8 evaluating criteria with defined scales per intent, composable across intents
[Specify Due Date]
[Documentation of criteria]
Milestone 4
Community Engagement
Awareness Roadmap and feedback tracking from the live sessions
[Specify Due Date]
[Summary of feedback & roadmap]
Milestone 5
Metrics Garden Accessibility
Live interactive interface for the Metrics Garden
[Specify Due Date]
[Link to the interface/proof of integration]
How Should Badgeholders Measure Impact Upon Completion of this Mission
Number of projects applying to RetroPGF4 with an Attestation for using the Metrics Garden.
Increase in the creation of on-chain impact data as per Metrics Garden descriptions.
Quantity and Quality of user feedback results from interactions with the website.
Number of Badgeholders using the composable evaluating criteria during RetroPGF4 voting as reported by Badgeholders.
Quantity and Quality of Badgeholder feedback on quality of applications participating in RetroPGF4.
Have You Engaged a Grant-as-a-Service Provider for this Mission Request
No
Has Anyone Other than the Proposing Delegate Contributed to this Mission Request
That is correct, I would not share any of the grant for the support I provide to executing teams of which I am not part of. If the project gets RetroPGF for their work, they can choose to share some of it with me but it’s up to them and there’s no expectation of this either.
My goal for explicitly mentioning that I will support anyone is so that 1) anyone can tap into the knowledge I developed while creating v.1 as they shape one of the v.2’s, 2) encourage those interested in applying to do so, knowing they will get support.
Realized I forgot to complete an edit so didn’t get to ask – so the idea here is that these teams would be able effectively to determine what is and is not important (or impactful), allowing badgeholders to get easy answers to complicated problems?
And I’m curious why you set this at two teams, not one or several.
On the audiences:
The result of these tools should also serve a broader range of Collective Members, the minimum being:
Badgeholders
RetroPGF future applicants
And it can even serve Council Members and the Milestone and Metrics SubCommitee in their reporting processes (more in terms of structure of the data being reported than the operationalization of the reports or reviews).
On the number of teams:
As I mentioned on the thread for proposals, I believe there isn’t one single right answer to Impact Evaluation, therefore it is good to have more than one team working on it, so experimentation is more efficient and moves faster.
Two would be ideal though to avoid cognitive overload and giving too many options, which can lead to decision paralysis and just be overwhelming. If there are too many options, learning from mistakes and iterating on early results (whether iterations are completed by those same teams or others) will be more complex to execute on. In addition to this, the design of the proposal is for each of the teams to build modular impact evaluation rubrics. Since rubrics can be modified and rearranged, this means that, effectively, there will be more than “2” evaluation rubrics produced by these two teams.
I hope this answers your questions and feel free to share any others that you or anyone may have.
Hey @LauNaMu , since this is a rather specific mission request, with a specific dependency/continuation of a past effort, can you provide some insight into the Impact Evaluation Framework and Metric Garden’s use so far? Where has it been successful/unsuccessful?
Thank you for your question. I have shared some initial findings here that answer some of the questions you’re sharing.
TLDR;
The Impact Evaluation Framework had valuable information that helped generate clear concepts of impact in the Optimism Ecosystem for Badgeholders, but it was hard to consume and could be overwhelming for people with limited bandwidth.
Pushing for Data driven evaluations was hard without standardized data points and while it provided an initial intuition for Badgeholders, it didn’t result in straightforward assessments, this is where the Metrics Garden comes into play to generate a bottoms up standardization on the metrics projects can use to measure their impact.
Please do share any additional questions that may come up.
The Grants Council has opened early submissions as an Indication of Interest for this mission request here
For your application to be considered, the Mission request must pass the Token House vote on February 14th. Submissions will not be considered if a Mission Request is not approved on the 14th.