July 27 Dolt Bounty retro
https://www.dolthub.com/repositories/pdap/datasets/bounties/default
What went well?
engagement & re-engagement with community + potential supporters
iterated on our schema / planned and made some changes
we filled out the agencies table to a high level of completion → a prerequisite to the datasets table
excellent support from the Dolt team!
learned that volunteer interest plays a big role in determining our focus
got a start on refining the review process
we split the bounty into two pieces so we had an opportunity to learn
What could have gone better?
there's not currently an automated review process or script
we discussed the difference between a dataset URL and an agency homepage
people ended up focusing on the agency homepage URL more than the datasets
one person ran away with the scoreboard by using machine readable lat/long
Next steps
Katie to share some metrics on engagement
Minimize variables + types of submissions allowed, particularly differences in scraped vs. manually gathered data
Plan next bounty for ~8/25 start
Formalize the fact that our team should do the reviews + scripts
Smaller bounties help us iterate on a faster loop
Look at attribution + scoreboard code and make contributions to how rewards are given out.
Announce $50/participant (pending Katie + Dolt approval)
Notes for bounty success
Pick a common schema for data (i.e. "incident reports")
Allow people to submit data linking back to the source (i.e. dataset)
Focus on discrete, scrapeable goals with relevant topics
e.g. hospital bounty was in response to a law which passed which required hospitals to publish data
national scope for a narrow data focus is a potential alternative to the narrow geo focus
helps with schema normalization + big stories + big moves
local is storytelling → action loop
Focus on additions vs edits, enforceable by requiring properties
Last updated