July 27 Dolt Bounty retro


What went well?

  • engagement & re-engagement with community + potential supporters

  • iterated on our schema / planned and made some changes

  • we filled out the agencies table to a high level of completion → a prerequisite to the datasets table

  • excellent support from the Dolt team!

  • learned that volunteer interest plays a big role in determining our focus

  • got a start on refining the review process

  • we split the bounty into two pieces so we had an opportunity to learn

What could have gone better?

  • there's not currently an automated review process or script

  • we discussed the difference between a dataset URL and an agency homepage

    • people ended up focusing on the agency homepage URL more than the datasets

  • one person ran away with the scoreboard by using machine readable lat/long

Next steps

  • Katie to share some metrics on engagement

  • Minimize variables + types of submissions allowed, particularly differences in scraped vs. manually gathered data

  • Plan next bounty for ~8/25 start

    • Formalize the fact that our team should do the reviews + scripts

    • Smaller bounties help us iterate on a faster loop

    • Look at attribution + scoreboard code and make contributions to how rewards are given out.

    • Announce $50/participant (pending Katie + Dolt approval)

Notes for bounty success

  1. Pick a common schema for data (i.e. "incident reports")

  2. Allow people to submit data linking back to the source (i.e. dataset)

  3. Focus on discrete, scrapeable goals with relevant topics

    • e.g. hospital bounty was in response to a law which passed which required hospitals to publish data

    • national scope for a narrow data focus is a potential alternative to the narrow geo focus

      • helps with schema normalization + big stories + big moves

      • local is storytelling → action loop

  4. Focus on additions vs edits, enforceable by requiring properties

Last updated