Notes from Scorecard BOF in Paris

Attendees: Alan Berg, Seth Theriault Peter Knoop, Gilamesh Noteboos, Jean-Fracious Leveque, Aaron Zercoski, Linda Place, Chris Kretler, Mark Norton, Oliver Heyer, Clay Fenlason, Megan May
Absent: David Haines
Purpose: Formulate WG to define Scorecard (ratings) and structured description. This documentation will serve two purposes
1) Advise Release Mgmt team on what tools to bundle as a release
2) Advise institutions in making local decisions on bundles

Currently we have

  • contrib - in release/turned on
  • provisional - in release/turned off
  • core - in release
    This does not adequately speak to the quality of the tools and that is why we have convened.

Criteria should be SMART
Specific
Measurable
Attainable
Relevant
Timely

Stephen envisions 5 summary ranges (red orange yellow, green, gold) for different dimensions to these scorecards. Envision that we should have 3-7. Take the details in these, roll them up and get a top level score.

1) DIMENSIONS DEFINED
------------------------------------------------------------------------------------------
User Experience (usability, accessibility, internationalization)

  • Consistency / best practices - this must be moving forward. not the status quo
  • Usability
  • Accessibly
  • ease of use
  • user testing (ie placed in front of real users)

2) Technical (code review, unit tests, adheres to code standards)

  • Browser support
  • code review
  • Unit testing
  • functional regression testing
  • Integration testing
  • Performance testing
  • Internationalization
  • Licensing
  • Outstanding JIRA bugs
  • Packaging (code structure)
  • static code review
  • Validation/spec conformance
  • DB support
  • DB best practices
  • security review
  • technical
  • Event tracking
  • May be possible to further clump these (ie testing)

3) Descriptive - documented, does it have help

  • help (bundled)
  • Test plan
  • Javadocs
  • If required technical / architectural doc
  • Wiki/website
  • Deployment doc
  • End user external docs.
  • tracking tasks in issue tracking system ((ie JIRA)
  • Events documented
  • Licensing Documented
  • Sakai.properties

4) Community support

  • Team size
  • Team diversity (institution, dev/ux/qa)
  • Responsiveness (Average time to respond to JIRAS)
  • Production experience - length, scale, diversity
  • Communications and openness.

Talked a lot about internationalization b/c it's an issue that spans a number of dimensions. For instance, there are technical hurdles

  • Lot's of talk about who owns the process. It was suggested that this was to be lead by QA WG. Mark pointed out that there used to be a group that dealt with processes. Others mentioned that they weren't the owners of the process once defined. Some thought the community should own, others the project teams . Someone needs to lead the process but decision was deferred.

Mark suggested that we run through this experience on a few pilot projects. He're

  • Sousa
  • Resources (TBD)
  • Site Stats
  • Polls
  • Blogwow
    --> This lead to discussion onto handling services (content, search). Decision on how to handle these are TBD

Megan elected Stephen to lead this past the conference

Action Items
------------------------------

  • Crosscheck the list with existing documentation
  • Create Collab list (Megan will send in)
  • post notes and send message to community