Gamification in metadata creation - how do we show “quality” and encourage improvement?

Gamification in metadata creation - how do we show “quality” and encourage improvement?

This post also available on my personal blog.

Encouraging the creation of good metadata can be a challenging exercise. Systems for metadata creation need to allow for blank fields to allow incremental or in progress content to be saved, however some fields may be semantically recommended or mandatory for standardisation. So, while a metadata editing tool needs to be flexible to allow evolving content, it also needs to provide feedback to drive improvement.

This leads to three questions:

  1. Can we automatically measure metadata quality?
  2. Can we use this data to encourage metadata editors to more actively participate and improve content?
  3. How can we best show the “quality” of metadata to an editor to achieve better content?

Gamification is a recognised term for encouraging changes in user behaviour through small incentives. The question I'd like to pose for Aristotle is, how can these principles be used to encourage the creation of good metadata. Obviously, 'good' is very subjective, but for metadata objects such as a Data Elements, at the bare minimum having an attached "Data Element Concept" and "Value Domain" is a prerequisite for a quality item. Likewise, a piece of metadata with a description is good, but a piece of metadata with a longer description is probably better (but not always which leads to further challenges).

For the moment, let’s assume that basic metrics for "good" metadata can be constructed and applied to a piece of metadata, and that these can be summed together to give a raw score based on the possible sum. This assumption means we can grade metadata completion and get a raw score like "77 passes out of a possible 82". From these we can derive all sorts of graphics or figures that can influence user behaviour, and it’s that which I'm interested in right now.

First of all, from these figures we can derive a percentage or rank - 77 out of 82 is about 94%, "9/10", "4.5/5" or "A-". This may mean a metadata item has all its relations, all fields are filled out, but one or two are a little shorter than our metrics would like. Perhaps though, the item is described perfectly - adding more text in this case is worse.

Secondly, there is the issue of how to present this visually once we've determined a score. There are probably many ways to present this, but for now I want to focus on two - symbols and progress bars. A symbol can be any repeated small graphic, but the best example is a star ranking where metadata is given a rank out of 5 stars.

Once a raw score is computed, we can then normalise this to a score out of 5 and show stars (or other symbols). However, initial discussions suggest that this presents a more abrupt ranking that discourages work-in-progress, rather than displaying further work to do.

An alternative is the use of progress bars to show the completion of an item. Again, this is computed from the raw score and normalised to a percent and then shown to the user. The images show different possible options including a percentage complete, an integer or rounded decimal rankings out of 10. Again, initial discussions suggest that percentages may encourage over work, where users on 94% might strive for 100% by 'gaming the system', opposed to users with metadata ranked 9.5/10. For example, if a metadata item has a well written short description but is under a predefined limit resulting in a score of 94%, we need to design a pattern to discourage an editor from ‘adding’ content free text to ‘score’ 100%. The use of colour is also a possible way to gauge progress analogous to stars that many users are familiar with, but raises the questions of how to define ‘cut-offs’ for quality.

Metadata quality tracking in practice

In this section we look at a number of possible options for presenting the quality metrics of metadata using the Aristotle Metadata Registry. At the moment these are just mock-ups, but sample work has shown that dynamic analysis of metadata to quantify “quality” is possible, so here we will address the matter of how to show this.

First of all, it is important to note that these quality rankings can be shown at every stage from first edit all the way through to final publication, so a one-size-fits-all approach may not be the best way. In the simple case, we can look at the difference between progress and a star rating. Alongside with all the basic details, metadata can be given a ranking right on the page as well as a status to give users immediate feedback on its fitness for use.

Secondly, we can look at simple presentation options. Here it’s important to note only one rating would be shown out of all the possible options. Stars offer fewer levels of granularity, and when coloured are bright and distracting. However, progress bars blend quite well, even when coloured, and give more options for embedding textual representations.

Lastly, we can see how these would look when a number of items are shown together. Using a ‘sparklines’ approach, we can use stars or bars to quickly highlight trouble spots in metadata when looking at a large number of items.

For the professional context of a registry, based on initial feedback, there are strengths in the progress bar style that make it more suited for use, however more feedback is required to make a conclusive argument.

Conclusion

This is intended to be be the first in a number of articles that address the issue of best practices in presenting “quality” in the creation of metadata and the implementation of these practices in the Aristotle Metadata Registry. As such, I welcome and encourage comments and feedback into this design process, both from Aristotle users and the broader community.

Key questions for feedback

  1. Question 1: How can we textually show a metadata quality rank to encourage more participation? Possible options: raw values (77/82), percent (94%), normalised (9/10), graded (A-) or something else?
  2. Question 2: How can we visually show metadata rank to encourage participation? Possible options: stars (or others symbols), progress bar, colours, text only or something else?
  3. Question 3: How do we positively encourage completion without adversely encouraging “gaming the system”?
  4. Future questions:
    How do we programmatically measure metadata quality?
    Based on a set of sub-components of quality how and when can we show a user how to improve metadata quality?
Charles Rosenbury

Designer, Architect, Philosopher

8y

I always like informative dashboards. And I like that part. I think the "gamification" concepts presented are meager. Gamification is not about "changes in user behaviour through small incentives." Gamification is about recognizing that individual people have different motivators, and trying to design a system which grants that level of acknowledgement. That includes both rewards and "punishments." Having a dashboard to record status is, in my opinion, a very small part of the needed solution. And it may even drive the wrong behavior if not being very careful. We get what we measure. Gamification is about making it "fun." Making people enjoy the work by making it an interesting effort. I don't really see that in the proposal.

Like
Reply
Gillian Kerr

Evaluation systems, LogicalOutcomes

9y

I really like this approach to include graphic representations of metadata quality within the dictionary. The colours alone encourage users to find out what 'quality' means, especially if it's red. Re the elements of metadata quality: there is (1) technical adequacy of the description (whether the method of measurement is accurate and understandable), (2) the technical quality of the indicator (whether it has been validated in terms of reliability, clarity etc.) and (3) the quality of the indicator in terms of a validated measure of the outcome it is supposed to be demonstrating. A measure can be a great indicator of one outcome and a terrible indicator of another outcome. Your general metadata quality rating could combine (1) and (2), but there would have to be different ratings to measure the quality of individual indicators in relation to individual outcomes. That second set of ratings is too complex to be included in Aristotle at this point in development, but it would be great if working groups could have the ability to add that field themselves and search on it. Re your questions: 1. I like the look of your bars, but the stars might be more feasible to rate. You could use ratings like the ones from evidence-based medicine: Evidence levels from 1 to 5. 2. Stars and progress bars are both good. Will think about the rest. - Gillian

Alan Pauza

Director at Asura Group | Information System Design | Business Architecture | Data Analytics | Program Management

9y

Very interesting perspective Sam - I'd considered aspects of gamification to help improve the quality of collected data vs metadata standards - that is - how near or far are we from 'error free' data submission for a particular NMDS - but hadn't thought of gamification of the actual metadata items. I guess I see the biggest 'system challenge is the appropriate adoption and use of ANY metadata; there is an implicit underlying assumption that metadata needs to conform to a 'baseline acceptance standard' before being approved for use.

Like
Reply

To view or add a comment, sign in

More articles by Samuel Spencer

Insights from the community

Others also viewed

Explore topics