The purpose of this page
…is to record some ongoing reflections on my experience of running two pre-tests of ParEvo carried out in late 2018 and early 2019.
Participants and others are encouraged to add their own comments, by using the Comment facility at the bottom of this page
Two pre-tests are underway
- One involves 11 participants developing a scenario involving the establishment of an MSC (Most Significant Change) process in a development programme in Nigeria. These volunteers were found via the MSC email list. They came from 7 countries and 64% were women.
- The other involves 11 participants developing a Brexit scenario following Britain failing to reach an agreement with the EU by March 2019. These participants were found via the MandE NEWS email list. They came from 9 countries and 46% were women.
For more background (especially if you have not been participating) see this 2008 post on the process design and this 2019 Conference abstract talking about these pre-tests
Reflections so far
Issues arising…
- How many participants should there be?
- In the current pre-tests, I have limited the number to around 10. My concern is that with larger numbers there will be too many story segments (and their storylines) for people to scan and make a single preferred selection. But improved methods of visualising the text contributions may help overcome this limitation. Another option is to allow/encourage individual participants to represent teams of people, e.g. different stakeholder groups. I have not yet tried this out.
- Do the same participants need to be involved in each iteration of the process?
- My initial concern is that not doing so would make some of the follow up quantitative analysis more difficult, but I am not so concerned about that now, its a manageable problem. On the other hand, it is likely that some people will have to drop out mid-process, and ideally, they could be replaced by others, thus maintaining the diversity of storylines.
- How do you select an appropriate topic for a scenario planning exercise?
- Ideally, it would be a topic that was of interest to all the participants and one which they felt some confidence in talking about, even if only in terms of imagined futures. One pre-test topic, the use of MSC in Nigeria, was within these bounds. But the other was more debatable: the fate of the UK after no resolution of BREXIT terms by 29th March 2019
- How should you solicit responses from participants?
- I started by sending a standard email to all the (MSC scenario) participants, but this has been cumbersome and has risks. It is too easy to lose track of who contributed what text, to add to what existing storyline. I am now using two-part single question survey via SurveyMonkey. This enables me to keep a mistake-free record of who contributed what to what, and who has responded and who has not. But this still involves sending multiple communications, including reminders, and I have sometimes confused what I am sending to whom. A more automated systems is definitely needed.
- How should you represent and share participants responses?
- This has been done in two forms. One is a tree diagram, showing all storylines, where participants can mouseover nodes to immediately see each text segment. Or they can click on each node to go to a separate web page and see complete storylines. These are both laborious to construct, but hopefully will soon be simplified and automated via some tech support which is now under discussion. PS: I have now resorted to only using the tree diagram with mouseover.
- Should all contributions be anonymous?
- There are two types of contributions: (a) the storyline segments contributed during each iteration of the process, (b) Comments made on these contributions, that can be enabled on the blog page that hosts each full storyline to date. This second type was an afterthought, whereas the first is central to the process.
- The first process of contributing to storylines designed to make authorship anonymous, so people would focus on the contents. I think this remains a good feature.
- The second process of allowing people to comment has pros and cons. The advantage is that it can enrich the discussion process, providing a meta-level to the main discussion which is the storyline development. The risk, however, is that if the comments are not enabled to be anonymous then a careful reader of the comments can sometimes work out who made which storyline contributions. I have tried to make comments anonymous but they still seem to reveal the identity of the person making the comment. This may be resolvable. PS: This option is now not available, while I am only using the tree diagram to show storylines. This may need to be changed.
- How many iterations should be completed?
- It has been suggested that participants should know this in advance, so that their story segments don’t leap in the future too quickly, or the reverse, progress the story too slowly. With the Brexit scenario pre-test I am inclined to agree. It might help to saying at the beginning that there will be 5 iterations, ending in the year 2025. With the MSC scenario pre-test I am less certain, it seems to be moving on at a pace I would not have predicted
- I am now thinking it may also be useful to spell out in advance the number of iterations that will take place. And perhaps even suggest each one will represent a given increment in time, say a month or a year, or…
- What limits should there be on the length of the text that participants submit?
- I have really wobbled on this issue, ranging from 100-word limits to 50-word limits to no voiced limits at all. Perhaps when people select which storyline to continue the length of the previous contributions will be something they take into account? I would like to hear participants views on this issue. Should there be word limits, and if so, what sort of limit?
- What sort of editorial intervention should there be by the facilitator, if any?
- I have been tempted, more than once, to ask some participants to reword and revise their contribution. I now limit myself to very basic spelling corrections, checked with the participant, if necessary. I was worried that some participants have a limited grasp of the scenario topic, but now think that just has to be part of the reality, some people have little to go on when anticipating specific the future, and others may have “completely the wrong idea”, according to others. As the facilitator, I now think I need to stand back and let things run.
- Another thought I had some time ago is that the facilitator could act as the spokesperson for “the wider context”, including any actors not represented by any of the participant’s contributions so far. At the beginning of a new iteration, they could provide some contextual text that participants are encouraged to bear in mind when designing their next contribution. If so, how / where should this context information be presented?
- How long should a complete exercise take?
- The current pre-tests are stretching out over a number of weeks. But I think this will be an exception. In a workshop setting where all participants (or teams of) have access to a laptop and internet, it should be possible to move through a quite a few iterations within a couple of hours. In other non-workshop settings perhaps a week will be long enough, if all participants have a stake in the process. Compacting the available time might generate more concentration and focus. The web app now under development should also radically reduce the turnaround time between iterations because manual work done by the facilitator will be automated.
- Is my aim to have participants evaluate the completed storylines realistic?
- After the last iteration, I plan to ask each participant, probably via an online survey page, to identify: (a) the most desirable storyline, (b) the most likely to happen storyline. But I am not sure if this will work. Will participants be willing to read every storyline from beginning to end? Or will they make judgments on the basis of the last addition to each storyline, which they will be more familiar with? And how much will this bias their judgments (and how could I identify if it does)?
- What about the contents??
- One concern I have is the apparent lack of continuity between some of the contributions to a storyline. Is this because the participants are very diverse? Or because I have not stressed the importance of continuity? Or because I can’t see the continuity that others can see?
- What else should we look for when evaluating the content as a whole? One consideration might be the types of stakeholders who are represented or referred to, and those which seem to be being ignored
- How should performance measures be used?
- Elsewhere I have listed a number of ways of measuring and comparing how people contribute and how storylines are developed. Up to now, I have thought of this primarily as a useful research tool, which could be used to analyze storylines after they have been developed.
- But after reading a paper on “gamification” of scenario planning it occurred to me that some of these measures could be more usefully promoted at the beginning of a scenario planning exercise, as measures that participants should be aware of and even seek to maximize when deciding how and where to contribute. For example, one measure is the number of extensions that have been added to a participant’s texts by other participants, and the even distribution of those contributions (known as variety and balance).
- Stories as predictions
- Most writers on scenario planning emphasize that scenarios are not meant to be predictions, but more like possibilities that need to be planned for
- But if ParEvo was used in a M&E context, could participants be usefully encouraged to write story segments as predictions, and then be rewarded in some way if they came true? This would probably require an exercise to focus on the relatively near future, say a year or two at the most, with each iteration perhaps only covering a month or so.
- Tagging of story segments
- It is common practice to use coding / tagging of text contents in other settings. Would it be useful with ParEvo? An ID tag is already essential, to be able to identify and link story segments.
- What other issues are arising and need discussion?
- Over to you…to comment below
- I also plan to have one to one skype conversations with participants, to get your views on the process and products