Make it make sense: the challenge of data analysis in global deliberation
The deliberative community is just beginning to experiment with how to do this, and only through collaborative spirit will we arrive at processes that can make a difference. In my role leading the data strategy for Iswe Foundation, I am supporting their efforts to convene a coalition and institute a permanent global citizen’s assembly, starting next year.
Massifying participation will be key to invigorating global deliberation. Assemblies will have a better chance of being seen as legitimate, fair, and publicly supported if they involve thousands or even millions of diverse participants. This raises an operational challenge: how to systematise political ideas from many people across the globe.
In a centralised global assembly, anything from 50 to 500 citizens from various countries engage in a single deliberation and produce recommendations or political actions by crossing languages and cultures. In a distributed assembly, multiple gatherings are convened locally that share a common but flexible methodology, allowing participants to discuss a common issue applied both to local and global contexts. Either way, a global deliberation process demands the organisation and synthesis of possibly thousands of ideas from diverse languages and cultures around the world.
How could we ever make sense of all that data to systematise citizens’ ideas and recommendations? Most people turn their heads to computational methods to help reduce complexity and identify patterns. First up, one technique for analysing text amounts to little more than simple counting, through which we can produce something like a frequency table or a wordcloud.
This is why, in most cases, participation analysts combine the Big Data methods with what I call Little Data (inspired by Norvaisas & Karpfen and mentor Coni Miranda)- examples, quotes, stories, photos, individual ideas, and other anecdotal insights that more richly illuminates citizens’ perspectives. This means walking a thin line between overwhelming readers with too much information (if we just focus on the Little Data) and losing the nuance that makes deliberation worthwhile (if we just focus on the Big Data).
It’s a fine line between overwhelming readers with too much information, and losing the nuance that makes deliberation worthwhile.
How do we go about picking the right examples for the Little Data? At least in my experience, in many cases the organiser or researcher just selects a bunch of examples based on their preferences. In other cases, people use statistical procedures that identify sentences that are supposedly representative of discussions, known as extractive summarisation. But the outcome often lacks diversity, as it typically churns out common denominators that end up being generic and bland, rather than exploring the breadth of perspectives.
Other projects have now also started using Large Language Models (LLMs) to actively create “exemplars” of citizen stances (abstractive summarisation). Under this approach, participation analysts can use pre-trained models (including commercial ones such as ChatGPT) by providing all the relevant text and asking it to produce a summary or to describe groups of opinions. However, it doesn’t make sense to fabricate amalgamated perspectives when we have actual citizen ideas that can be traced back to their local context.
Projects like Talk to The City and Fora from Cortico utilise all of these methods: statistical modelling, LLMs, and also doing their best to share unadulterated quotes and videos. This approach is definitely promising but is still in development. For example, we still do not know how trustworthy LLMs are in summarising text.
There are also more creative options. If we think of the themes or most prevalent words we identified through statistical modelling as initial filters, a more participatory, decentralised approach could work. If we believe in the democratic importance of decentralisation and participation, our data analysis strategy should also refrain from leaving all the data interpretation to a single organisation. For example, if we identify in our analysis that the concept of “clean water” is prevalent, we could take all assembly outputs that refer to that concept and send them to NGOs worldwide that specialise in clean water as “guest curators” of their top ideas.
We could even engage with citizen science approaches and have contributors help with flagging and ordering the data. This means we can engage volunteers to help analyse the data by letting them organise, classify and select ideas to show. A very similar approach is being developed by Fora in inviting as many people as possible to become “sensemakers” in their projects.
About the Author
Iñaki Goñi is a data associate at the Iswe Foundation and a doctoral candidate in Science and Technology Studies at the University of Edinburgh, where he studies technology and democracy. He works in large-scale participation and participatory technologies from a critical perspective.
Supporters
The Journal of Deliberative Democracy and Deliberative Democracy Digest are supported by:
Contact
General queries
Please get in touch with our editor Lucy Parry.
Mailing Address
Journal of Deliberative Democracy
Centre for Deliberative Democracy and Global Governance
Ann Harding Conference Centre
University Drive South
University of Canberra, ACT 2617