The last few years have seen massive hype and massive paranoia over Artificial Intelligence. While AI continues to progress rapidly, the hype is beginning to subside. There’s greater realism about the speed of advance of fields like driverless cars, and about just how long human drivers are likely to be needed alongside algorithms. The fears are also being modulated too, as it’s recognised that AI will more often reshape jobs rather than replacing them entirely.

The future of minds and machines: how AI can enhance collective intelligence

Explore the full report

feature page thumbnails-21.png

This more nuanced position is also opening up important new thinking on the relationship between AI and Collective Intelligence: specifically, interest is turning to how AI can help large groups to think together rather than providing an alternative to them. Over the last four months, we’ve started mapping out this new field - what AI-enabled Collective Intelligence (CI) looks like at the moment, where the main opportunities for innovation are, which challenges to be aware of and how these could be managed. In this blog, which provides the background to a panel discussion on the AI/CI relationship at our 2019 Collective Intelligence Conference, we explore some early lessons from our research.

There are very few fields where AI on its own is likely to be able to solve complex problems. But AI can be useful in providing inputs - for example, making predictions about likely patterns of climate change. And AI can help large groups to deliberate, think and decide more effectively, increasing both the depth and breadth of human collective intelligence.

This emerging field - which has had much less attention and funding than the use of AI in prediction and analysis -is our main focus here. As we show AI can help groups think more sharply - whether charities, companies, parliaments or cities - and it could help solve the problems of scale that often impede efforts at large scale deliberation and democracy.

The piece summarises some of the promising examples and shows where they could be heading, from humanitarian aid to science, and shows the future potential for other areas of CI such as digital democracy. It shows that the field is now going beyond the limitations of a previous generation of work on human-computer interfaces and design that focused primarily on how individuals could interact with AI (eg doctors, lawyers or teachers) focusing instead on groups.

While presenting the opportunities for AI to augment and amplify collective intelligence (CI) - particularly the issues of handling scale - it also sets early lessons on some of the design tensions that anyone experimenting with combinations of AI and CI needs to be aware of.

CI is already shaping most of AI

AI today is much more dependent on people and crowdsourcing than is often appreciated. A huge amount of hidden collective labour contributes to the latest advances in AI, and specifically, machine learning. Supervised approaches to machine learning rely on large labelled datasets as their training material. Whether these activities are compensated through crowd-work platforms such as Amazon Mechanical Turk and FigureEight or obtained through internet traffic with tools such as reCAPTCHA, they often remain the untold story behind the success of AI. Flipping this on its head, many crowd-labelling collective intelligence initiatives have thus naturally found themselves in possession of vast datasets that are suitable training material for AI tools. But this is just one of the many types of interactions that are possible between AI and the crowd.

How AI can enable CI

In Kenya and Uganda, more than 1 million farmers are members of the WeFarm platform which allows small-scale farmers to ask each other questions on anything related to agriculture and then receive crowdsourced bespoke content and ideas from other farmers around the world within minutes. To deal with the more than 40,000 daily questions and answers the network uses Natural Language Processing based on machine learning to best match questions to the responders within the network.

WeFarm and the other examples in this feature are good illustrations of how AI can enable CI.

At its simplest collective intelligence can be understood as the enhanced capacity that is created when people work together, often with the help of technology, to mobilise a wider range of information, ideas and insights. Within this CI-based approaches provide new approaches and opportunities in four ways through increasing our ability to:

  • Understand Problems Generate contextualised insights, facts and information on the dynamics of a situation.
  • Seek Solutions Find novel approaches or tested solutions from elsewhere. Or incentivise innovators to create new ways of tackling the problem.
  • Decide and Act Make decisions with, or informed by, collaborative input from a wide range of people and/or relevant experts.
  • Learn and Adapt Monitor the implementation of initiatives by involving citizens in generating data, and share knowledge to improve the ability of others.

Our playbook for collective intelligence design explores the wider opportunities in these four categories of collective intelligence in more detail.

A common challenge in tech-enabled CI-based solutions is scale and making sense of the different people, ideas, contributions and different types of data within a network. Especially as many solutions rely on multiple contributions of data from different sources, such as combining satellite, weather station and citizen-generated data to better understand changes to the environment. While this is primarily a software engineering challenge, once these data are cleaned and organised they can provide a rich input for current “data-hungry” AI methods such as natural language processing, computer vision and speech and audio processing.

Recent advances in these methods could play a significant role in enhancing CI by increasing the efficiency and scale of data processing, making more accurate predictions about future events or identifying new patterns and relationships between datasets. This combination of CI and AI can help facilitate more timely reactions and decision making as well as a more nuanced understanding of the complex dynamic of situations and how they change in real-time. The table below summarises some of our early thoughts on the mapping between CI challenges and how AI might be used to overcome them.

A table describing mapping between CI challenges and how AI might be used to overcome them

A table describing mapping between CI challenges and how AI might be used to overcome them. Download this table as a PDF

Five types of AI-CI interaction

At present, there is no established framework for understanding the interaction between AI and CI. Our efforts to map existing practice and academic research have indicated at least five ways we can begin to understand this relationship, which we describe below (we take a lot of inspiration from this great resource by Rai et al on digital labour platforms). Although these categories of interaction will undoubtedly grow as the field evolves, we hope that they can act as a useful starting point for those interested in exploring the current AI-enabled CI landscape and future opportunities.

Machines and groups of people taking turns to solve problems together

In this form of interaction people knowingly interact with each other and the AI but take turns to solve a task. This is done either by combining different capabilities of human and machine intelligence or through a system of feedback loops between the crowd and AI (to allow for continuous improvement of the system).

One example of this is the Early Warning Project which uses both crowd forecasting and statistical modelling to generate predictions about the risk of mass atrocities worldwide. In combination, the methods offer complementary insights and counterbalance each other’s weaknesses.

Machines working on tasks and data generated by crowds and sensors

In this second type of interaction networks of humans and sensors passively generate or actively collect data that is used as the input for a machine learning algorithm. Insights and lessons from this analysis are then utilised by the wider community of users of the platform to generate new knowledge.

This interaction also sometimes makes use of crowd-microtasking to generate labelled datasets as a training input for supervised ML models or use unsupervised AI methods to produce structured, organised insights.

Examples of this category of AI-CI interaction range from large-scale projects with active participation from users such as various projects on Zooniverse and MapwithAI to projects like OneSoil where the project makes sense of passively gathered sensor data.

People and machines solving tasks together at the same time

Instead of taking turns, this form of collaboration happens in real time where AI and humans both contribute to the same task at the same time. The generative design software for collaborative design developed by Autodesk is one example of this. In this case the AI gives designers and other users real time suggestions for different possible permutations of a solution and design alternatives based on the parameters that it is given. If the designer changes the parameters, by for example changing the width of a room, the AI will generate a new set of alternative design options.

Enabling better matching and search within a collective

AI can also play a vital role in enabling more efficient and streamlined collective intelligence projects by helping people better navigate lots of different kinds of information and tasks.

In this type of interaction, AI is used for back-end functionanility to improve the experience of individuals on online platforms. This can be achieved in many different ways, for example by better matching with others who have common interests, optimising search functions (SyrianArchive) or optimising training processes and task assignment for human contributors to citizen science projects (GravitySpy on Zooniverse). We see this type of AI contribution as “greasing the wheels” of a CI process.

Using CI to audit and support the development of better AI

Finally CI initiatives can be used to support the collaborative or competitive development of AI tools and use crowd contributions to ensure that these are better and fairer.

Example of this include online challenges that focus on AI development to solve a challenge such as the DeepFakesDetection Challenge and MalmoCollaborativeAI which is specifically set up as a game to reward the development of more collaborative AI. Another example is augmenting AI by providing more diverse data for training of AI which can turn lead to the development of AI tools that are more aligned with the public interest. Mozilla's Commonvoice project , for example, is creating an entirely new dataset driven by crowdsourcing voice contribution and validation to create an audio ML model that is both more transparent and representative of the population.

Don’t lose the crowd - design tensions at the AI-CI intersection

While there are many opportunities, the combination of AI and CI also brings with it a number of design tensions. We offer a preview of some of the main trade-offs below, with particular attention to the impact that AI could have on motivation to participate, crowd dynamics and responsibilities in high-stakes contexts.

Deploying algorithms for project efficiency vs considering the potential impact on engagement of volunteers/participants

Several examples of AI integration in CI projects perform tasks that may previously have been performed by volunteers. The risk of dis-incentivising the volunteers, by making the tasks too hard or making tasks monotonous, has been highlighted in particular by citizen science projects such as the Zooniverse platform and Cochrane Crowd. Apart from the risk of damaging relationships with volunteers who have dedicated large amounts of time (sometimes years) to helping these projects, there is also the potential loss of auxiliary social impact e.g. science education in the case of citizen science, through increased automation of microtasks.

Balancing the trade-off between designing for speed vs careful and deliberate deployment

Algorithms typically optimise for accuracy and/or speed. In the context of CI prioritising for these characteristics may not always be as relevant. For example, when citizens are brought together to discuss contentious or complex issues, as is the case in digital or deliberative democracy initiatives, you may choose to optimise for transparency and inclusiveness, which may actually slow down the process.

Developing algorithms with the highest possible accuracy vs something that is “good enough” to act as a triage alongside crowds

In AI and software development in both industry and academia, an obsession with optimising tools can distract from other drivers that should be weighted equally as optimisation criteria. For example, iterative incremental improvements to the accuracy of the algorithms can be among the primary outcome measures that drives investment and allocation of resources. However, within CI projects the AI can be used alongside significant human contributions, for example with additional verification by experts and the crowd. For CI projects the priority will therefore often be to get a working AI tool with “high enough” accuracy that enables a tangible improvement to scaling issues rather than needing to invest extra resources in ongoing refinement of a tool for an extra 1-2% accuracy.

CI-based solutions always require bringing AI into the “messy” real world where the stakes are high

Some of the most cutting-edge AI methods are currently being developed and tested in closed lab settings or in industry contexts where interpretability is less of a priority than methodological advances and datasets can lack real-world messiness.

One of the most promising techniques at the AI research frontier, deep learning, is often criticised for its evasion of interrogation. The scale of deployment typically seen in real-world CI initiatives increases the potential risk when things go wrong, which can happen when methods from the lab are brought into messy real-world contexts. This places a higher burden of responsibility on CI project leads to ensure that the AI tools used within them are well understood. Our research has highlighted cases where deep learning methods were trialled and discarded in favour of interpretable classical ML approaches in order to meet the stricter accountability norms imposed by working with the public sector. This risk aversion is understandable but might also be preventing more imaginative exploration of AI-CI interaction.

What happens when AI-CI goes wrong

Some of the most common failures of successful AI integration stem from forgetting to adequately consider ongoing human interactions and group behaviour when deploying AI tools. Examples of initiatives that didn’t work as intended include Google Flu Trends, which was hailed as a success of search-query-scraping before it emerged that it was vulnerable to overfitting and changes in search behaviour. Citizen science efforts that focus entirely on building an automated tool can fail to consider volunteer needs such as adequate training on tasks, which in return yields poor data quality insufficient for training AI. Even well intentioned projects such as the “Detox” tool developed by Google Jigsaw and Wikipedia that used crowdsourcing in combination with machine learning to identify toxic comments may only be effective for short periods until “bad actors” figure out how to counteract them. This vulnerability to gaming is a common feature of automated methods when they are not updated frequently enough to remain sensitive to a shifting context.

Finally, current “data-hungry” AI methods rely on large amounts of clean, machine-readable data. Even in public sector contexts where such data exists, deployment of AI encounters long delays if there are many stakeholders involved in negotiating the data sharing that is necessary. For example, the New York City Fire Department, which has long promised an enhanced AI-enabled version (Firecast 3.0) of its model for predicting fire risks has faced many difficulties due to organisational culture.

Making the most of the AI+CI opportunity

Between the £1 billion allocated in the 2018 industrial strategy and the more than £800 Million raised by AI companies in the first half of 2019 AI remain one of the most well-funded areas for technology research and development. However, in spite of this little funding is going towards CI opportunities.

At Nesta, we have been exploring how this could be done. In April 2019 we announced our first 12 grants for collective intelligence experiments. Earlier this month we, in partnership with our co-funders Wellcome Trust, Cloudera Foundation and Omidyar Network, launched our second fund making an additional £500,000 available to organisations interested in experimenting with AI/CI based solutions.

Others have also begun developing a research agenda in this space, with a growing interest from academia in exploring some of the more interesting and imaginative interactions between humans and machines. Examples of this include GovLab’s recent study on identifying citizens’ needs by combining AI and CI , the work of the MIT Centre for CI and studies of the impact of using machine learning on the Zooniverse citizen science platform. However, amongst the billions spent on AI in the UK and internationally, studies like these are a rarity.

Case studies

AI and Collective Intelligence: case studies

Explore 20 case studies that bring together AI and collective intelligence

feature page thumbnails-24.png

We have just started exploring this area and will be publishing more on this project in the coming months as we continue our research. If you’re interested in finding out more, would like to share comments or are interested in collaborating with us, we’d love to hear from you! Please comment below or contact Aleks Berditchevskaia.

A big thank you to Geoff Mulgan, Kostas Stathoulopoulos and Jack Orlick for their contributions, feedback and questions, all of which made this blog infinitely better.

Author

Aleks Berditchevskaia

Aleks Berditchevskaia

Aleks Berditchevskaia

Principal Researcher, Centre for Collective Intelligence Design

Aleks Berditchevskaia is the Principal Researcher at Nesta’s Centre for Collective Intelligence Design.

View profile
Peter Baeck

Peter Baeck

Peter Baeck

Director of the Centre for Collective Intelligence Design

Peter leads work that explores how combining human and machine intelligence can develop innovative solutions to social challenges.

View profile