1887

AI and the Future of Skills, Volume 2

Methods for Evaluating AI Capabilities

image of AI and the Future of Skills, Volume 2

As artificial intelligence (AI) expands its scope of applications across society, understanding its impact becomes increasingly critical. The OECD's AI and the Future of Skills (AIFS) project is developing a comprehensive framework for regularly measuring AI capabilities and comparing them to human skills. The resulting AI indicators should help policymakers anticipate AI’s impacts on education and work.

This volume describes the second phase of the project: exploring three different approaches to assessing AI. First, the project explored the use of education tests for the assessment by asking computer experts to evaluate AI’s performance on OECD’s tests in reading, mathematics and science. Second, the project extended the rating of AI capabilities to tests used to certify workers for occupations. These tests present complex practical tasks and are potentially useful for understanding the application of AI in the workplace. Third, the project explored measures from direct AI evaluations. It commissioned experts to develop methods for selecting high-quality direct measures, categorising them according to AI capabilities and systematising them into single indicators. The report discusses the advantages and challenges in using these approaches and describes how they will be integrated into developing indicators of AI capabilities.

English

Assessing AI capabilities with education tests

This chapter introduces three exploratory studies that assessed the capabilities of artificial intelligence (AI) through standardised education tests designed for humans. The first two studies, conducted in 2016 and 2021/22, asked experts to evaluate AI’s performance on the literacy and numeracy tests of the OECD’s Survey of Adult Skills (PIAAC). The third study collected expert judgements of whether AI can solve science questions from the OECD's Programme for International Student Assessment (PISA). The studies aimed to refine the assessment framework for eliciting expert knowledge on AI using established educational assessments. They explored different test formats, response methodologies and rating instructions, along with two distinct assessment approaches. A “behavioural approach” used in the PIAAC studies emphasised smaller expert groups engaging in discussions, and a "mathematical approach" adopted in the PISA study relied more heavily on quantitative data from a larger expert pool. This chapter presents the results of the studies and discusses the advantages and disadvantages of their methodological approaches.

English

Graphs

This is a required field
Please enter a valid email address
Approval was a Success
Invalid data
An Error Occurred
Approval was partially successful, following selected items could not be processed due to error