File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: What Is It for a Machine Learning Model to Have a Capability?

TitleWhat Is It for a Machine Learning Model to Have a Capability?
Authors
Issue Date9-Jul-2024
PublisherThe University of Chicago Press
Citation
The British Journal for the Philosophy of Science, 2024 How to Cite?
Abstract

What can contemporary machine learning (ML) models do? Given the proliferation of ML models in society, answering this question matters to a variety of stakeholders, both public and private. The evaluation of models' capabilities is rapidly emerging as a key subfield of modern ML, buoyed by regulatory attention and government grants. Despite this, the notion of an ML model possessing a capability has not been interrogated: what are we saying when we say that a model is able to do something? And what sorts of evidence bear upon this question? In this paper, we aim to answer these questions, using the capabilities of large language models (LLMs) as a running example. Drawing on the large philosophical literature on abilities, we develop an account of ML models' capabilities which can be usefully applied to the nascent science of model evaluation. Our core proposal is a conditional analysis of model abilities (CAMA): crudely, a machine learning model has a capability to X just when it would reliably succeed at doing X if it 'tried'. The main contribution of the paper is making this proposal precise in the context of ML, resulting in an operationalisation of CAMA applicable to LLMs. We then put CAMA to work, showing that it can help make sense of various features of ML model evaluation practice, as well as suggest procedures for performing fair inter-model comparisons.


Persistent Identifierhttp://hdl.handle.net/10722/356818
ISSN
2023 Impact Factor: 3.2
2023 SCImago Journal Rankings: 1.446

 

DC FieldValueLanguage
dc.contributor.authorHarding, Jacqueline-
dc.contributor.authorSharadin, Nathaniel-
dc.date.accessioned2025-06-19T00:35:14Z-
dc.date.available2025-06-19T00:35:14Z-
dc.date.issued2024-07-09-
dc.identifier.citationThe British Journal for the Philosophy of Science, 2024-
dc.identifier.issn0007-0882-
dc.identifier.urihttp://hdl.handle.net/10722/356818-
dc.description.abstract<p>What can contemporary machine learning (ML) models do? Given the proliferation of ML models in society, answering this question matters to a variety of stakeholders, both public and private. The evaluation of models' capabilities is rapidly emerging as a key subfield of modern ML, buoyed by regulatory attention and government grants. Despite this, the notion of an ML model possessing a capability has not been interrogated: what are we saying when we say that a model is able to do something? And what sorts of evidence bear upon this question? In this paper, we aim to answer these questions, using the capabilities of large language models (LLMs) as a running example. Drawing on the large philosophical literature on abilities, we develop an account of ML models' capabilities which can be usefully applied to the nascent science of model evaluation. Our core proposal is a conditional analysis of model abilities (CAMA): crudely, a machine learning model has a capability to X just when it would reliably succeed at doing X if it 'tried'. The main contribution of the paper is making this proposal precise in the context of ML, resulting in an operationalisation of CAMA applicable to LLMs. We then put CAMA to work, showing that it can help make sense of various features of ML model evaluation practice, as well as suggest procedures for performing fair inter-model comparisons.<br></p>-
dc.languageeng-
dc.publisherThe University of Chicago Press-
dc.relation.ispartofThe British Journal for the Philosophy of Science-
dc.titleWhat Is It for a Machine Learning Model to Have a Capability?-
dc.typeArticle-
dc.identifier.doi10.1086/732153-
dc.identifier.eissn1464-3537-
dc.identifier.issnl0007-0882-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats