- Title
- AI apology : interactive multi-objective reinforcement learning for human-aligned AI
- Creator
- Harland, Hadassah; Dazeley, Richard; Nakisa, Bahareh; Cruz, Francisco; Vamplew, Peter
- Date
- 2023
- Type
- Text; Journal article
- Identifier
- http://researchonline.federation.edu.au/vital/access/HandleResolver/1959.17/197188
- Identifier
- vital:18817
- Identifier
-
https://doi.org/10.1007/s00521-023-08586-x
- Identifier
- ISSN:0941-0643 (ISSN)
- Abstract
- For an Artificially Intelligent (AI) system to maintain alignment between human desires and its behaviour, it is important that the AI account for human preferences. This paper proposes and empirically evaluates the first approach to aligning agent behaviour to human preference via an apologetic framework. In practice, an apology may consist of an acknowledgement, an explanation and an intention for the improvement of future behaviour. We propose that such an apology, provided in response to recognition of undesirable behaviour, is one way in which an AI agent may both be transparent and trustworthy to a human user. Furthermore, that behavioural adaptation as part of apology is a viable approach to correct against undesirable behaviours. The Act-Assess-Apologise framework potentially could address both the practical and social needs of a human user, to recognise and make reparations against prior undesirable behaviour and adjust for the future. Applied to a dual-auxiliary impact minimisation problem, the apologetic agent had a near perfect determination and apology provision accuracy in several non-trivial configurations. The agent subsequently demonstrated behaviour alignment with success that included up to complete avoidance of the impacts described by these objectives in some scenarios. © 2023, The Author(s).
- Publisher
- Springer Science and Business Media Deutschland GmbH
- Relation
- Neural Computing and Applications Vol. 35, no. 23 (2023), p. 16917-16930
- Rights
- All metadata describing materials held in, or linked to, the repository is freely available under a CC0 licence
- Rights
- http://creativecommons.org/licenses/by/4.0/
- Rights
- Copyright © 2023, The Author(s)
- Rights
- Open Access
- Subject
- 4602 Artificial intelligence; 4603 Computer vision and multimedia computation; 4611 Machine learning; AI apology; AI safety; Human alignment; Impact minimisation; Multi-objective reinforcement learning
- Full Text
- Reviewed
- Funder
- Open Access funding enabled and organized by CAUL and its Member Institutions.
- Hits: 2883
- Visitors: 2926
- Downloads: 99
Thumbnail | File | Description | Size | Format | |||
---|---|---|---|---|---|---|---|
View Details Download | SOURCE1 | Published version | 1 MB | Adobe Acrobat PDF | View Details Download |