How Good Are We at Assessing the Trustworthiness of LLMs?

Mathilde Smidt, Olivia Figge Anegaard, Anders Søgaard*

*Corresponding author af dette arbejde

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningpeer review

Abstract

Chain-of-thought prompting has been proposed as a technique that would increase people's trust in instruction-tuned large language models such as ChatGPT-3.5 or LLaMA-2. We find, somewhat surprisingly, that while people prefer chain-of-thought explanations, such explanations only increase trust when they are not read, but decrease trust when they are read. Moreover, the question type is predictive of trust. Across these two psychological biases, much of people's trust in instruction-tuned large language models seems independent of the content of their responses.

OriginalsprogEngelsk
TitelSocial Robots with AI : Prospects, Risks, and Responsible Methods, Proceedings of Robophilosophy 2024
RedaktørerJohanna Seibt, Peter Fazekas, Oliver Santiago Quick
ForlagIOS Press BV
Publikationsdato2025
Sider238-243
ISBN (Elektronisk)9781643685670
DOI
StatusUdgivet - 2025
Begivenhed6th Social Robots with AI: Prospects, Risks, and Responsible Methods Robophilosophy, RP 2024 - Hybrid, Aarhus, Danmark
Varighed: 19 aug. 202423 aug. 2024

Konference

Konference6th Social Robots with AI: Prospects, Risks, and Responsible Methods Robophilosophy, RP 2024
Land/OmrådeDanmark
ByHybrid, Aarhus
Periode19/08/202423/08/2024
NavnFrontiers in Artificial Intelligence and Applications
Vol/bind397
ISSN0922-6389

Bibliografisk note

Publisher Copyright:
© 2025 IOS Press. All rights reserved.

Citationsformater