On the Reliability of Large Language Models to Misinformed and Demographically-Informed Prompts

11 Pages Posted: 26 Jan 2024 Last revised: 29 Apr 2024

See all articles by Toluwani Aremu

Toluwani Aremu

Mohamed bin Zayed University of Artificial Intelligence

Oluwakemi Akinwehinmi

University of Lleida

Chukwuemeka Nwagu

Dalhousie University

Syed Ishtiaque Ahmed

University of Toronto - Department of Computer Science

Rita Orji

Dalhousie University

Pedro Arnau Del Amo

University of Lleida

Abdulmotaleb El Saddik

University of Ottawa

Abstract

We investigate and observe the behaviour and performance of Large Language Model (LLM)-backed chatbots in addressing misinformed prompts and questions with demographic information within the domains of Climate Change and Mental Health. Through a combination of quantitative and qualitative methods, we assess the chatbots' ability to discern the veracity of statements, their adherence to facts, and the presence of bias or misinformation in their responses. Our quantitative analysis using True/False questions reveals that these chatbots can be relied on to give the right answers to these close-ended questions. However, the qualitative insights, gathered from domain experts, shows that there are still concerns regarding privacy, ethical implications, and the necessity for chatbots to direct users to professional services. We conclude that while these chatbots hold significant promise, their deployment in sensitive areas necessitates careful consideration, ethical oversight, and rigorous refinement to ensure they serve as a beneficial augmentation to human expertise rather than an autonomous solution.

Note:

Funding Information: We declare that our study received no external funding. It was conducted independently by the authors.

Conflict of Interests: We confirm that there are no competing interests among the authors. This has been explicitly stated in a document attached to our submission to the journal "Computers in Human Behavior." We reiterate here that there are no financial, personal, or professional conflicts that could be construed as influencing the research.

Keywords: Large Language Models, AI for Climate Change, AI for Mental Health, Responsible AI

Suggested Citation

Aremu, Toluwani and Akinwehinmi, Oluwakemi and Nwagu, Chukwuemeka and Ishtiaque Ahmed, Syed and Orji, Rita and Arnau Del Amo, Pedro and Saddik, Abdulmotaleb El, On the Reliability of Large Language Models to Misinformed and Demographically-Informed Prompts. Available at SSRN: https://ssrn.com/abstract=4699451 or http://dx.doi.org/10.2139/ssrn.4699451

Toluwani Aremu (Contact Author)

Mohamed bin Zayed University of Artificial Intelligence ( email )

United Arab Emirates

Oluwakemi Akinwehinmi

University of Lleida ( email )

Chukwuemeka Nwagu

Dalhousie University ( email )

6225 University Avenue
Halifax, B3H 4H7
Canada

Syed Ishtiaque Ahmed

University of Toronto - Department of Computer Science ( email )

Sandford Fleming Building
King’s College Road, Room 3302
Toronto, Ontario M5S 3G4
Canada

Rita Orji

Dalhousie University ( email )

6225 University Avenue
Halifax, B3H 4H7
Canada

Pedro Arnau Del Amo

University of Lleida ( email )

Victor Siurana 1
Lleida, 25003
Spain

Abdulmotaleb El Saddik

University of Ottawa ( email )

2292 Edwin Crescent
Ottawa, K2C 1H7
Canada

Do you have a job opening that you would like to promote on SSRN?

Paper statistics

Downloads
128
Abstract Views
435
Rank
424,879
PlumX Metrics