Can ChatGPT Decipher Fedspeak?

28 Pages Posted: 7 Apr 2023 Last revised: 11 Apr 2024

See all articles by Anne Lundgaard Hansen

Anne Lundgaard Hansen

Federal Reserve Bank of Richmond - Quantitative Supervision & Research

Sophia Kazinnik

Federal Reserve Banks - Quantitative Supervision & Research

Date Written: April 10, 2024

Abstract

This paper investigates the ability of Generative Pre-training Transformer (GPT) models to decipher Fedspeak, a term used to describe the technical language used by the Federal Reserve to communicate on monetary policy decisions. We evaluate the ability of GPT models to classify the policy stance of Federal Open Market Committee announcements relative to human assessment. We show that GPT models deliver a considerable improvement in classification performance over other commonly used methods. We then demonstrate how the GPT-4 model can provide explanations for its classifications that are on par with human reasoning. Finally, we show that the GPT-4 model can be used to identify macroeconomic shocks using the narrative approach of Romer and Romer (1989, 2023).

Keywords: Natural Language Processing (NLP), Generative Pre-training Transformer (GPT), Federal Reserve Communication, Applications, Artificial Intelligence (AI)

JEL Classification: E52, E58, C88

Suggested Citation

Hansen, Anne Lundgaard and Kazinnik, Sophia, Can ChatGPT Decipher Fedspeak? (April 10, 2024). Available at SSRN: https://ssrn.com/abstract=4399406 or http://dx.doi.org/10.2139/ssrn.4399406

Anne Lundgaard Hansen

Federal Reserve Bank of Richmond - Quantitative Supervision & Research ( email )

United States
7734407018 (Phone)

HOME PAGE: http://sites.google.com/view/anneh/

Sophia Kazinnik (Contact Author)

Federal Reserve Banks - Quantitative Supervision & Research ( email )

United States

Do you have negative results from your research you’d like to share?

Paper statistics

Downloads
8,977
Abstract Views
27,821
Rank
1,247
PlumX Metrics