Focusing on Fine-Tuning: Understanding the Four Pathways for Shaping Generative AI

29 Pages Posted: 26 Mar 2024

See all articles by Paul Ohm

Paul Ohm

Georgetown University Law Center

Date Written: February 25, 2024

Abstract

Those who design and deploy generative-AI models, such as Large Language Models like GPT-4 or image diffusion models like Stable Diffusion, can shape model behavior in four distinct stages: pretraining, fine-tuning, in-context learning, and input-and-output filtering. The four stages differ among many dimensions, including cost, access, and persistence of change. Pretraining is always very expensive and in-context learning is nearly costless. Pretraining and fine-tuning change the model in a more persistent manner, while in-context learning and filters make less durable alterations. These are but two of many such distinctions reviewed in this article.

Legal scholars, policymakers, and judges need to understand the differences between the four stages as they try to shape and direct what these models do. Although legal and policy interventions can (and probably will) occur during all four stages, many will best be directed at the fine-tuning stage. Fine-tuning will often represent the best balance between power, precision, and disruption of the approaches.

Keywords: artificial intelligence, machine learning, AI, ML, generative AI, regulation, pretraining, fine-tuning

Suggested Citation

Ohm, Paul, Focusing on Fine-Tuning: Understanding the Four Pathways for Shaping Generative AI (February 25, 2024). Columbia Science and Technology Law Review, Forthcoming, Available at SSRN: https://ssrn.com/abstract=4738261 or http://dx.doi.org/10.2139/ssrn.4738261

Paul Ohm (Contact Author)

Georgetown University Law Center ( email )

600 New Jersey Avenue, NW
Washington, DC 20001
United States
202-662-9685 (Phone)

Do you have negative results from your research you’d like to share?

Paper statistics

Downloads
35
Abstract Views
91
PlumX Metrics