Lessons from ChainForge, a Visual Tool for Prompt Engineering and LLM Hypothesis Testing

Abstract

I showcase ChainForge (chainforge.ai), an open-source tool for open-ended testing of hypotheses about large language model (LLM) outputs. First, I cover the motivations behind ChainForge and demo the tool. Then, I will detail results from in-lab and in-the-wild usage studies conducted with colleagues at Harvard CS. We find that there are three stages of prompt engineering —opportunistic exploration, limited evaluation, and iterative refinement —and argue that future designers need to explicitly consider each stage when developing LLM sensemaking tools. Finally, I cover more informal lessons learned, such as surprising ways people have been using ChainForge (at least to us!), trade-offs between a tool’s breadth of applicability and the learning curve for specific user groups; and future research directions.

Date
Apr 26, 2024 11:00 AM — 11:30 AM
Location
Polytechnique Montreal
2500 Chem. de Polytechnique, Montréal, QC H3T 1J4
Ian Arawjo
Ian Arawjo
Assistant Professor - Université de Montréal

Ian is an assistant Professor of Human-Computer Interaction at the University of Montréal in the Department of Computer Science and Operations Research (DIRO), where I am also affiliated with Mila. I lead the Montréal HCI group. In the recent past, I was a Postdoctoral Fellow at Harvard University, working with Professor Elena Glassman in the Harvard HCI group.