WorkDifferentWithAI.com Academic Paper Alert!
Written by Wanqin Ma, Chenyang Yang, Christian Kästner
Category: AI for IT
Article Section: AI Development and Operations; Automated Code Refactoring
Publication Date: 2023-11-18
SEO Description: Exploring updated regression testing methods for improving prompt design in evolving Large Language Model (LLM) APIs.
Keywords
Large Language Models, Regression Testing, API Evolution, Prompt Design, Non-Determinism
AI-Generated Paper Summary
Generated by Ethical AI Researcher GPT
Author’s Abstract
Large Language Models (LLMs) are increasingly integrated into software applications. Downstream application developers often access LLMs through APIs provided as a service. However, LLM APIs are often updated silently and scheduled to be deprecated, forcing users to continuously adapt to evolving models. This can cause performance regression and affect prompt design choices, as evidenced by our case study on toxicity detection. Based on our case study, we emphasize the need for and re-examine the concept of regression testing for evolving LLM APIs. We argue that regression testing LLMs requires fundamental changes to traditional testing approaches, due to different correctness notions, prompting brittleness, and non-determinism in LLM APIs.