Fine-Tuning Models: Uncovering Hidden Capabilities
Fine-tuning large pre-trained models rarely alters underlying capabilities, instead adding a "wrapper" to perform new tasks without changing core knowledge.
This is a Plain English Papers summary of a research paper called Fine-tuning veils rather than tailors underlying model powers. If you like these kinds of analysis, you should join AImodels.fyi or follow me on Twitter. Overview Fine-tuning large pre-trained models has become the standard approach for developing machine learning systems, including those intended for safe deployment. However, there has been little research exploring how fine-tuning affects the underlying capabilities learned by a model during pre-training. This paper aims to address this gap by analyzing fine-tuning i...