Large language model adaptation through few-shots learning and post-hoc calibration
Keywords:
few-shot learning, calibration, large language model adaptationAbstract
In the context of adapting language models to a specific task, using prompt engineering often yields performance gains without requiring access to the internal parameters of the model. Another form of adaptation, much less studied in the literature, is achieved through posthoc calibration techniques, where only the model’s output scores are accessed and modified via a function to enhance task performance. This “gray-box” approach, which only utilizes the values from the model’s output layer, offers a computationally inexpensive alternative to supervised learning techniques and has not been thoroughly investigated in the literature. This work presents some preliminary results in which the combination of prompt engineering and post-hoc calibration shows improvements in multiple-choice social behavior question tasks on two large-scale language models (Phi-1.5 and Phi-2). The results obtained so far suggest that these two techniques are complementary, paving the way for the development of techniques that systematically combine prompt engineering with post-hoc calibration to improve model performance.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Juan Ignacio Tollo

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
Acorde a estos términos, el material se puede compartir (copiar y redistribuir en cualquier medio o formato) y adaptar (remezclar, transformar y crear a partir del material otra obra), siempre que a) se cite la autoría y la fuente original de su publicación (revista y URL de la obra), b) no se use para fines comerciales y c) se mantengan los mismos términos de la licencia.











