no code implementations • 30 May 2024 • Avelina Asada Hadji-Kyriacou, Ognjen Arandjelovic
Pre-trained Language Models (LMs) exhibit strong zero-shot and in-context learning capabilities; however, their behaviors are often difficult to control.
no code implementations • 14 Dec 2023 • Avelina Asada Hadji-Kyriacou, Ognjen Arandjelovic
This paper introduces a novel Parameter-Efficient Fine-Tuning (PEFT) framework for multi-modal, multi-task transfer learning with pre-trained language models.