Publication result detail

Task Prompt Vectors: Effective Initialization through Multi-Task Soft Prompt Transfer

BELANEC, R.; OSTERMANN, S.; SRBA, I.; BIELIKOVÁ, M.

Original Title

Task Prompt Vectors: Effective Initialization through Multi-Task Soft Prompt Transfer

English Title

Task Prompt Vectors: Effective Initialization through Multi-Task Soft Prompt Transfer

Type

Paper in proceedings outside WoS and Scopus

Original Abstract

Prompt tuning is a parameter-efficient method for adapting large language models (LLMs), where only a small continuous soft prompt is finetuned. In recent works, soft prompts have usually been trained in a task-specific way, leaving their multi-task capabilities underexplored. Our work aims to make soft prompts more task modular based on recent research on task vectors, where arithmetic operations are applied on full model weights to achieve the desired multi-task performance. To this end, we introduce Task Prompt Vectors, created by the element-wise difference between weights of tuned soft prompts and their random initialization. Experimental results on an extensive set of 19 datasets show that task prompt vectors can be used in low-resource settings to initialize prompt tuning on similar tasks effectively. In addition, we show that task prompt vectors are independent of the random initialization of prompt tuning on 3 different language model architectures. This key property of random initialization independence allows prompt arithmetics with the pre-trained vectors from different tasks. In this way, the arithmetic addition of task prompt vectors from multiple tasks represents a competitive and computationally more effective alternative to state-of-the-art solutions.

English abstract

Prompt tuning is a parameter-efficient method for adapting large language models (LLMs), where only a small continuous soft prompt is finetuned. In recent works, soft prompts have usually been trained in a task-specific way, leaving their multi-task capabilities underexplored. Our work aims to make soft prompts more task modular based on recent research on task vectors, where arithmetic operations are applied on full model weights to achieve the desired multi-task performance. To this end, we introduce Task Prompt Vectors, created by the element-wise difference between weights of tuned soft prompts and their random initialization. Experimental results on an extensive set of 19 datasets show that task prompt vectors can be used in low-resource settings to initialize prompt tuning on similar tasks effectively. In addition, we show that task prompt vectors are independent of the random initialization of prompt tuning on 3 different language model architectures. This key property of random initialization independence allows prompt arithmetics with the pre-trained vectors from different tasks. In this way, the arithmetic addition of task prompt vectors from multiple tasks represents a competitive and computationally more effective alternative to state-of-the-art solutions.

Authors

BELANEC, R.; OSTERMANN, S.; SRBA, I.; BIELIKOVÁ, M.

Released

04.10.2025

Publisher

Springer, Berlin, Heidelberg

ISBN

978-3-662-72242-8

Pages from

77

Pages to

94

Pages count

561

URL

BibTex

@inproceedings{BUT198002,
  author="BELANEC, R. and OSTERMANN, S. and SRBA, I. and BIELIKOVÁ, M.",
  title="Task Prompt Vectors: Effective Initialization through Multi-Task Soft Prompt Transfer",
  year="2025",
  pages="77--94",
  publisher="Springer, Berlin, Heidelberg",
  doi="10.1007/978-3-662-72243-5\{_}5",
  isbn="978-3-662-72242-8",
  url="https://link.springer.com/chapter/10.1007/978-3-662-72243-5_5"
}