The University of Southampton
University of Southampton Institutional Repository

Do prompt positions really matter?

Do prompt positions really matter?
Do prompt positions really matter?
Prompt-based models have gathered a lot of attention from researchers due to their remarkable advancements in the fields of zero-shot and few-shot learning. Developing an effective prompt template plays a critical role. However, prior studies have mainly focused on prompt vocabulary searching or embedding initialization within a predefined template with the prompt position fixed. In this empirical study, we conduct the most comprehensive analysis to date of prompt position for diverse Natural Language Processing (NLP) tasks. Our findings quantify the substantial impact prompt position has on model performance. We observe that the prompt positions used in prior studies are often sub-optimal, and this observation is consistent even in widely used instruction-tuned models. These findings suggest prompt position optimisation as a valuable research direction to augment prompt engineering methodologies and prompt position-aware instruction tuning as a potential way to build more robust models in the future.
4102–4130
Association for Computational Linguistics (ACL)
Mao, Junyu
dde288a4-c7c9-423b-8271-e1b5d7a2d675
Middleton, Stuart E.
404b62ba-d77e-476b-9775-32645b04473f
Niranjan, Mahesan
5cbaeea8-7288-4b55-a89c-c43d212ddd4f
Mao, Junyu
dde288a4-c7c9-423b-8271-e1b5d7a2d675
Middleton, Stuart E.
404b62ba-d77e-476b-9775-32645b04473f
Niranjan, Mahesan
5cbaeea8-7288-4b55-a89c-c43d212ddd4f

Mao, Junyu, Middleton, Stuart E. and Niranjan, Mahesan (2024) Do prompt positions really matter? In Findings of the Association for Computational Linguistics: NAACL 2024. Association for Computational Linguistics (ACL). 4102–4130 .

Record type: Conference or Workshop Item (Paper)

Abstract

Prompt-based models have gathered a lot of attention from researchers due to their remarkable advancements in the fields of zero-shot and few-shot learning. Developing an effective prompt template plays a critical role. However, prior studies have mainly focused on prompt vocabulary searching or embedding initialization within a predefined template with the prompt position fixed. In this empirical study, we conduct the most comprehensive analysis to date of prompt position for diverse Natural Language Processing (NLP) tasks. Our findings quantify the substantial impact prompt position has on model performance. We observe that the prompt positions used in prior studies are often sub-optimal, and this observation is consistent even in widely used instruction-tuned models. These findings suggest prompt position optimisation as a valuable research direction to augment prompt engineering methodologies and prompt position-aware instruction tuning as a potential way to build more robust models in the future.

Text
2305.14493v4 - Accepted Manuscript
Available under License Creative Commons Attribution.
Download (435kB)
Text
2024.findings-naacl.258 - Version of Record
Available under License Creative Commons Attribution.
Download (314kB)

More information

Published date: 16 June 2024
Venue - Dates: 2024 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Mexico City, Mexico, 2024-06-16 - 2024-06-24

Identifiers

Local EPrints ID: 492080
URI: http://eprints.soton.ac.uk/id/eprint/492080
PURE UUID: 3ced360f-1a50-4a95-80fc-91871f66f899
ORCID for Stuart E. Middleton: ORCID iD orcid.org/0000-0001-8305-8176
ORCID for Mahesan Niranjan: ORCID iD orcid.org/0000-0001-7021-140X

Catalogue record

Date deposited: 16 Jul 2024 16:39
Last modified: 20 Jul 2024 01:42

Export record

Contributors

Author: Junyu Mao
Author: Mahesan Niranjan ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×