Efficient Structured Prediction with Transformer Encoders

Finetuning is a useful method for adapting Transformer-based text encoders to new tasks but can be computationally expensive for structured prediction tasks that require tuning at the token level. Furthermore, finetuning is inherently inefficient in updating all base model parameters, which prevent...

Full description

Saved in:
Bibliographic Details
Main Author: Ali Basirat
Format: Article
Language:English
Published: Linköping University Electronic Press 2024-12-01
Series:Northern European Journal of Language Technology
Online Access:https://nejlt.ep.liu.se/article/view/4932
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832591235099918336
author Ali Basirat
author_facet Ali Basirat
author_sort Ali Basirat
collection DOAJ
description Finetuning is a useful method for adapting Transformer-based text encoders to new tasks but can be computationally expensive for structured prediction tasks that require tuning at the token level. Furthermore, finetuning is inherently inefficient in updating all base model parameters, which prevents parameter sharing across tasks. To address these issues, we propose a method for efficient task adaptation of frozen Transformer encoders based on the local contribution of their intermediate layers to token representations. Our adapter uses a novel attention mechanism to aggregate intermediate layers and tailor the resulting representations to a target task. Experiments on several structured prediction tasks demonstrate that our method outperforms previous approaches, retaining over 99% of the finetuning performance at a fraction of the training cost. Our proposed method offers an efficient solution for adapting frozen Transformer encoders to new tasks, improving performance and enabling parameter sharing across different tasks.
format Article
id doaj-art-625983470b724f5b9c9647453e516535
institution Kabale University
issn 2000-1533
language English
publishDate 2024-12-01
publisher Linköping University Electronic Press
record_format Article
series Northern European Journal of Language Technology
spelling doaj-art-625983470b724f5b9c9647453e5165352025-01-22T15:24:15ZengLinköping University Electronic PressNorthern European Journal of Language Technology2000-15332024-12-0110110.3384/nejlt.2000-1533.2024.4932Efficient Structured Prediction with Transformer EncodersAli Basirat0University of Copenhagen Finetuning is a useful method for adapting Transformer-based text encoders to new tasks but can be computationally expensive for structured prediction tasks that require tuning at the token level. Furthermore, finetuning is inherently inefficient in updating all base model parameters, which prevents parameter sharing across tasks. To address these issues, we propose a method for efficient task adaptation of frozen Transformer encoders based on the local contribution of their intermediate layers to token representations. Our adapter uses a novel attention mechanism to aggregate intermediate layers and tailor the resulting representations to a target task. Experiments on several structured prediction tasks demonstrate that our method outperforms previous approaches, retaining over 99% of the finetuning performance at a fraction of the training cost. Our proposed method offers an efficient solution for adapting frozen Transformer encoders to new tasks, improving performance and enabling parameter sharing across different tasks. https://nejlt.ep.liu.se/article/view/4932
spellingShingle Ali Basirat
Efficient Structured Prediction with Transformer Encoders
Northern European Journal of Language Technology
title Efficient Structured Prediction with Transformer Encoders
title_full Efficient Structured Prediction with Transformer Encoders
title_fullStr Efficient Structured Prediction with Transformer Encoders
title_full_unstemmed Efficient Structured Prediction with Transformer Encoders
title_short Efficient Structured Prediction with Transformer Encoders
title_sort efficient structured prediction with transformer encoders
url https://nejlt.ep.liu.se/article/view/4932
work_keys_str_mv AT alibasirat efficientstructuredpredictionwithtransformerencoders