It is all in the [MASK]: Simple instruction-tuning enables BERT-like masked language models as generative classifiers
While encoder-only models such as BERT and ModernBERT are ubiquitous in real-world NLP applications, their conventional reliance on task-specific classification heads can limit their applicability compared to decoder-based large language models (LLMs). In this work, we introduce ModernBERT-Large-Ins...
Saved in:
| Main Authors: | Benjamin Clavié, Nathan Cooper, Benjamin Warner |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Elsevier
2025-06-01
|
| Series: | Natural Language Processing Journal |
| Subjects: | |
| Online Access: | http://www.sciencedirect.com/science/article/pii/S2949719125000263 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
MPTTF-BERT: Multi-Prefix-Tuning Template Fusion by BERT for Zero-shot English Text Relation Extraction Model
by: Fangqi Song
Published: (2025-05-01) -
An approach to automatic answering for English reading comprehension tests
by: Phat Tien Bui, et al.
Published: (2024-09-01) -
Positional embeddings and zero-shot learning using BERT for molecular-property prediction
by: Medard Edmund Mswahili, et al.
Published: (2025-02-01) -
Enhancing QA System Evaluation: An In-Depth Analysis of Metrics and Model-Specific Behaviors
by: Heesop Kim, et al.
Published: (2025-03-01) -
Japanese Short Answer Grading for Japanese Language Learners Using the Contextual Representation of BERT
by: Dyah Lalita Luhurkinanti, et al.
Published: (2025-01-01)