Yet another approximation of human semantic judgments using LLMs... but with quantized local models on novel data

This study investigates the automatic generation of semantic norms on word specificity using various quantized open-source local Large Language Models (LLMs), including a comparison with a proprietary model (i.e. GPT-4). Word specificity norms on English are still not public, thus they are not inclu...

Full description

Saved in:
Bibliographic Details
Main Authors: Andrea Amelio Ravelli, Marianna Marcella Bolognesi
Format: Article
Language:English
Published: Accademia University Press 2024-12-01
Series:IJCoL
Online Access:https://journals.openedition.org/ijcol/1439
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This study investigates the automatic generation of semantic norms on word specificity using various quantized open-source local Large Language Models (LLMs), including a comparison with a proprietary model (i.e. GPT-4). Word specificity norms on English are still not public, thus they are not included in the training datasets of all tested models. This offers a novel contribution by assessing LLMs ability to generalize beyond pre-trained knowledge. Our findings reveal that smaller, local quantized models such as Llama3, Phi3, and Mistral underperform in generating human-like judgments of word specificity, while a larger model such as Mixtral, even if slightly less accurate than GPT-4, represents a viable alternative to proprietary models if adequate computational resources are accessible. These findings open up new perspectives for research on linguistic features and on the scalablility of semantic norms without relying on proprietary models.
ISSN:2499-4553