Probing for Hyperbole in Pre-Trained Language Models

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Standard

Probing for Hyperbole in Pre-Trained Language Models. / Schneidermann, Nina Skovgaard; Hershcovich, Daniel; Pedersen, Bolette Sandford.

Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop),. Association for Computational Linguistics, 2023. s. 200–211.

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Harvard

Schneidermann, NS, Hershcovich, D & Pedersen, BS 2023, Probing for Hyperbole in Pre-Trained Language Models. i Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop),. Association for Computational Linguistics, s. 200–211, 61st Annual Meeting of the Association for Computational Linguistics, ACL 2023, Toronto, Canada, 09/07/2023. https://doi.org/10.18653/v1/2023.acl-srw.30

APA

Schneidermann, N. S., Hershcovich, D., & Pedersen, B. S. (2023). Probing for Hyperbole in Pre-Trained Language Models. I Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop), (s. 200–211). Association for Computational Linguistics. https://doi.org/10.18653/v1/2023.acl-srw.30

Vancouver

Schneidermann NS, Hershcovich D, Pedersen BS. Probing for Hyperbole in Pre-Trained Language Models. I Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop),. Association for Computational Linguistics. 2023. s. 200–211 https://doi.org/10.18653/v1/2023.acl-srw.30

Author

Schneidermann, Nina Skovgaard ; Hershcovich, Daniel ; Pedersen, Bolette Sandford. / Probing for Hyperbole in Pre-Trained Language Models. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop),. Association for Computational Linguistics, 2023. s. 200–211

Bibtex

@inproceedings{197e70ea5d924bc595148ffcbbb03f25,
title = "Probing for Hyperbole in Pre-Trained Language Models",
abstract = "Hyperbole is a common figure of speech, which is under-explored in NLP research. In this study, we conduct edge and minimal description length (MDL) probing experiments on three pre-trained language models (PLMs) in an attempt to explore the extent to which hyperbolic information is encoded in these models. We use both word-in-context and sentence-level representations as model inputs as a basis for comparison. We also annotate 63 hyperbole sentences from the HYPO dataset according to an operational taxonomy to conduct an error analysis to explore the encoding of different hyperbole categories. Our results show that hyperbole is to a limited extent encoded in PLMs, and mostly in the final layers. They also indicate that hyperbolic information may be better encoded by the sentence-level representations, which, due to the pragmatic nature of hyperbole, may therefore provide a more accurate and informative representation in PLMs. Finally, the inter-annotator agreement for our annotations, a Cohen{\textquoteright}s Kappa of 0.339, suggest that the taxonomy categories may not be intuitive and need revision or simplification.",
author = "Schneidermann, {Nina Skovgaard} and Daniel Hershcovich and Pedersen, {Bolette Sandford}",
year = "2023",
doi = "10.18653/v1/2023.acl-srw.30",
language = "English",
pages = "200–211",
booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop),",
publisher = "Association for Computational Linguistics",
note = "61st Annual Meeting of the Association for Computational Linguistics, ACL 2023 ; Conference date: 09-07-2023 Through 14-07-2023",

}

RIS

TY - GEN

T1 - Probing for Hyperbole in Pre-Trained Language Models

AU - Schneidermann, Nina Skovgaard

AU - Hershcovich, Daniel

AU - Pedersen, Bolette Sandford

PY - 2023

Y1 - 2023

N2 - Hyperbole is a common figure of speech, which is under-explored in NLP research. In this study, we conduct edge and minimal description length (MDL) probing experiments on three pre-trained language models (PLMs) in an attempt to explore the extent to which hyperbolic information is encoded in these models. We use both word-in-context and sentence-level representations as model inputs as a basis for comparison. We also annotate 63 hyperbole sentences from the HYPO dataset according to an operational taxonomy to conduct an error analysis to explore the encoding of different hyperbole categories. Our results show that hyperbole is to a limited extent encoded in PLMs, and mostly in the final layers. They also indicate that hyperbolic information may be better encoded by the sentence-level representations, which, due to the pragmatic nature of hyperbole, may therefore provide a more accurate and informative representation in PLMs. Finally, the inter-annotator agreement for our annotations, a Cohen’s Kappa of 0.339, suggest that the taxonomy categories may not be intuitive and need revision or simplification.

AB - Hyperbole is a common figure of speech, which is under-explored in NLP research. In this study, we conduct edge and minimal description length (MDL) probing experiments on three pre-trained language models (PLMs) in an attempt to explore the extent to which hyperbolic information is encoded in these models. We use both word-in-context and sentence-level representations as model inputs as a basis for comparison. We also annotate 63 hyperbole sentences from the HYPO dataset according to an operational taxonomy to conduct an error analysis to explore the encoding of different hyperbole categories. Our results show that hyperbole is to a limited extent encoded in PLMs, and mostly in the final layers. They also indicate that hyperbolic information may be better encoded by the sentence-level representations, which, due to the pragmatic nature of hyperbole, may therefore provide a more accurate and informative representation in PLMs. Finally, the inter-annotator agreement for our annotations, a Cohen’s Kappa of 0.339, suggest that the taxonomy categories may not be intuitive and need revision or simplification.

U2 - 10.18653/v1/2023.acl-srw.30

DO - 10.18653/v1/2023.acl-srw.30

M3 - Article in proceedings

SP - 200

EP - 211

BT - Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop),

PB - Association for Computational Linguistics

T2 - 61st Annual Meeting of the Association for Computational Linguistics, ACL 2023

Y2 - 9 July 2023 through 14 July 2023

ER -

ID: 367292720