Sökning: onr:"swepub:oai:DiVA.org:uu-497148" >
Probing Pre-trained...
Probing Pre-trained Language Models for Semantic Attributes and their Values
-
- Beloucif, Meriem (författare)
- MIN Faculty, Universität Hamburg,Computational Linguistics
-
- Biemann, Chris (författare)
- MIN Faculty, Universität Hamburg
-
(creator_code:org_t)
- Stroudsburg, PA, USA : Association for Computational Linguistics, 2021
- 2021
- Engelska.
-
Ingår i: Findings of the Association for Computational Linguistics: EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 16-20 November, 2021. - Stroudsburg, PA, USA : Association for Computational Linguistics. - 9781955917100 ; , s. 2554-2559
- Relaterad länk:
-
https://doi.org/10.1...
-
visa fler...
-
https://aclanthology...
-
https://uu.diva-port... (primary) (Raw object)
-
https://aclanthology...
-
https://urn.kb.se/re...
-
https://doi.org/10.1...
-
visa färre...
Abstract
Ämnesord
Stäng
- Pretrained Language Models (PTLMs) yield state-of-the-art performance on many Natural Language Processing tasks, including syntax, semantics and commonsense reasoning. In this paper, we focus on identifying to what extent do PTLMs capture semantic attributes and their values, e.g. the relation between rich and high net worth. We use PTLMs to predict masked tokens using patterns and lists of items from Wikidata in order to verify how likely PTLMs encode semantic attributes along with their values. Such inferences based on semantics are intuitive for us humans as part of our language understanding. Since PTLMs are trained on large amounts of Wikipedia data, we would assume that they can generate similar predictions. However, our findings reveal that PTLMs perform still much worse than humans on this task. We show an analysis which explains how to exploit our methodology to integrate better context and semantics into PTLMs using knowledge bases.
Ämnesord
- NATURVETENSKAP -- Data- och informationsvetenskap -- Språkteknologi (hsv//swe)
- NATURAL SCIENCES -- Computer and Information Sciences -- Language Technology (hsv//eng)
Nyckelord
- BERT
- Pretrained Language Models
- Computational Linguistics
- Machine learning
- Maskininlärning
- Linguistics
- Lingvistik
Publikations- och innehållstyp
- ref (ämneskategori)
- kon (ämneskategori)
Hitta via bibliotek
Till lärosätets databas