SwePub
Sök i LIBRIS databas

  Extended search

WFRF:(Bender Emily Professor)
 

Search: WFRF:(Bender Emily Professor) > Linguistically Info...

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Linguistically Informed Neural Dependency Parsing for Typologically Diverse Languages

de Lhoneux, Miryam, 1990- (author)
Uppsala universitet,Institutionen för lingvistik och filologi,Computational Linguistics
Nivre, Joakim (thesis advisor)
Uppsala universitet,Institutionen för lingvistik och filologi
Stymne, Sara (thesis advisor)
Uppsala universitet,Institutionen för lingvistik och filologi
show more...
Bender, Emily, Professor (opponent)
University of Washington, Department of Linguistics
show less...
 (creator_code:org_t)
ISBN 9789151307671
Uppsala : Acta Universitatis Upsaliensis, 2019
English 178 s.
Series: Studia Linguistica Upsaliensia, 1652-1366 ; 24
  • Doctoral thesis (other academic/artistic)
Abstract Subject headings
Close  
  • This thesis presents several studies in neural dependency parsing for typologically diverse languages, using treebanks from Universal Dependencies (UD). The focus is on informing models with linguistic knowledge. We first extend a parser to work well on typologically diverse languages, including morphologically complex languages and languages whose treebanks have a high ratio of non-projective sentences, a notorious difficulty in dependency parsing. We propose a general methodology where we sample a representative subset of UD treebanks for parser development and evaluation. Our parser uses recurrent neural networks which construct information sequentially, and we study the incorporation of a recursive neural network layer in our parser. This follows the intuition that language is hierarchical. This layer turns out to be superfluous in our parser and we study its interaction with other parts of the network. We subsequently study transitivity and agreement information learned by our parser for auxiliary verb constructions (AVCs). We suggest that a parser should learn similar information about AVCs as it learns for finite main verbs. This is motivated by work in theoretical dependency grammar. Our parser learns different information about these two if we do not augment it with a recursive layer, but similar information if we do, indicating that there may be benefits from using that layer and we may not yet have found the best way to incorporate it in our parser. We finally investigate polyglot parsing. Training one model for multiple related languages leads to substantial improvements in parsing accuracy over a monolingual baseline. We also study different parameter sharing strategies for related and unrelated languages. Sharing parameters that partially abstract away from word order appears to be beneficial in both cases but sharing parameters that represent words and characters is more beneficial for related than unrelated languages.

Subject headings

HUMANIORA  -- Språk och litteratur -- Jämförande språkvetenskap och allmän lingvistik (hsv//swe)
HUMANITIES  -- Languages and Literature -- General Language Studies and Linguistics (hsv//eng)

Keyword

Dependency parsing
multilingual NLP
Universal Dependencies
Linguistically informed NLP
Computational Linguistics
Datorlingvistik

Publication and Content Type

vet (subject category)
dok (subject category)

Find in a library

To the university's database

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view