PolyLM: Learning about polysemy through language modeling

dc.contributor.authorAnsell, Alanen_NZ
dc.contributor.authorBravo-Marquez, Felipeen_NZ
dc.contributor.authorPfahringer, Bernharden_NZ
dc.contributor.editorMerlo, Pen_NZ
dc.contributor.editorTiedemann, Jen_NZ
dc.contributor.editorTsarfaty, Ren_NZ
dc.coverage.spatialonlineen_NZ
dc.date.accessioned2023-11-28T21:30:24Z
dc.date.available2023-11-28T21:30:24Z
dc.date.issued2021en_NZ
dc.description.abstractTo avoid the “meaning conflation deficiency” of word embeddings, a number of models have aimed to embed individual word senses. These methods at one time performed well on tasks such as word sense induction (WSI), but they have since been overtaken by task-specific techniques which exploit contextualized embeddings. However, sense embeddings and contextualization need not be mutually exclusive. We introduce PolyLM, a method which formulates the task of learning sense embeddings as a language modeling problem, allowing contextualization techniques to be applied. PolyLM is based on two underlying assumptions about word senses: firstly, that the probability of a word occurring in a given context is equal to the sum of the probabilities of its individual senses occurring; and secondly, that for a given occurrence of a word, one of its senses tends to be much more plausible in the context than the others. We evaluate PolyLM on WSI, showing that it performs considerably better than previous sense embedding techniques, and matches the current state-of-the-art specialized WSI method despite having six times fewer parameters. Code and pre-trained models are available at https://github.com/AlanAnsell/PolyLM.en_NZ
dc.format.mimetypeapplication/pdf
dc.identifier.isbn9781954085022en_NZ
dc.identifier.urihttps://hdl.handle.net/10289/16207
dc.language.isoen
dc.publisherACLen_NZ
dc.relation.isPartOfProc 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume (EACL 2021)en_NZ
dc.relation.urihttps://www.aclweb.org/anthology/2021.eacl-main.45en_NZ
dc.rights© 2021 Association for Computational Linguistics. This work is licensed Creative Commons Attribution 4.0 International License.
dc.sourceEACL 2021en_NZ
dc.titlePolyLM: Learning about polysemy through language modelingen_NZ
dc.typeConference Contribution
dspace.entity.typePublication
pubs.begin-page563
pubs.end-page574
pubs.finish-date2021-04-23en_NZ
pubs.publication-statusPublisheden_NZ
pubs.start-date2021-04-19en_NZ

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
2021.eacl-main.45.pdf
Size:
429.98 KB
Format:
Adobe Portable Document Format
Description:
Published version

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Research Commons Deposit Agreement 2017.pdf
Size:
188.11 KB
Format:
Adobe Portable Document Format
Description: