Con-S2V: A Generic Framework for Incorporating Extra-Sentential Context into Sen2Vec

If you need an accessible version of this item, please submit a remediation request.
Date
2017
Language
English
Embargo Lift Date
Committee Members
Degree
Degree Year
Department
Grantor
Journal Title
Journal ISSN
Volume Title
Found At
Springer
Abstract

We present a novel approach to learn distributed representation of sentences from unlabeled data by modeling both content and context of a sentence. The content model learns sentence representation by predicting its words. On the other hand, the context model comprises a neighbor prediction component and a regularizer to model distributional and proximity hypotheses, respectively. We propose an online algorithm to train the model components jointly. We evaluate the models in a setup, where contextual information is available. The experimental results on tasks involving classification, clustering, and ranking of sentences show that our model outperforms the best existing models by a wide margin across multiple datasets.

Description
item.page.description.tableofcontents
item.page.relation.haspart
Cite As
Saha, T. K., Joty, S., & Hasan, M. A. (2017). Con-S2V: A Generic Framework for Incorporating Extra-Sentential Context into Sen2Vec. In Machine Learning and Knowledge Discovery in Databases (pp. 753–769). Springer, Cham. https://doi.org/10.1007/978-3-319-71249-9_45
ISSN
Publisher
Series/Report
Sponsorship
Major
Extent
Identifier
Relation
Journal
Machine Learning and Knowledge Discovery in Databases
Source
Author
Alternative Title
Type
Article
Number
Volume
Conference Dates
Conference Host
Conference Location
Conference Name
Conference Panel
Conference Secretariat Location
Version
Author's manuscript
Full Text Available at
This item is under embargo {{howLong}}