File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Domain Adapted Word Embeddings for Improved Sentiment Classification

TitleDomain Adapted Word Embeddings for Improved Sentiment Classification
Authors
Issue Date2018
Citation
Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2018, p. 51-59 How to Cite?
AbstractGeneric word embeddings are trained on large-scale generic corpora; Domain Specific (DS) word embeddings are trained only on data from a domain of interest. This paper proposes a method to combine the breadth of generic embeddings with the specificity of domain specific embeddings. The resulting embeddings, called Domain Adapted (DA) word embeddings, are formed by first aligning corresponding word vectors using Canonical Correlation Analysis (CCA) or the related nonlinear Kernel CCA (KCCA) and then combining them via convex optimization. Results from evaluation on sentiment classification tasks show that the DA embeddings substantially outperform both generic, DS embeddings when used as input features to standard or state-of-the-art sentence encoding algorithms for classification.
Persistent Identifierhttp://hdl.handle.net/10722/341283
ISSN

 

DC FieldValueLanguage
dc.contributor.authorSarma, Prathusha K.-
dc.contributor.authorLiang, Yingyu-
dc.contributor.authorSethares, William A.-
dc.date.accessioned2024-03-13T08:41:36Z-
dc.date.available2024-03-13T08:41:36Z-
dc.date.issued2018-
dc.identifier.citationProceedings of the Annual Meeting of the Association for Computational Linguistics, 2018, p. 51-59-
dc.identifier.issn0736-587X-
dc.identifier.urihttp://hdl.handle.net/10722/341283-
dc.description.abstractGeneric word embeddings are trained on large-scale generic corpora; Domain Specific (DS) word embeddings are trained only on data from a domain of interest. This paper proposes a method to combine the breadth of generic embeddings with the specificity of domain specific embeddings. The resulting embeddings, called Domain Adapted (DA) word embeddings, are formed by first aligning corresponding word vectors using Canonical Correlation Analysis (CCA) or the related nonlinear Kernel CCA (KCCA) and then combining them via convex optimization. Results from evaluation on sentiment classification tasks show that the DA embeddings substantially outperform both generic, DS embeddings when used as input features to standard or state-of-the-art sentence encoding algorithms for classification.-
dc.languageeng-
dc.relation.ispartofProceedings of the Annual Meeting of the Association for Computational Linguistics-
dc.titleDomain Adapted Word Embeddings for Improved Sentiment Classification-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85089289412-
dc.identifier.spage51-
dc.identifier.epage59-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats