Automatically extracting polarity-bearing topics for cross-domain sentiment classification

Yulan He, Harith Alani, Chenghua Lin

Research output: Chapter in Book/Report/Conference proceedingOther chapter contribution

Abstract

Joint sentiment-topic (JST) model was previously proposed to detect sentiment and topic simultaneously from text. The only supervision required by JST model learning is domain-independent polarity word priors. In this paper, we modify the JST model by incorporating word polarity priors through modifying the topic-word Dirichlet priors. We study the polarity-bearing topics extracted by JST and show that by augmenting the original feature space with polarity-bearing topics, the in-domain supervised classifiers learned from augmented feature representation achieve the state-of-the-art performance of 95% on the movie review data and an average of 90% on the multi-domain sentiment dataset. Furthermore, using feature augmentation and selection according to the information gain criteria for cross-domain sentiment classification, our proposed approach performs either better or comparably compared to previous approaches. Nevertheless, our approach is much simpler and does not require difficult parameter tuning.
Original languageEnglish
Title of host publicationProceeding : HLT '11 proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies
EditorsDekang Lin
Place of PublicationStroudsburg, PA (US)
PublisherAssociation for Computational Linguistics
Pages123-131
Number of pages9
Volume1
ISBN (Print)978-1-932432-87-9
Publication statusPublished - 2011
Event49th annual meeting of the association for computational linguistics, HLT '11 - Portland, OR, United States
Duration: 19 Jun 201124 Jun 2011

Conference

Conference49th annual meeting of the association for computational linguistics, HLT '11
CountryUnited States
CityPortland, OR
Period19/06/1124/06/11

Fingerprint

Bearings (structural)
Classifiers
Tuning

Cite this

He, Y., Alani, H., & Lin, C. (2011). Automatically extracting polarity-bearing topics for cross-domain sentiment classification. In D. Lin (Ed.), Proceeding : HLT '11 proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies (Vol. 1, pp. 123-131). Stroudsburg, PA (US): Association for Computational Linguistics.
He, Yulan ; Alani, Harith ; Lin, Chenghua. / Automatically extracting polarity-bearing topics for cross-domain sentiment classification. Proceeding : HLT '11 proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies. editor / Dekang Lin. Vol. 1 Stroudsburg, PA (US) : Association for Computational Linguistics, 2011. pp. 123-131
@inbook{93fa14cee3be4494bc37974f515af9f9,
title = "Automatically extracting polarity-bearing topics for cross-domain sentiment classification",
abstract = "Joint sentiment-topic (JST) model was previously proposed to detect sentiment and topic simultaneously from text. The only supervision required by JST model learning is domain-independent polarity word priors. In this paper, we modify the JST model by incorporating word polarity priors through modifying the topic-word Dirichlet priors. We study the polarity-bearing topics extracted by JST and show that by augmenting the original feature space with polarity-bearing topics, the in-domain supervised classifiers learned from augmented feature representation achieve the state-of-the-art performance of 95{\%} on the movie review data and an average of 90{\%} on the multi-domain sentiment dataset. Furthermore, using feature augmentation and selection according to the information gain criteria for cross-domain sentiment classification, our proposed approach performs either better or comparably compared to previous approaches. Nevertheless, our approach is much simpler and does not require difficult parameter tuning.",
author = "Yulan He and Harith Alani and Chenghua Lin",
year = "2011",
language = "English",
isbn = "978-1-932432-87-9",
volume = "1",
pages = "123--131",
editor = "Dekang Lin",
booktitle = "Proceeding : HLT '11 proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies",
publisher = "Association for Computational Linguistics",

}

He, Y, Alani, H & Lin, C 2011, Automatically extracting polarity-bearing topics for cross-domain sentiment classification. in D Lin (ed.), Proceeding : HLT '11 proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies. vol. 1, Association for Computational Linguistics, Stroudsburg, PA (US), pp. 123-131, 49th annual meeting of the association for computational linguistics, HLT '11, Portland, OR, United States, 19/06/11.

Automatically extracting polarity-bearing topics for cross-domain sentiment classification. / He, Yulan; Alani, Harith; Lin, Chenghua.

Proceeding : HLT '11 proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies. ed. / Dekang Lin. Vol. 1 Stroudsburg, PA (US) : Association for Computational Linguistics, 2011. p. 123-131.

Research output: Chapter in Book/Report/Conference proceedingOther chapter contribution

TY - CHAP

T1 - Automatically extracting polarity-bearing topics for cross-domain sentiment classification

AU - He, Yulan

AU - Alani, Harith

AU - Lin, Chenghua

PY - 2011

Y1 - 2011

N2 - Joint sentiment-topic (JST) model was previously proposed to detect sentiment and topic simultaneously from text. The only supervision required by JST model learning is domain-independent polarity word priors. In this paper, we modify the JST model by incorporating word polarity priors through modifying the topic-word Dirichlet priors. We study the polarity-bearing topics extracted by JST and show that by augmenting the original feature space with polarity-bearing topics, the in-domain supervised classifiers learned from augmented feature representation achieve the state-of-the-art performance of 95% on the movie review data and an average of 90% on the multi-domain sentiment dataset. Furthermore, using feature augmentation and selection according to the information gain criteria for cross-domain sentiment classification, our proposed approach performs either better or comparably compared to previous approaches. Nevertheless, our approach is much simpler and does not require difficult parameter tuning.

AB - Joint sentiment-topic (JST) model was previously proposed to detect sentiment and topic simultaneously from text. The only supervision required by JST model learning is domain-independent polarity word priors. In this paper, we modify the JST model by incorporating word polarity priors through modifying the topic-word Dirichlet priors. We study the polarity-bearing topics extracted by JST and show that by augmenting the original feature space with polarity-bearing topics, the in-domain supervised classifiers learned from augmented feature representation achieve the state-of-the-art performance of 95% on the movie review data and an average of 90% on the multi-domain sentiment dataset. Furthermore, using feature augmentation and selection according to the information gain criteria for cross-domain sentiment classification, our proposed approach performs either better or comparably compared to previous approaches. Nevertheless, our approach is much simpler and does not require difficult parameter tuning.

UR - http://www.scopus.com/inward/record.url?scp=84255206668&partnerID=8YFLogxK

UR - http://dl.acm.org/citation.cfm?id=2002472.2002489&coll=DL&dl=ACM&CFID=173576560&CFTOKEN=50339695

M3 - Other chapter contribution

AN - SCOPUS:84255206668

SN - 978-1-932432-87-9

VL - 1

SP - 123

EP - 131

BT - Proceeding : HLT '11 proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies

A2 - Lin, Dekang

PB - Association for Computational Linguistics

CY - Stroudsburg, PA (US)

ER -

He Y, Alani H, Lin C. Automatically extracting polarity-bearing topics for cross-domain sentiment classification. In Lin D, editor, Proceeding : HLT '11 proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies. Vol. 1. Stroudsburg, PA (US): Association for Computational Linguistics. 2011. p. 123-131