2019
Report  Open Access

Word-Class Embeddings for Multiclass Text Classification

Moreo Fernandez A. D., Esuli A., Sebastiani F.

Word-Class Embeddings  Word embeddings  Distributional hypothesis  Multiclass text classification 

Pre-trained word embeddings encode general word semantics and lexical regularities of natural language, and have proven useful across many NLP tasks, including word sense disambiguation, machine translation, and sentiment analysis, to name a few. In supervised tasks such as multiclass text classification (the focus of this article) it seems appealing to enhance word representations with ad-hoc embeddings that encode task-specific information. We propose (supervised) word-class embeddings (WCEs), and show that, when concatenated to (unsupervised) pre-trained word embeddings, they substantially facilitate the training of deep-learning models in multiclass classification by topic. We show empirical evidence that WCEs yield a consistent improvement in multiclass classification accuracy, using four popular neural architectures and six widely used and publicly available datasets for multiclass text classification. Our code that implements WCEs is publicly available at this https URL

Source: Research report, ARIADNEplus, 823914, pp.1–29, 2019



Back to previous page
BibTeX entry
@techreport{oai:it.cnr:prodotti:415579,
	title = {Word-Class Embeddings for Multiclass Text Classification},
	author = {Moreo Fernandez A. D. and Esuli A. and Sebastiani F.},
	institution = {Research report, ARIADNEplus, 823914, pp.1–29, 2019},
	year = {2019}
}
CNR ExploRA

Bibliographic record

ISTI Repository

Deposited version Open Access

Also available from

arxiv.orgOpen Access