What is it about?

In this paper, we propose the KG-driven verbalizer, a novel approach to bridge the gap between the label word space and the class label space based on their in-between semantic relatedness. Our verbalizer is not handcrafted or trained, but automatically built and embedding-based. Commonsense knowledge is integrated into our approach because the semantic similarity is measured in a hybrid vector space that is an ensemble of word2vec, GloVe and ConceptNet-PPMI. We test the validness of our approach in the task of zero-shot text classification upon a humanitarian aid-related tweet dataset. Satisfactory results are achieved, which proves the effectiveness of the KG-driven verbalizer.

Featured Image

Why is it important?

The combination of pre-training and fine-tuning has become a general paradigm for handling Natural Language Processing (NLP)tasks. This approach needs two preconditions large training datasets and high configuration of the training environment. it is not applicable in scenarios like classification under low data resources. Insufficient labelled data or even unseen classes. Ver-balizer, supports pre-trained Language Models (PLMs) with task-specific prompts can effectively solve the problem. We propose a KG-driven verbalizer that leverages commonsense knowledge to con-struct the embedding-based correspondence between label words and classes. Specifically, we transform the in-between projection into semantic relevance in the commonsense-injected embedding space. For zero-shot text classification, experimental results exhibit the effectiveness of our KG-driven verbalizer on HumAID dataset compared with other baselines.

Perspectives

I believe working on inject knowledge graph into PLMs can increase the efficiency of language model traning process and eventually enhance model's performance in NPL tasks.

Doctor Gangmin Li
University of Bedfordshire

Read the Original

This page is a summary of: Injecting Commonsense Knowledge into Prompt Learning for Zero-Shot Text Classification, February 2023, ACM (Association for Computing Machinery),
DOI: 10.1145/3587716.3587787.
You can read the full text:

Read

Contributors

The following have contributed to this page