[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Label-semantics Aware Generative Approach for Domain-Agnostic Multilabel Classification

Created by
  • Haebom

Author

Subhendu Khatuya, Shashwat Naidu, Saptarshi Ghosh, Pawan Goyal, Niloy Ganguly

Outline

In this paper, we propose a domain-independent, robust, and efficient generative model framework, LAGAMC, to address the challenges of manual classification of rapidly growing text data. Instead of treating existing labels as atomic symbols, we train a model to generate these descriptions based on input text using predefined label descriptions. In the inference process, a fine-tuned sentence transformer is used to match the generated descriptions with the predefined labels. We integrate a dual-objective loss function that combines cross-entropy loss and the cosine similarity of the generated sentences and the predefined target descriptions to ensure both semantic alignment and accuracy. LAGAMC is suitable for real-world applications due to its parameter efficiency and versatility on various datasets. Experimental results show that it outperforms the existing state-of-the-art models on all the evaluated datasets, achieving 13.94% performance improvement in Micro-F1 and 24.85% performance improvement in Macro-F1 compared to the closest baseline model.

Takeaways, Limitations

Takeaways:
It outperforms existing state-of-the-art (SOTA) performance on various datasets.
High parameter efficiency makes it suitable for practical applications.
Domain independent, so it can be applied to various fields.
Semantically rich classification possible by leveraging label descriptions.
Limitations:
Performance may be affected by the quality of label descriptions (lack of discussion on the quality of predefined label descriptions).
Additional experiments may be needed to determine the generalization performance of the proposed model. (Experimental results on various datasets are presented, but there is no clear discussion on generalization performance.)
Possible performance degradation for certain types of text data (lack of experimental results for specific text types)
👍