Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Revealing the impact of synthetic native samples and multi-tasking strategies in Hindi-English code-mixed humour and sarcasm detection

Created by
  • Haebom

Author

Debajyoti Mazumder, Aakash Kumar, Jasabanta Patro

Outline

This paper reports the results of experiments with various strategies to improve code-mixed humor and sarcasm detection. We explored three approaches: (i) native language sample mixing, (ii) multi-task learning (MTL), and (iii) prompting and instruction fine-tuning of a large-scale multilingual language model (VMLM). Native language sample mixing involved adding monolingual task samples to the code-mixed training set, while MTL training involved using native language and code-mixed samples from a semantically related task (hate detection in this study). Finally, we evaluated the effectiveness of VMLM through contextual prompting and instruction fine-tuning, performed over a few trials. Experimental results showed that adding native language samples improved humor and sarcasm detection (up to 6.76% and 8.64% F1-score increases, respectively). Training the MLM within the MTL framework further improved humor and sarcasm detection (up to 10.67% and 12.35% F1-score increases, respectively). In contrast, VMLM's prompting and instruction fine-tuning did not outperform other approaches. Additionally, ablation studies and error analysis were used to identify areas where model improvements were needed, and the code was made public to ensure reproducibility.

Takeaways, Limitations

Takeaways:
We demonstrate that native language sample mixing and multi-task learning (MTL) can significantly improve code-mixed humor and sarcasm detection performance.
We found that multi-task learning (MTL) was more effective than mixing native language samples.
This study presents a practical improvement approach for code-mixed text analysis.
The reproducibility of the study was improved through open code.
Limitations:
Further analysis is needed to determine why VMLM's prompting and direction fine-tuning was not as effective as expected.
There is a lack of specifics regarding areas where model improvement is needed, as revealed through ablation studies and error analysis.
Generalization performance may vary depending on the characteristics of the dataset and model used.
👍