[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Consistency of Responses and Continuations Generated by Large Language Models on Social Media

Created by
  • Haebom

Author

Wenlu Fan, Yuqi Zhu, Chenyang Wang, Bin Wang, Wentao Xu

Outline

This study investigated emotional consistency and semantic coherence by analyzing climate change-related conversations on social media (Twitter, Reddit) using large-scale language models (LLMs), Gemma and Llama. We examined how LLMs handled emotional content and maintained semantic relationships in continuation and response tasks, examining emotional transitions, intensity patterns, and semantic similarity between human-authored and LLM-generated content. Gemma tended to amplify negative emotions, especially anger, but preserved positive emotions, such as optimism. Llama better preserved emotions across a wider emotional spectrum. Both models generated responses with attenuated emotional intensity compared to human-authored content and showed a bias toward positive emotions in the response task. Both models maintained strong semantic similarity to the original text, but there was a difference in performance between the continuation and response tasks.

Takeaways, Limitations

Takeaways:
Provides insight into LLM's emotional and semantic processing capabilities
Presentation of __T81466_____ on Social Media Environment and Human-AI Interaction Design
Increased understanding of LLM's emotional patterns (negative emotion amplification, positive emotion bias) and ability to maintain semantic similarity
Presenting differences in Gemma and Llama's ability to process emotions and maintain meaning
Limitations:
The LLM models used in the analysis are limited to two: Gemma and Llama.
There are limitations in generalizing when using only climate change-related conversation data.
Possible lack of clear explanation of how to measure emotional intensity and semantic similarity
Possible lack of detailed description of specific design and evaluation methods for sustained and responsive tasks
👍