Papers
Topics
Authors
Recent
Search
2000 character limit reached

Decoding the Silent Majority: Inducing Belief Augmented Social Graph with Large Language Model for Response Forecasting

Published 20 Oct 2023 in cs.CL, cs.AI, and cs.LG | (2310.13297v1)

Abstract: Automatic response forecasting for news media plays a crucial role in enabling content producers to efficiently predict the impact of news releases and prevent unexpected negative outcomes such as social conflict and moral injury. To effectively forecast responses, it is essential to develop measures that leverage the social dynamics and contextual information surrounding individuals, especially in cases where explicit profiles or historical actions of the users are limited (referred to as lurkers). As shown in a previous study, 97% of all tweets are produced by only the most active 25% of users. However, existing approaches have limited exploration of how to best process and utilize these important features. To address this gap, we propose a novel framework, named SocialSense, that leverages a LLM to induce a belief-centered graph on top of an existent social network, along with graph-based propagation to capture social dynamics. We hypothesize that the induced graph that bridges the gap between distant users who share similar beliefs allows the model to effectively capture the response patterns. Our method surpasses existing state-of-the-art in experimental evaluations for both zero-shot and supervised settings, demonstrating its effectiveness in response forecasting. Moreover, the analysis reveals the framework's capability to effectively handle unseen user and lurker scenarios, further highlighting its robustness and practical applicability.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (56)
  1. Dogu Araci. 2019. Finbert: Financial sentiment analysis with pre-trained language models. arXiv preprint arXiv:1908.10063.
  2. Predicting responses to microblog posts. In proceedings of the 2012 conference of the north American chapter of the Association for Computational Linguistics: human language technologies, pages 602–606.
  3. Twitterbert: Framework for twitter sentiment analysis based on pre-trained language model representations. In Emerging Trends in Intelligent Computing and Informatics: Data Science, Intelligent Information Systems and Smart Computing 4, pages 428–437. Springer.
  4. Annye Braca and Pierpaolo Dondio. 2023. Persuasive communication systems: a machine learning approach to predict the effect of linguistic styles and persuasion techniques. Journal of Systems and Information Technology, (ahead-of-print).
  5. Hou Pong Chan and Irwin King. 2018. Thread popularity prediction and tracking with a permutation-invariant model. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31 - November 4, 2018, pages 3392–3401. Association for Computational Linguistics.
  6. Can LMs Generalize to Future Data? An Empirical Analysis on Text Summarization. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics.
  7. A survey on embedding dynamic graphs. ACM Comput. Surv., 55(2):10:1–10:37.
  8. Active prompting with chain-of-thought for large language models.
  9. Infosurgeon: Cross-media fine-grained information consistency checking for fake news detection. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1683–1698.
  10. Normsage: Multi-lingual multi-cultural norm discovery from conversations on-the-fly. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing.
  11. NewsClaims: A new benchmark for claim detection from news with attribute knowledge. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 6002–6018, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  12. Emotional influence prediction of news posts. In Twelfth International AAAI Conference on Web and Social Media.
  13. Chatgpt outperforms crowd-workers for text-annotation tasks. arXiv preprint arXiv:2303.15056.
  14. Moral foundations theory. Atlas of moral psychology, 211.
  15. Lm-switch: Lightweight language model conditioning in word embedding space. arXiv preprint arXiv:2305.12798.
  16. Deep reinforcement learning with a combinatorial action space for predicting popular reddit threads. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, EMNLP 2016, Austin, Texas, USA, November 1-4, 2016, pages 1838–1848. The Association for Computational Linguistics.
  17. Deberta: Decoding-enhanced bert with disentangled attention. arXiv preprint arXiv:2006.03654.
  18. Tiffany Hsu and Stuart A. Thompson. 2023. Disinformation researchers raise alarms about a.i. chatbots.
  19. Heterogeneous graph transformer. In Proceedings of the web conference 2020, pages 2704–2710.
  20. Zero-shot faithful factual error correction. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 9-14, 2023, pages 5660–5676. Association for Computational Linguistics.
  21. Manitweet: A new benchmark for identifying manipulation of news on social media. CoRR, abs/2305.14225.
  22. Sandeepa Kannangara. 2018. Mining twitter for fine-grained political opinion polarity classification, ideology detection and sarcasm detection. In Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining, pages 751–752.
  23. Large language models are zero-shot reasoners. arXiv preprint arXiv:2205.11916.
  24. Multi-view models for political ideology detection of news articles. arXiv preprint arXiv:1809.03485.
  25. Senti2pop: sentiment-aware topic popularity prediction on social media. In 2019 IEEE International conference on data mining (ICDM), pages 1174–1179. IEEE.
  26. Unsupervised belief representation learning with information-theoretic variational graph auto-encoders. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 1728–1738.
  27. Defining a new nlp playground. ACL Findings.
  28. Kevin Hsin-Yih Lin and Hsin-Hsi Chen. 2008. Ranking reader emotions using pairwise loss minimization and emotional distribution regression. In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, pages 136–144, Honolulu, Hawaii. Association for Computational Linguistics.
  29. Emoticon smoothed language models for twitter sentiment analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 26, pages 1678–1684.
  30. On the variance of the adaptive learning rate and beyond. arXiv preprint arXiv:1908.03265.
  31. Summary of chatgpt/gpt-4 research and perspective towards the future of large language models. arXiv preprint arXiv:2304.01852.
  32. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
  33. Partner personas generation for dialogue response generation. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5200–5212.
  34. A survey on computational propaganda detection. arXiv preprint arXiv:2007.08024.
  35. The behaviors and attitudes of us adults on twitter.
  36. Semeval-2018 task 1: Affect in tweets. In Proceedings of the 12th international workshop on semantic evaluation, pages 1–17.
  37. Propaganda detection in text data based on nlp and machine learning. In MoMLeT+ DS, pages 132–144.
  38. Quinn Owen and Max Zahn. 2023. Avoiding potential ’extinction event’ from ai requires action, us official says.
  39. Pytorch: An imperative style, high-performance deep learning library. arXiv preprint arXiv:1912.01703.
  40. Smartbook: Ai-assisted situation report generation. arXiv preprint arXiv:2303.14337.
  41. Shalom H Schwartz. 1992. Universals in the content and structure of values: Theoretical advances and empirical tests in 20 countries. In Advances in experimental social psychology, volume 25, pages 1–65. Elsevier.
  42. Measuring the effect of influential messages on varying personas. arXiv preprint arXiv:2305.16470.
  43. Incorporating task-specific concept knowledge into script learning. arXiv preprint arXiv:2209.00068.
  44. Attention is all you need. In Advances in neural information processing systems, pages 5998–6008.
  45. Generating diversified comments via reader-aware topic modeling and saliency detection. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 13988–13996.
  46. Sentiment forecasting in dialog. In Proceedings of the 28th International Conference on Computational Linguistics, pages 2448–2458.
  47. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics.
  48. Cross-document misinformation detection based on event graph reasoning. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 543–558, Seattle, United States. Association for Computational Linguistics.
  49. Personalized response generation via generative split memory network. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 1956–1970, Online. Association for Computational Linguistics.
  50. Harnessing the power of llms in practice: A survey on chatgpt and beyond. arXiv preprint arXiv:2304.13712.
  51. Read, attend and comment: A deep architecture for automatic news comment generation.
  52. Shehel Yoosuf and Yin Yang. 2019. Fine-grained propaganda detection with fine-tuned bert. In Proceedings of the second workshop on natural language processing for internet freedom: censorship, disinformation, and propaganda, pages 87–91.
  53. Vibe: Topic-driven temporal adaptation for twitter classification. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP), Singapore. Association for Computational Linguistics.
  54. Calibrate before use: Improving few-shot performance of language models. In International Conference on Machine Learning, pages 12697–12706. PMLR.
  55. Least-to-most prompting enables complex reasoning in large language models. arXiv preprint arXiv:2205.10625.
  56. Graph neural networks: A review of methods and applications. AI open, 1:57–81.
Citations (14)

Summary

  • The paper presents SocialSense, a framework that leverages large language models for latent persona extraction to build belief-augmented social graphs for improved response forecasting.
  • It integrates belief nodes and social interactions using graph neural networks, achieving up to 10.28% improvement in correlation metrics and 5.99% in F1 scores.
  • The approach effectively forecasts responses for various user segments, including lurkers and unseen users, in both supervised and zero-shot settings.

Decoding the Silent Majority: Inducing Belief Augmented Social Graph with LLM for Response Forecasting

Introduction

The paper "Decoding the Silent Majority: Inducing Belief Augmented Social Graph with LLM for Response Forecasting" introduces SocialSense, a framework aimed at enhancing automatic response forecasting in news media scenarios. The primary objective is to predict responses to news messages by leveraging belief-augmented social networks. This approach addresses the limitation of existing methods that often overlook the subtle social dynamics and belief systems that influence how users interact with news content.

Methodology

SocialSense is built upon three key components:

  1. Latent Persona Extraction: Utilizing LLMs, such as ChatGPT, latent personas of users are extracted by analyzing their social media profiles and historical posts. This includes identifying moral and human values, beliefs, and opinions that shape user behavior. The initial stage involves designing a specific prompt Pl\mathrm{P}_l to facilitate the extraction of this latent information.
  2. Belief-Augmented Social Network: The framework develops a belief-centered network by integrating induced and existing social interaction data. The network structure comprises nodes representing news media, users, and beliefs, connected by edges indicating interactions. This structure emphasizes the importance of bridging users with similar beliefs, even if they reside in distant and disconnected social communities.
  3. Graph Information Propagation: Utilizing Graph Neural Networks (GNN), information is propagated through the network to update user representations. A heterogeneous graph transformer mechanism captures interactions among varied node types, facilitating nuanced information flow and enhancing prediction accuracy. Figure 1

    Figure 1: Distribution of belief values.

Task Formulation

The task is defined as a multi-class prediction problem. Given a user persona and a news message, the objective is to predict the user's sentiment polarity ϕp\phi_p and intensity ϕint\phi_{int}. The framework aims to capture how users with similar beliefs respond to news, with evaluations conducted in both zero-shot and supervised settings.

Experimental Validation

SocialSense's effectiveness is validated through experiments on a curated dataset from Twitter, involving over 18,000 users. Evaluations were conducted to gauge the framework's performance in both supervised and zero-shot scenarios.

  • Supervised Scenario: The framework showed superior performance over baseline models like DeBERTa and RoBERTa, achieving higher Spearman and Pearson correlations for sentiment intensity prediction, and better F1 scores for sentiment polarity predictions.
  • Zero-Shot Scenario: By simulating graph propagation with social prompts, SocialSenseZero_\text{Zero} demonstrated enhanced prediction accuracy, underscoring the importance of incorporating social context in forecasts. Figure 2

    Figure 2: Prompt template Pl\mathrm{P}_l used for extracting user latent profiles.

Results and Discussion

The results exhibit that integrating belief nodes within the social network significantly influences the model's performance. The proposed belief-centered approach outperforms competitors by a margin of 10.28% in correlation metrics and 5.99% in F1 scores, illuminating the importance of belief systems in user behavior forecasting.

  • Lurker and Unseen User Scenarios: SocialSense excels in scenarios involving lurkers and unseen users, with improvements observed due to the belief-centered graph's ability to infer accurate user responses without extensive historical data.
  • Ablation Studies: Component-level evaluations highlighted the vital roles of belief nodes, user-news edges, and the methodology of user representation initialization. Figure 3

    Figure 3: Prompt template Ps\mathrm{P}_s used for aggregating neighbor information.

Conclusion

The proposed SocialSense framework provides a robust solution for automatic response forecasting. By bridging socially distant yet belief-aligned users, the framework captures intricate beliefs and social dynamics, thereby enhancing predictive accuracy. Future work may explore the generalization of this approach to different domains and the development of dynamic social prompting strategies for evolving social media data.

In summary, SocialSense represents a significant advancement in the understanding and prediction of user responses in the context of news media, driven by a detailed modeling of belief systems and social interactions.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.