Papers
Topics
Authors
Recent
Search
2000 character limit reached

Leveraging Logical Rules in Knowledge Editing: A Cherry on the Top

Published 24 May 2024 in cs.CL, cs.AI, and cs.LG | (2405.15452v2)

Abstract: Multi-hop Question Answering (MQA) under knowledge editing (KE) is a key challenge in LLMs. While best-performing solutions in this domain use a plan and solve paradigm to split a question into sub-questions followed by response generation, we claim that this approach is sub-optimal as it fails for hard to decompose questions, and it does not explicitly cater to correlated knowledge updates resulting as a consequence of knowledge edits. This has a detrimental impact on the overall consistency of the updated knowledge. To address these issues, in this paper, we propose a novel framework named RULE-KE, i.e., RULE based Knowledge Editing, which is a cherry on the top for augmenting the performance of all existing MQA methods under KE. Specifically, RULE-KE leverages rule discovery to discover a set of logical rules. Then, it uses these discovered rules to update knowledge about facts highly correlated with the edit. Experimental evaluation using existing and newly curated datasets (i.e., RKE-EVAL) shows that RULE-KE helps augment both performances of parameter-based and memory-based solutions up to 92% and 112.9%, respectively.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. Prompt-saw: Leveraging relation-aware graphs for textual prompt compression. arXiv preprint arXiv:2404.00489, 2024.
  2. Editing factual knowledge in language models. In Conference on Empirical Methods in Natural Language Processing, 2021.
  3. Recall and learn: Fine-tuning deep pretrained language models with less forgetting. In Conference on Empirical Methods in Natural Language Processing, 2020.
  4. Multi-hop question answering under temporal knowledge editing. arXiv preprint arXiv:2404.00492, 2024.
  5. Calibrating noise to sensitivity in private data analysis. In Theory of Cryptography: Third Theory of Cryptography Conference, TCC 2006, New York, NY, USA, March 4-7, 2006. Proceedings 3, pages 265–284. Springer, 2006.
  6. Pokemqa: Programmable knowledge editing for multi-hop question answering. arXiv preprint arXiv:2312.15194, 2023.
  7. Improving sequential model editing with fact retrieval. In Houda Bouamor, Juan Pino, and Kalika Bali, editors, Findings of the Association for Computational Linguistics: EMNLP 2023, pages 11209–11224, Singapore, December 2023. Association for Computational Linguistics.
  8. Do language models have beliefs? methods for detecting, updating, and visualizing model beliefs. ArXiv, abs/2111.13654, 2021.
  9. Faithful question answering with monte-carlo planning. In Annual Meeting of the Association for Computational Linguistics, 2023.
  10. Wilke: Wise-layer knowledge editor for lifelong knowledge editing. ArXiv, abs/2402.10987, 2024.
  11. Differentially private natural language models: Recent advances and future directions. arXiv preprint arXiv:2301.09112, 2023.
  12. Improving faithfulness for vision transformers. arXiv preprint arXiv:2311.17983, 2023.
  13. Seat: stable and explainable attention. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37(11), pages 12907–12915, 2023.
  14. High dimensional differentially private stochastic optimization with heavy-tailed data. In Proceedings of the 41st ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems, pages 227–236, 2022.
  15. Privacy-preserving sparse generalized eigenvalue problem. In International Conference on Artificial Intelligence and Statistics, pages 5052–5062. PMLR, 2023.
  16. Meta-learning online adaptation of language models. ArXiv, abs/2305.15076, 2023.
  17. Jie Huang and Kevin Chen-Chuan Chang. Towards reasoning in large language models: A survey. ArXiv, abs/2212.10403, 2022.
  18. See the unseen: Better context-consistent knowledge-editing by noises. ArXiv, abs/2401.07544, 2024.
  19. Cheonsu Jeong. A study on the implementation of generative ai services using an enterprise data-based llm application architecture. arXiv preprint arXiv:2309.01105, 2023.
  20. Dense passage retrieval for open-domain question answering. ArXiv, abs/2004.04906, 2020.
  21. Decomposed prompting: A modular approach for solving complex tasks. ArXiv, abs/2210.02406, 2022.
  22. End-to-end neural entity linking. arXiv preprint arXiv:1808.07699, 2018.
  23. Faithful vision-language interpretation via concept bottleneck models. In The Twelfth International Conference on Learning Representations, 2023.
  24. Fast and exact rule mining with amie 3. The Semantic Web, 12123:36 – 52, 2020.
  25. A survey on complex knowledge base question answering: Methods, challenges and solutions. ArXiv, abs/2105.11644, 2021.
  26. Pmet: Precise model editing in a transformer. ArXiv, abs/2308.08742, 2023.
  27. A survey on multi-hop question answering and generation. ArXiv, abs/2204.09140, 2022.
  28. Locating and editing factual associations in gpt. Advances in Neural Information Processing Systems, 35:17359–17372, 2022.
  29. Mass-editing memory in a transformer. In The Eleventh International Conference on Learning Representations, 2022.
  30. Memory-based model editing at scale. ArXiv, abs/2206.06520, 2022.
  31. Training language models to follow instructions with human feedback. ArXiv, abs/2203.02155, 2022.
  32. Rnnlogic: Learning logic rules for reasoning on knowledge graphs. ArXiv, abs/2010.04029, 2020.
  33. J. Ross Quinlan. Learning logical definitions from relations. Machine Learning, 5:239–266, 1990.
  34. A survey of hallucination in large foundation models. arXiv preprint arXiv:2309.05922, 2023.
  35. Faster rates of private stochastic convex optimization. In International Conference on Algorithmic Learning Theory, pages 995–1002. PMLR, 2022.
  36. Rule: Neural-symbolic knowledge graph reasoning with rule embedding. ArXiv, abs/2210.14905, 2022.
  37. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
  38. Wikidata: a free collaborative knowledgebase. Communications of the ACM, 57(10):78–85, 2014.
  39. Differentially private (gradient) expectation maximization algorithm with statistical guarantees. arXiv preprint arXiv:2010.13520, 2020.
  40. Estimating smooth glm in non-interactive local differential privacy model with public unlabeled data. In Algorithmic Learning Theory, pages 1207–1213. PMLR, 2021.
  41. Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models. In Annual Meeting of the Association for Computational Linguistics, 2023.
  42. Chain of thought prompting elicits reasoning in large language models. ArXiv, abs/2201.11903, 2022.
  43. Transformers: State-of-the-art natural language processing. In Conference on Empirical Methods in Natural Language Processing, 2019.
  44. How does selection leak privacy: Revisiting private selection and improved results for hyper-parameter tuning. arXiv preprint arXiv:2402.13087, 2024.
  45. Practical differentially private and byzantine-resilient federated learning. Proceedings of the ACM on Management of Data, 1(2):1–26, 2023.
  46. An llm can fool itself: A prompt-based adversarial attack. arXiv preprint arXiv:2310.13345, 2023.
  47. Moral: Moe augmented lora for llms’ lifelong learning. arXiv preprint arXiv:2402.11260, 2024.
  48. Human-ai interactions in the communication era: Autophagy makes large models achieving local optima. arXiv preprint arXiv:2402.11271, 2024.
  49. Dialectical alignment: Resolving the tension of 3h and security threats of llms. arXiv preprint arXiv:2404.00486, 2024.
  50. Can we edit factual knowledge by in-context learning? ArXiv, abs/2305.12740, 2023.
  51. Mquake: Assessing knowledge editing in language models via multi-hop questions. arXiv preprint arXiv:2305.14795, 2023.
  52. Modifying memories in transformer models. ArXiv, abs/2012.00363, 2020.
Citations (5)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.