Chrome Extension
WeChat Mini Program
Use on ChatGLM

Beyond Graphs: Can Large Language Models Comprehend Hypergraphs?

Yifan Feng, Chengwu Yang, Xingliang Hou,Shaoyi Du,Shihui Ying,Zongze Wu,Yue Gao

Computing Research Repository (CoRR)(2025)

Tsinghua University | Xi&x27an Jiaotong University | Shanghai University | Shenzhen University

Cited 0|Views6
Abstract
Existing benchmarks like NLGraph and GraphQA evaluate LLMs on graphs by focusing mainly on pairwise relationships, overlooking the high-order correlations found in real-world data. Hypergraphs, which can model complex beyond-pairwise relationships, offer a more robust framework but are still underexplored in the context of LLMs. To address this gap, we introduce LLM4Hypergraph, the first comprehensive benchmark comprising 21,500 problems across eight low-order, five high-order, and two isomorphism tasks, utilizing both synthetic and real-world hypergraphs from citation networks and protein structures. We evaluate six prominent LLMs, including GPT-4o, demonstrating our benchmark’s effectiveness in identifying model strengths and weaknesses. Our specialized prompt- ing framework incorporates seven hypergraph languages and introduces two novel techniques, Hyper-BAG and Hyper-COT, which enhance high-order reasoning and achieve an average 4% (up to 9%) performance improvement on structure classification tasks. This work establishes a foundational testbed for integrating hypergraph computational capabilities into LLMs, advancing their comprehension.
More
Translated text
PDF
Bibtex
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn
Chat Paper

要点】:论文提出了LLM4Hypergraph,首个针对大型语言模型在超图理解能力上的全面基准,发现了现有模型在高阶关系理解上的不足,并通过两种新颖技术提高了模型性能。

方法】:作者开发了一种特殊的提示框架,包含七种超图语言,并引入了两种新技术Hyper-BAG和Hyper-COT,以增强高阶推理能力。

实验】:研究通过21,500个问题,涵盖八种低阶、五种高阶以及两种同构任务,使用了合成和现实世界中的引用网络及蛋白质结构超图,评估了六个著名的大型语言模型,包括GPT-4o,实验结果显示了提出的框架在结构分类任务上平均提高了4个点。