结果(博弈论)
萃取(化学)
事件(粒子物理)
计算机科学
自然语言处理
数学
色谱法
化学
物理
数理经济学
量子力学
作者
Jun Gao,Huan Zhao,Wei Wang,Changlong Yu,Ruifeng Xu
出处
期刊:Cornell University - arXiv
日期:2024-02-17
被引量:3
标识
DOI:10.48550/arxiv.2402.11430
摘要
In this study, we present EventRL, a reinforcement learning approach developed to enhance event extraction for large language models (LLMs). EventRL utilizes outcome supervision with specific reward functions to tackle prevalent challenges in LLMs, such as instruction following and hallucination, manifested as the mismatch of event structure and the generation of undefined event types. We evaluate EventRL against existing methods like Few-Shot Prompting (FSP) (based on GPT4) and Supervised Fine-Tuning (SFT) across various LLMs, including GPT-4, LLaMa, and CodeLLaMa models. Our findings show that EventRL significantly outperforms these conventional approaches by improving the performance in identifying and structuring events, particularly in handling novel event types. The study emphasizes the critical role of reward function selection and demonstrates the benefits of incorporating code data for better event extraction. While increasing model size leads to higher accuracy, maintaining the ability to generalize is essential to avoid overfitting.
科研通智能强力驱动
Strongly Powered by AbleSci AI