RAGAR, Your Falsehood RADAR: RAG-Augmented Reasoning for Political Fact-Checking using Multimodal Large Language Models

research-papers

RAGAR, Your Falsehood RADAR: RAG-Augmented Reasoning for Political Fact-Checking using Multimodal Large Language Models

Original Paper: https://arxiv.org/abs/2404.12065 By: M. Abdul Khaliq, P. Chang, M. Ma, B. Pflugfelder, F. Miletić Abstract: The escalating challenge of misinformation, particularly in the context of political discourse, necessitates advanced solutions for fact-checking. We introduce innovative approaches to enhance the reliability and efficiency of multimodal

By Athina AI
CYBERSECEVAL 2: A Wide-Ranging Cybersecurity Evaluation Suite for Large Language Models

research-papers

CYBERSECEVAL 2: A Wide-Ranging Cybersecurity Evaluation Suite for Large Language Models

Original Paper: https://ai.meta.com/research/publications/cyberseceval-2-a-wide-ranging-cybersecurity-evaluation-suite-for-large-language-models/ By: Manish Bhatt∗, Sahana Chennabasappa∗, Yue Li∗, Cyrus Nikolaidis∗, Daniel Song∗, Shengye Wan∗, Faizan Ahmad, Cornelius Aschermann, Yaohui Chen, Dhaval Kapil, David Molnar, Spencer Whitman, Joshua Saxe∗ ∗Co-equal primary author Abstract: Large language models (LLMs) introduce new security risks, but there

By Athina AI
Ever: Mitigating Hallucination in Large Language Models through Real-Time Verification and Rectification

research-papers

Ever: Mitigating Hallucination in Large Language Models through Real-Time Verification and Rectification

Original Paper: https://arxiv.org/html/2311.09114v2 By: Haoqiang Kang, Juntong Ni, Huaxiu Yao Abstract: Large Language Models (LLMs) have demonstrated remarkable proficiency in generating fluent text. However, they often encounter the challenge of generating inaccurate or hallucinated content. This issue is common in both non-retrieval-based generation and retrieval-augmented

By Athina AI