1

Logic-LM: Empowering Large Language Models with Symbolic Solvers for Faithful Logical Reasoning

Large Language Models (LLMs) have shown human-like reasoning abilities but still struggle with complex logical problems. This paper introduces a novel framework, Logic-LM, which integrates LLMs with symbolic reasoning to improve logical …

On the Risk of Misinformation Pollution with Large Language Models

In this paper, we comprehensively investigate the potential misuse of modern Large Language Models (LLMs) for generating credible-sounding misinformation and its subsequent impact on information-intensive applications, particularly Open-Domain …

SCITAB: A Challenging Benchmark for Compositional Reasoning and Claim Verification on Scientific Tables

Scientific fact-checking is crucial for ensuring the accuracy, reliability, and trustworthiness of scientific claims. However, existing benchmarks are limited in terms of their claim diversity, reliance on text-based evidence, and oversimplification …

QACHECK: A Demonstration System for Question-Guided Multi-Hop Fact-Checking

Fact-checking real-world claims often requires complex, multi-step reasoning due to the absence of direct evidence to support or refute them. However, existing fact-checking systems often lack transparency in their decision-making, making it …

INSTRUCTSCORE: Explainable Text Generation Evaluation with Finegrained Feedback

Automatically evaluating the quality of language generation is critical. Although recent learned metrics show high correlation with human judgement, these metrics can not explain their verdict or associate the scores with defects in generated text. …

MAF: Multi-Aspect Feedback for Improving Reasoning in Large Language Models

Language Models (LMs) have shown impressive performance in various natural language tasks. However, when it comes to natural language reasoning, LMs still face challenges such as hallucination, generating incorrect intermediate reasoning steps, and …

Doolittle: Benchmarks and Corpora for Academic Writing Formalization

Improving the quality of academic writing is a meaningful but challenging task. Conventional methods of language refinement focus on narrow, specific linguistic features within isolated sentences, such as grammatical errors and improper word use. We …

Dynamic Topic Modeling with Contrastive Topic Evolution and Unassociated Word Sampling

Dynamic topic models capture the evolution of topics in sequential corpora, which have been applied in various tasks like trend analysis. However, existing models suffer from producing repetitive and unassociated topics that fail to reveal the …

Fact-Checking Complex Claims with Program-Guided Reasoning

Fact-checking real-world claims often requires collecting multiple pieces of evidence and applying complex multi-step reasoning. In this paper, we present Program-Guided Fact-Checking (ProgramFC), a novel fact-checking model that decomposes complex …

Attacking Open-domain Question Answering by Injecting Misinformation

With a rise in false, inaccurate, and misleading information in propaganda, news, and social media, real-world Question Answering (QA) systems face the challenges of synthesizing and reasoning over misinformation-polluted contexts to derive correct …