Can Large Language Models Infer and Disagree Like Humans?

05/23/2023
by   noah-lee, et al.
0

Large Language Models (LLMs) have shown stellar achievements in solving a broad range of tasks. When generating text, it is common to sample tokens from these models: whether LLMs closely align with the human disagreement distribution has not been well-studied, especially within the scope of Natural Language Inference (NLI). In this paper, we evaluate the performance and alignment of LLM distribution with humans using two different techniques: Monte Carlo Reconstruction (MCR) and Log Probability Reconstruction (LPR). As a result, we show LLMs exhibit limited ability in solving NLI tasks and simultaneously fail to capture human disagreement distribution, raising concerns about their natural language understanding (NLU) ability and their representativeness of human users.

READ FULL TEXT
research
12/19/2022

Emergent Analogical Reasoning in Large Language Models

The recent advent of large language models - large neural networks train...
research
12/21/2022

Language models are better than humans at next-token prediction

Current language models are considered to have sub-human capabilities at...
research
05/31/2021

Language Model Evaluation Beyond Perplexity

We propose an alternate approach to quantifying how well language models...
research
12/01/2021

Towards More Robust Natural Language Understanding

Natural Language Understanding (NLU) is a branch of Natural Language Pro...
research
11/21/2022

Validating Large Language Models with ReLM

Although large language models (LLMs) have been touted for their ability...
research
09/24/2021

Text-based NP Enrichment

Understanding the relations between entities denoted by NPs in text is a...
research
09/01/2023

No Train Still Gain. Unleash Mathematical Reasoning of Large Language Models with Monte Carlo Tree Search Guided by Energy Function

Large language models (LLMs) exhibit impressive language understanding a...

Please sign up or login with your details

Forgot password? Click here to reset