Papers
arxiv:2211.15268

Scientific and Creative Analogies in Pretrained Language Models

Published on Nov 28, 2022
Authors:
,
,
,

Abstract

State-of-the-art language models perform poorly on a novel analogy dataset containing diverse and complex mappings across dissimilar domains.

AI-generated summary

This paper examines the encoding of analogy in large-scale pretrained language models, such as BERT and GPT-2. Existing analogy datasets typically focus on a limited set of analogical relations, with a high similarity of the two domains between which the analogy holds. As a more realistic setup, we introduce the Scientific and Creative Analogy dataset (SCAN), a novel analogy dataset containing systematic mappings of multiple attributes and relational structures across dissimilar domains. Using this dataset, we test the analogical reasoning capabilities of several widely-used pretrained language models (LMs). We find that state-of-the-art LMs achieve low performance on these complex analogy tasks, highlighting the challenges still posed by analogy understanding.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2211.15268
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2211.15268 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 1

Collections including this paper 4