ID: cs/0001002

Minimum Description Length and Compositionality

January 4, 2000

View on ArXiv
Wlodek Zadrozny
Computer Science
Computation and Language
Artificial Intelligence

We present a non-vacuous definition of compositionality. It is based on the idea of combining the minimum description length principle with the original definition of compositionality (that is, that the meaning of the whole is a function of the meaning of the parts). The new definition is intuitive and allows us to distinguish between compositional and non-compositional semantics, and between idiomatic and non-idiomatic expressions. It is not ad hoc, since it does not make any references to non-intrinsic properties of meaning functions (like being a polynomial). Moreover, it allows us to compare different meaning functions with respect to how compositional they are. It bridges linguistic and corpus-based, statistical approaches to natural language understanding.

Similar papers 1

Are Representations Built from the Ground Up? An Empirical Examination of Local Composition in Language Models

October 7, 2022

90% Match
Emmy Liu, Graham Neubig
Computation and Language

Compositionality, the phenomenon where the meaning of a phrase can be derived from its constituent parts, is a hallmark of human language. At the same time, many phrases are non-compositional, carrying a meaning beyond that of each part in isolation. Representing both of these types of phrases is critical for language understanding, but it is an open question whether modern language models (LMs) learn to do so; in this work we examine this question. We first formulate a probl...

Find SimilarView on arXiv

What Meaning-Form Correlation Has to Compose With

December 7, 2020

89% Match
Timothee Mickus, Timothée Bernard, Denis Paperno
Computation and Language

Compositionality is a widely discussed property of natural languages, although its exact definition has been elusive. We focus on the proposal that compositionality can be assessed by measuring meaning-form correlation. We analyze meaning-form correlation on three sets of languages: (i) artificial toy languages tailored to be compositional, (ii) a set of English dictionary definitions, and (iii) a set of English sentences drawn from literature. We find that linguistic phenome...

Find SimilarView on arXiv

Challenging the principle of compositionality in interpreting natural language texts

September 8, 2006

89% Match
Françoise LIPN Gayral, Daniel LIPN Kayser, François LIPN Lévy
Computation and Language

The paper aims at emphasizing that, even relaxed, the hypothesis of compositionality has to face many problems when used for interpreting natural language texts. Rather than fixing these problems within the compositional framework, we believe that a more radical change is necessary, and propose another approach.

Find SimilarView on arXiv

A Multivariate Model for Representing Semantic Non-compositionality

August 15, 2019

89% Match
Meghdad Farahmand
Computation and Language
Artificial Intelligence

Semantically non-compositional phrases constitute an intriguing research topic in Natural Language Processing. Semantic non-compositionality --the situation when the meaning of a phrase cannot be derived from the meaning of its components, is the main characteristic of such phrases, however, they bear other characteristics such as high statistical association and non-substitutability. In this work, we present a model for identifying non-compositional phrases that takes into a...

Find SimilarView on arXiv

Distributed Representations for Compositional Semantics

November 12, 2014

88% Match
Karl Moritz Hermann
Computation and Language

The mathematical representation of semantics is a key issue for Natural Language Processing (NLP). A lot of research has been devoted to finding ways of representing the semantics of individual words in vector spaces. Distributional approaches --- meaning distributed representations that exploit co-occurrence statistics of large corpora --- have proved popular and successful across a number of tasks. However, natural language usually comes in structures beyond the word level,...

Find SimilarView on arXiv

Paraphrase-Supervised Models of Compositionality

January 31, 2018

88% Match
Avneesh Saluja, Chris Dyer, Jean-David Ruvini
Computation and Language

Compositional vector space models of meaning promise new solutions to stubborn language understanding problems. This paper makes two contributions toward this end: (i) it uses automatically-extracted paraphrase examples as a source of supervision for training compositional models, replacing previous work which relied on manual annotations used for the same purpose, and (ii) develops a context-aware model for scoring phrasal compositionality. Experimental results indicate that...

Find SimilarView on arXiv

A Study of Metrics of Distance and Correlation Between Ranked Lists for Compositionality Detection

March 10, 2017

88% Match
Christina Lioma, Niels Dalum Hansen
Computation and Language

Compositionality in language refers to how much the meaning of some phrase can be decomposed into the meaning of its constituents and the way these constituents are combined. Based on the premise that substitution by synonyms is meaning-preserving, compositionality can be approximated as the semantic similarity between a phrase and a version of that phrase where words have been replaced by their synonyms. Different ways of representing such phrases exist (e.g., vectors [1] or...

Find SimilarView on arXiv

Geometric Signatures of Compositionality Across a Language Model's Lifetime

October 2, 2024

88% Match
Jin Hwa Lee, Thomas Jiralerspong, Lei Yu, ... , Cheng Emily
Computation and Language
Artificial Intelligence
Information Theory
Machine Learning
Information Theory

Compositionality, the notion that the meaning of an expression is constructed from the meaning of its parts and syntactic rules, permits the infinite productivity of human language. For the first time, artificial language models (LMs) are able to match human performance in a number of compositional generalization tasks. However, much remains to be understood about the representational mechanisms underlying these abilities. We take a high-level geometric approach to this probl...

Find SimilarView on arXiv

The paradox of the compositionality of natural language: a neural machine translation case study

August 12, 2021

88% Match
Verna Dankers, Elia Bruni, Dieuwke Hupkes
Computation and Language
Artificial Intelligence
Machine Learning

Obtaining human-like performance in NLP is often argued to require compositional generalisation. Whether neural networks exhibit this ability is usually studied by training models on highly compositional synthetic data. However, compositionality in natural language is much more complex than the rigid, arithmetic-like version such data adheres to, and artificial compositionality tests thus do not allow us to determine how neural models deal with more realistic forms of composi...

Find SimilarView on arXiv

Challenges for Distributional Compositional Semantics

July 10, 2012

88% Match
Daoud Clarke
Computation and Language
Artificial Intelligence

This paper summarises the current state-of-the art in the study of compositionality in distributional semantics, and major challenges for this area. We single out generalised quantifiers and intensional semantics as areas on which to focus attention for the development of the theory. Once suitable theories have been developed, algorithms will be needed to apply the theory to tasks. Evaluation is a major problem; we single out application to recognising textual entailment and ...

Find SimilarView on arXiv