ID: cmp-lg/9505019

Measuring semantic complexity

May 8, 1995

View on ArXiv

Similar papers 4

A New Semantic Theory of Natural Language

September 10, 2017

83% Match
Kun Xing
Computation and Language
Logic in Computer Science

Formal Semantics and Distributional Semantics are two important semantic frameworks in Natural Language Processing (NLP). Cognitive Semantics belongs to the movement of Cognitive Linguistics, which is based on contemporary cognitive science. Each framework could deal with some meaning phenomena, but none of them fulfills all requirements proposed by applications. A unified semantic theory characterizing all important language phenomena has both theoretical and practical signi...

Find SimilarView on arXiv

A Survey on Semantic Parsing

December 3, 2018

83% Match
Aishwarya Kamath, Rajarshi Das
Computation and Language

A significant amount of information in today's world is stored in structured and semi-structured knowledge bases. Efficient and simple methods to query them are essential and must not be restricted to only those who have expertise in formal query languages. The field of semantic parsing deals with converting natural language utterances to logical forms that can be easily executed on a knowledge base. In this survey, we examine the various components of a semantic parsing syst...

Find SimilarView on arXiv

A Learning Approach to Natural Language Understanding

June 1, 1994

83% Match
Roberto AT&T Bell Laboratories Pieraccini, Esther AT&T Bell Laboratories Levin
Computation and Language

In this paper we propose a learning paradigm for the problem of understanding spoken language. The basis of the work is in a formalization of the understanding problem as a communication problem. This results in the definition of a stochastic model of the production of speech or text starting from the meaning of a sentence. The resulting understanding algorithm consists in a Viterbi maximization procedure, analogous to that commonly used for recognizing speech. The algorithm ...

Find SimilarView on arXiv

A Gentle Introduction and Survey on Computing with Words (CWW) Methodologies

August 13, 2022

83% Match
Prashant K. Gupta, Javier Andreu-Perez
Artificial Intelligence
Computation and Language

Human beings have an inherent capability to use linguistic information (LI) seamlessly even though it is vague and imprecise. Computing with Words (CWW) was proposed to impart computing systems with this capability of human beings. The interest in the field of CWW is evident from a number of publications on various CWW methodologies. These methodologies use different ways to model the semantics of the LI. However, to the best of our knowledge, the literature on these methodol...

Find SimilarView on arXiv

Challenging the principle of compositionality in interpreting natural language texts

September 8, 2006

83% Match
Françoise LIPN Gayral, Daniel LIPN Kayser, François LIPN Lévy
Computation and Language

The paper aims at emphasizing that, even relaxed, the hypothesis of compositionality has to face many problems when used for interpreting natural language texts. Rather than fixing these problems within the compositional framework, we believe that a more radical change is necessary, and propose another approach.

Find SimilarView on arXiv

A Categorization of Complexity Classes for Information Retrieval and Synthesis Using Natural Logic

February 28, 2024

83% Match
Gregory Coppola
Information Retrieval

Given the emergent reasoning abilities of large language models, information retrieval is becoming more complex. Rather than just retrieve a document, modern information retrieval systems advertise that they can synthesize an answer based on potentially many different documents, conflicting data sources, and using reasoning. But, different kinds of questions have different answers, and different answers have different complexities. In this paper, we introduce a novel framewor...

Find SimilarView on arXiv

Natural Language Processing - A Survey

September 25, 2012

83% Match
Kevin Mote
Computation and Language

The utility and power of Natural Language Processing (NLP) seems destined to change our technological society in profound and fundamental ways. However there are, to date, few accessible descriptions of the science of NLP that have been written for a popular audience, or even for an audience of intelligent, but uninitiated scientists. This paper aims to provide just such an overview. In short, the objective of this article is to describe the purpose, procedures and practical ...

Find SimilarView on arXiv

Is Complexity an Illusion?

March 31, 2024

83% Match
Michael Timothy Bennett
Artificial Intelligence

Simplicity is held by many to be the key to general intelligence. Simpler models tend to "generalise", identifying the cause or generator of data with greater sample efficiency. The implications of the correlation between simplicity and generalisation extend far beyond computer science, addressing questions of physics and even biology. Yet simplicity is a property of form, while generalisation is of function. In interactive settings, any correlation between the two depends on...

Find SimilarView on arXiv

Syntax-Semantics Interaction Parsing Strategies. Inside SYNTAGMA

January 21, 2016

83% Match
Daniel Christen
Computation and Language

This paper discusses SYNTAGMA, a rule based NLP system addressing the tricky issues of syntactic ambiguity reduction and word sense disambiguation as well as providing innovative and original solutions for constituent generation and constraints management. To provide an insight into how it operates, the system's general architecture and components, as well as its lexical, syntactic and semantic resources are described. After that, the paper addresses the mechanism that perfor...

Find SimilarView on arXiv

Why machines do not understand: A response to S{\o}gaard

July 7, 2023

83% Match
Jobst Landgrebe, Barry Smith
Computation and Language
Artificial Intelligence

Some defenders of so-called `artificial intelligence' believe that machines can understand language. In particular, S{\o}gaard has argued in this journal for a thesis of this sort, on the basis of the idea (1) that where there is semantics there is also understanding and (2) that machines are not only capable of what he calls `inferential semantics', but even that they can (with the help of inputs from sensors) `learn' referential semantics \parencite{sogaard:2022}. We show t...

Find SimilarView on arXiv