CHAPTER 14 The Representation of Sentence Meaning

该博客是《Speech and Language Processing ed3》第14章的读书笔记,介绍了句子意义表示相关内容。涉及计算需求、模型论语义学、一阶逻辑、事件和状态表示、描述逻辑等。指出创建形式化意义表示是计算语言学处理意义的主要方法,一阶逻辑是实用的表示语言。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

CHAPTER 14 The Representation of Sentence Meaning

Speech and Language Processing ed3 读书笔记

The meaning of linguistic expressions can be captured in formal structures called meaning representations. Correspondingly, the frameworks that specify the syntax and semantics of these representations are called meaning representation languages.

The process of creating and assigning representations to linguistic inputs is called semantic analysis, and the entire enterprise of designing meaning representations and associated semantic analyzers is referred to as computational semantics.

14.1 Computational Desiderata for Representations

14.1.1 Verifiability

The notion of verifiability describes a system’s ability to compare the state of affairs described by a representation to the state of affairs in some world as modeled in a knowledge base.

14.1.2 Unambiguous Representations

ambiguity vs vagueness

ambiguity: has different meanings

vagueness: to some extent has the meaning of

14.1.3 Canonical Form

14.1.4 Inference and Variables

14.1.5 Expressiveness

14.2 Model-Theoretic Semantics

A model is a formal construct that stands for the particular state of affairs in the world. Expressions in a meaning representation language can be mapped in a systematic way to the elements of the model. If the model accurately captures the facts we’re interested in concerning some state of affairs, then a consistent mapping between the meaning
representation and the model provides the bridge between the meaning representation and world being considered.

First, some terminology. The vocabulary of a meaning representation consists of two parts: the non-logical vocabulary and the logical vocabulary. The non-logical vocabulary consists of the open-ended set of names for the objects, properties, and relations that make up the world we’re trying to represent. These appear in various schemes as predicates, nodes, labels on links, or labels in slots in frames, The logical vocabulary consists of the closed set of symbols, operators, quantifiers, links, etc., that provide the formal means for composing expressions in a given meaning representation language.

We’ll start by requiring that each element of the non-logical vocabulary have a denotation in the model. By denotation, we simply mean that every element of the non-logical vocabulary corresponds to a fixed, well-defined part of the model. Let’s start with objects, the most basic notion in most representational schemes. The domain of a model is simply the set of objects that are part of the application, or state of affairs, being represented. Each distinct concept, category, or individual in an application denotes a unique element in the domain. A domain is therefore formally a set.

  • Objects denote elements of the domain
  • Properties denote sets of elements of the domain
  • Relations denote sets of tuples of elements of the domain

A function that maps from the non-logical vocabulary of our meaning representation to the proper denotations in the model is called an interpretation.

One possible state of affairs using this scheme is given in Fig. 14.2.

1561173169719

truth-conditional semantics for conjunction operator

14.3 First-Order Logic

14.3.1 Basic Elements of First-Order Logic

Figure 14.3, which provides a complete context-free grammar for the particular syntax of FOL that we will use, is our roadmap for this section.

1561188226499

14.3.2 Variables and Quantifiers

14.3.3 Lambda Notation

The lambda notation extends the syntax of FOL to include expressions of the following form:
λ x . P ( x ) \lambda x.P(x) λx.P(x)
The usefulness of these λ \lambda λ-expressions is based on the ability to apply them to logical terms to yield new FOL expressions where the formal parameter variables are bound to the specified terms. This process is known as λ \lambda λ-reduction and consists of a simple textual replacement of the λ \lambda λ variables with the specified FOL terms, accompanied by the subsequent removal of the λ \lambda λ. The following expressions illustrate the application of a λ \lambda λ-expression to the constant A A A, followed by the result of performing a λ \lambda λ-reduction on this expression:
λ x . P ( x ) ( A ) P ( A ) \lambda x.P(x)(A) \\ P(A) λx.P(x)(A)P(A)
An important and useful variation of this technique is the use of one λ \lambda λ-expression as the body of another as in the following expression:
λ x . λ y . N e a r ( x , y ) \lambda x.\lambda y.Near(x,y) λx.λy.Near(x,y)

λ x . λ y . N e a r ( x , y ) ( B a c a r o ) λ y . N e a r ( B a c a r o , y ) \lambda x.\lambda y.Near(x,y)(Bacaro)\\ \lambda y.Near(Bacaro, y) λx.λy.Near(x,y)(Bacaro)λy.Near(Bacaro,y)

λ y . N e a r ( B a c a r o , y ) ( C e n t r o ) N e a r ( B a c a r o , C e n t r o ) \lambda y.Near(Bacaro,y)(Centro) \\Near(Bacaro,Centro) λy.Near(Bacaro,y)(Centro)Near(Bacaro,Centro)

14.3.4 The Semantics of First-Order Logic

14.3.5 Inference

Modus ponens

Forward chaining, backward chaining, resolution

plausible reasoning (abduction)

14.4 Event and State Representations

In the neo-Davidsonian approach to event representations:

  • Events are captured with predicates that take a single event variable as an argument.
  • There is no need to specify a fixed number of arguments for a given FOL predicate; rather, as many roles and fillers can be glued on as are provided in the input.
  • No more roles are postulated than are mentioned in the input.
  • The logical connections among closely related inputs that share the same predicate are satisfied without the need for additional inference.

14.4.1 Representing Time

temporal logic

tense logic

(14.48) I arrived in New York.
(14.49) I am arriving in New York.
(14.50) I will arrive in New York.
KaTeX parse error: Expected 'EOF', got '\and' at position 31: …n\ Arriving(e) \̲a̲n̲d̲ ̲ ̲Arriver(e, Spea…
1561255982340

14.4.2 Aspect

Stative: I know my departure gate.
Activity: John is flying.
Accomplishment: Sally booked her flight.
Achievement: She found her gate.

14.5 Description Logics

14.6 Summary

This chapter has introduced the representational approach to meaning. The following are some of the highlights of this chapter:

  • A major approach to meaning in computational linguistics involves the creation of formal meaning representations that capture the meaning-related content of linguistic inputs. These representations are intended to bridge the gap from language to common-sense knowledge of the world.
  • The frameworks that specify the syntax and semantics of these representations are called meaning representation languages. A wide variety of such languages are used in natural language processing and artificial intelligence.
  • Such representations need to be able to support the practical computational requirements of semantic processing. Among these are the need to determine the truth of propositions, to support unambiguous representations, to represent variables, to support inference, and to be sufficiently expressive.
  • Human languages have a wide variety of features that are used to convey meaning. Among the most important of these is the ability to convey a predicate-argument structure.
  • First-Order Logic is a well-understood, computationally tractable meaning representation language that offers much of what is needed in a meaning representation language.
  • Important elements of semantic representation including states and events can be captured in FOL.
  • Semantic networks and frames can be captured within the FOL framework.
  • Modern Description Logics consist of useful and computationally tractable subsets of full First-Order Logic. The most prominent use of a description logic is the Web Ontology Language (OWL), used in the specification of the Semantic Web.
### 大规模掩码视觉表征学习的极限与挑战 大规模掩码视觉表征学习(Masked Visual Representation Learning, MVRL)在计算机视觉领域取得了显著进展,但仍面临诸多局限性和挑战。 #### 数据需求与计算资源消耗 MVRL依赖于大量标注数据来训练深层神经网络。然而,获取高质量的大规模图像数据集不仅成本高昂而且耗时费力。此外,处理这些海量的数据需要强大的硬件支持和长时间的运算周期,这对研究机构和个人开发者构成了巨大障碍[^1]。 #### 表征能力瓶颈 尽管通过自监督方法可以有效减少对手动标签的需求并提高泛化性能,但在某些复杂场景下,当前模型可能无法捕捉到足够的语义信息或空间关系特征,从而影响最终效果。例如,在细粒度分类任务中,仅依靠局部区域遮挡策略难以充分表达目标对象的整体特性[^2]。 #### 泛化性不足 现有技术往往针对特定类型的变换进行了优化设计,当遇到未曾见过的新颖变化形式时表现不佳。比如旋转角度较大、尺度差异明显等情况可能导致预训练阶段学到的知识失效,进而降低迁移至下游应用的效果稳定性。 #### 跨模态融合难题 为了实现更加鲁棒可靠的多源感知理解功能,如何有效地将来自不同感官通道的信息结合起来成为了一个亟待解决的问题之一。目前大多数工作主要集中在单一视域内的探索上,对于跨媒体间交互作用机制的研究相对较少,这限制了其实际应用场景范围扩展的可能性。 ```python import torch.nn as nn class MaskedImageModel(nn.Module): def __init__(self): super(MaskedImageModel, self).__init__() # Define layers here def forward(self, x): pass # Implement forward propagation logic ```
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值