(Book Chapter)Deep Learning in Knowledge Graph #31
Labels
KGP/KGC(T)
Knowledge Graph Population/Construction Task
KRL/KGE((T/M)
Knowledge Representation Learning Task & Knowledge Graph Embedding Method
NEL(T)
Named Entity Linking Task
Survey
Survey/Review
Resource
Knowledge Representation Learning (KRL) / Knowledge Embedding (KE)
Deep Learning in Knowledge Graph
This is the note of Chapter 5 from Deep Learning in Natural Language Processing
5.1 Introduction
5.1.1 Basic Concepts
5.1.2 Typical Knowledge Graphs
5.2 Knowledge Representation Learning
Goal: Embedding the entities and relations in KG.
Recent studies reveal that translation-based representation learning methods are efficient and effective to encode relational facts in KG with low-dimensional representations of both entities and relations, which can alleviate the issue of data sparsity and be further employed to knowledge acquisition, fusion, and inference
Translation-based representation learning methods:
TransE只考虑direct relations between entities.于是有了下面考虑不同relation path的方法
Relation-path-based methods:
上面这些值考虑structure information in KG, 忽视了rich multisource information such as textual information, type information, and visual information.
5.3 Neural Relation Extraction
Goal: automatically finding unknown relational facts
Relation extraction (RE): Relation extraction aims at extracting relational data from plaintexts. In recent years, as the development of deep learning (Bengio 2009) techniques, neural relation extraction adopts an end-to-end neural network to model the relation extraction task.
The framework of neural relation extraction includes a sentence encoder to capture the semantic meaning of the input sentence and represents it as a sentence vector, and a relation extractor to generate the probability distribution of extracted relations according to sentence vectors.
Neural relation extraction (NRE) has two main tasks including sentence-level NRE and document-level NRE
5.3.1 Sentence-Level NRE
Sentence-level NRE aims at predicting the semantic relations between the entity (or nominal) pair in a sentence.
三个部分
5.3.1.1 Input Encoder
这里介绍了四种embedding:
New York is a city of United States
, the relative distance from the word city to New York is 3 and United States is −2.5.3.1.2 Sentence Encoder
sentence encoder负责把输入的embedding变为一个向量来表示一个句子。
5.3.1.3 Relation Classifier
5.3.2 Document-Level NRE
上面的一些方法总是受限于不足的训练样本。为了解决这个问题,研究者提出了distant supervision假设,通过KG来自动生成训练样本。
The intuition of distant supervision assumption is that all sentences that contain two entities will express their relations in KGs. For example, (New York, city of, United States) is a relational fact in KGs. Distant supervision assumption will regard all sentences that contain these two entities as valid instances for relation city of. It offers a natural way to utilize information from multiple sentences (document-level) rather than single sentence (sentence-level) to decide if a relation holds between two entities.
Therefore, document-level NRE aims to predict the semantic relations between an entity pair using all involved sentences. 其实就是multi-instance learning的另一种说法。
Document-Level NRE有四个部分:
5.3.2.1 Document Encoder
The document encodes all sentence vectors into either single vector S.
5.3.2.2 Relation Classifier
5.4 Bridging Knowledge with Text: Entity Linking
实体链接就是研究如何将指代词链接到知识库。比如Jobs leaves Apple这个句子,我们的KB里已经有Steve Jobs这个实体了,如果把"Jobs"链接到"Steve Jobs",其实也是在消除歧义。
The main challenges for entity linking are the name ambiguity problem(实体歧义) and the name variation problem(实体变化).
5.4.1 The Entity Linking Framework
这部分是传统的EL流程,没有涉及深度学习。
Given a document d and a Knowledge Graph K B, an entity linking system links the name mentions in the document as follows. 下面分几个步骤
Name Mention Identification
Candidate Entity Selection
Local Compatibility Computation
5.4.2 Deep Learning for Entity Linking
One main problem of EL is the name ambiguity problem, thus, the key challenge is how to compute the compatibility between a name mention and an entity by effectively using contextual evidences.
现在的EL很大程度上依赖于local compatibility model。即用一些手工制作的特征来表达不同的contextual evidenes. 但是这些feature-engineering-based approaches有缺点:
为了解决这些缺点,提出了基于DL的方法。
5.4.2.1 Representing Heterogeneous Evidences via Neural Networks
NN的一个强项在于对input能有一个很好的表达,比如词向量。
By encoding all contextual evidences in the continuous vector space which are suitable for entity linking, neural networks avoid the need of designing handcrafted features. In following, we introduce how to represent different types of contextual evidences in detail.
但是这种平均的方法没有考虑到单词的位置关系(这个很重要)。
5.4.2.2 Modeling Semantic Interactions Between Contextual Evidences
An EL system needs to take all different types of contextual evidences into consideration。如何利用好更多的背景证据。
Generally, two strategies have been used to model the semantic interactions between different contextual evidences:
5.4.2.3 Learning Local Compatibility Measures
学习局部相容条件,要有一个对应的局部相容指标。
We can see that mention’s evidence and entity’s evidence will be first encoded into a continuous feature space using contextual evidence representation neural networks, then compatibility signals between mention and entity will be com-puted using semantic interaction modeling neural networks, and finally, all these signals will be summarized into the local compatibility score.
The text was updated successfully, but these errors were encountered: