[go: up one dir, main page]

CN111506706A - Relationship similarity based upper and lower meaning relationship forest construction method - Google Patents

Relationship similarity based upper and lower meaning relationship forest construction method Download PDF

Info

Publication number
CN111506706A
CN111506706A CN202010296825.8A CN202010296825A CN111506706A CN 111506706 A CN111506706 A CN 111506706A CN 202010296825 A CN202010296825 A CN 202010296825A CN 111506706 A CN111506706 A CN 111506706A
Authority
CN
China
Prior art keywords
relationship
similarity
num
sub
matrix
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010296825.8A
Other languages
Chinese (zh)
Other versions
CN111506706B (en
Inventor
张英杰
方义秋
李春江
葛娜
林颂策
余德华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
Original Assignee
Chongqing University of Post and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications filed Critical Chongqing University of Post and Telecommunications
Priority to CN202010296825.8A priority Critical patent/CN111506706B/en
Publication of CN111506706A publication Critical patent/CN111506706A/en
Application granted granted Critical
Publication of CN111506706B publication Critical patent/CN111506706B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3344Query execution using natural language analysis
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

本发明涉及一种基于关系相似度的上下义关系森林构建方法,属于自然语言处理和知识图谱领域。该方法包括:输入所有三元组,通过传统多层感知机获得所有三元组中关系的实体对概率集合C,并对C作笛卡尔,对笛卡尔积C×C用公式计算得出相似度量值(矩阵S)并用门限值ThresholdA进行过滤,筛选出关系相似度集合;然后结合开放域实体数据集,利用训练好的相似关系在改进的多层感知模型上预测开放域实体中关系对应的实体数量,并用门限值ThresholdB进行过滤,得到Result矩阵;最后根据Result矩阵和关系相似度集合构建关系森林。本发明能够检索知识图谱中相似关系实体;提高关系抽取任务中相似关系抽取的精度。

Figure 202010296825

The invention relates to a method for constructing an upper-lower-sense relational forest based on relational similarity, and belongs to the fields of natural language processing and knowledge graphs. The method includes: inputting all triples, obtaining the entity pair probability set C of relations in all triples through a traditional multilayer perceptron, making a Cartesian for C, and calculating the similarity of the Cartesian product C×C with the formula The measurement value (matrix S) is filtered with the threshold value ThresholdA to filter out the relation similarity set; then combined with the open domain entity data set, the trained similarity relation is used to predict the relation correspondence in the open domain entity on the improved multi-layer perceptual model The number of entities is filtered by the threshold ThresholdB to obtain the Result matrix; finally, the relationship forest is constructed according to the Result matrix and the relationship similarity set. The invention can retrieve similar relationship entities in the knowledge map, and improve the accuracy of similar relationship extraction in the relationship extraction task.

Figure 202010296825

Description

一种基于关系相似度的上下义关系森林构建方法A method of constructing upper and lower relational forests based on relational similarity

技术领域technical field

本发明属于自然语言处理和知识图谱领域,涉及一种基于关系相似度的上下义关系森林 构建方法。The invention belongs to the field of natural language processing and knowledge graph, and relates to a method for constructing an upper-lower-sense relational forest based on relational similarity.

背景技术Background technique

目前学术界和工业界的自然语言处理和知识图谱是一个热度较高的研究领域,自然语言 处理中对于实体对关系的研究得到了长足的发展。At present, natural language processing and knowledge graphs in academia and industry are a hot research field, and the research on entity-to-relationship in natural language processing has made great progress.

现有关于关系相似度的研究主要集中在以下几种方法:Existing research on relational similarity mainly focuses on the following methods:

2006年Peter D.Turney在Similarity of Semantic Relation文中提出关系相似度的概念,作为 关系相似度研究的先驱Turney认为,测量关系相似度即测量关系对应的实体对的相似度,比 如:mason和stone的关系(此处称为关系A),carpenter和wood的关系(此处称为关系B), A与B的关系相似程度根据他们对应的实体对(mason:stone)和实体对(carpenter:wood) 来测量。In 2006, Peter D.Turney proposed the concept of relational similarity in the article Similarity of Semantic Relation. As a pioneer of relational similarity research, Turney believes that measuring relational similarity is to measure the similarity of entity pairs corresponding to the relation, such as: mason and stone's The relationship (referred to as Relation A here), the relationship between carpenter and wood (referred to as Relation B here), the similarity of the relationship between A and B is based on their corresponding entity pair (mason: stone) and entity pair (carpenter: wood) to measure.

2013年Alisa Zhila在NAACL-HLT会议中提出了一种测量关系相似度的方法,Alisa在 Combining Heterogeneous Models for Measuring Relational Similarity文中提出的用向量法来测 量关系相似度的方法为此项研究取得了里程碑的意义。文中提出用RNN循环神经网络语言模 型学习关系词向量,源语料库是由关系对应的各个实体组成的,将实体对输入到RNN循环网 络得到一个实体对嵌入矩阵,矩阵的维度是:实体数×维度,在此文章的实验中维度参数取 1600,假设实体对Wi=(Wi1,Wi2),Wj=(Wj1,Wj2)对应的词嵌入为(vi1,vi2)和(vj1,vj2),则Wi对 应的方向向量为vi=vi2-vi1,Wj对应的方向向量为vj=vj2-vj1,相似度测量公式为:In 2013, Alisa Zhila proposed a method for measuring relational similarity at the NAACL-HLT conference, and Alisa's method for measuring relational similarity using the vector method in Combining Heterogeneous Models for Measuring Relational Similarity achieved a milestone in this research. meaning. In this paper, it is proposed to use the RNN recurrent neural network language model to learn the relationship word vector. The source corpus is composed of various entities corresponding to the relationship, and the entity pair is input into the RNN recurrent network to obtain an entity pair embedding matrix. The dimension of the matrix is: number of entities × dimension , in the experiment of this article, the dimension parameter is set to 1600, assuming that the entity pair Wi = (W i1 , Wi2 ), W j = (W j1 , W j2 ) The corresponding word embeddings are (v i1 , v i2 ) and ( v j1 , v j2 ), then the direction vector corresponding to Wi i is v i =v i2 -v i1 , the direction vector corresponding to W j is v j =v j2 -v j1 , and the similarity measurement formula is:

Figure BDA0002452505050000011
Figure BDA0002452505050000011

2019年清华大学Weize Chen在ACL会议上发表论文Quantifying Similaritybetween Relations with Fact Distribution,此论文中提出用多层感知机求出实体对对应的关系概率即关 系1对应的头尾实体概率为Pθ1=(h,t|r1),关系2对应的头尾实体概率为Pθ2=(h,t|r2),则关系 1和关系2之间的相似度用KL散度来测量:In 2019, Weize Chen of Tsinghua University published a paper Quantifying Similarity between Relations with Fact Distribution at the ACL conference. In this paper, it is proposed to use the multilayer perceptron to find the relationship probability corresponding to the entity pair, that is, the probability of the head and tail entities corresponding to the relationship 1 is P θ1 = ( h,t|r 1 ), the head and tail entity probability corresponding to relation 2 is P θ2 =(h,t|r 2 ), then the similarity between relation 1 and relation 2 is measured by KL divergence:

Figure BDA0002452505050000012
Figure BDA0002452505050000012

求出KL散度后,用函数g(x)进行双向求相似度得到最终关系1和关系2的相似值S(r1,r2):After finding the KL divergence, use the function g(x) to find the similarity in both directions to obtain the similarity value S(r 1 , r 2 ) of the final relation 1 and relation 2:

S(r1,r2)=g(DKL(Pθ1(h,t|r1)||(Pθ2(h,t|r2)),DKL((Pθ1(h,t|r2)||(Pθ2(h,t|r1))) (3)S(r 1 ,r 2 )=g(D KL (P θ1 (h,t|r 1 )||(P θ2 (h,t|r 2 )),D KL ((P θ1 (h,t| r 2 )||(P θ2 (h,t|r 1 ))) (3)

从上述几种方案中可以看出,关系相似度目前在关系抽取任务中应用较广,目前的技术 是将关系相似度直接应用到任务中,比如:利用关系相似度进行关系去重,构建知识图谱等。 但是此类应用缺乏应用结构,直接运用到相应的任务中会导致效果不明显,或者不能根据相 应的任务有效的应用关系相似度技术。It can be seen from the above solutions that relational similarity is currently widely used in relation extraction tasks. The current technology is to directly apply relational similarity to tasks, such as: using relational similarity to deduplicate relationships and build knowledge map etc. However, such applications lack application structure, and direct application to the corresponding tasks will lead to insignificant effects, or the relational similarity technology cannot be effectively applied according to the corresponding tasks.

发明内容SUMMARY OF THE INVENTION

有鉴于此,本发明的目的在于提供一种基于关系相似度的上下义关系森林构建方法,根 据关系相似度和一定的规则来构建关系森林。构建好关系森林后各个任务在应用关系相似度 技术时,检索关系森林来获得相应的关系对象。使得关系相似度技术与其下游任务进行解耦 合,且此技术高内聚。①由此缓解目前关系抽取任务中由于关系相似,关系抽取精度不高的 问题。②在构建知识图谱中可以利用构建好的关系森林检索相似关系的实体。③在关系推理 任务中利用关系森林获得相似关系来推理获得下一个对象。In view of this, the purpose of the present invention is to provide a method for constructing a relational forest of upper and lower sense based on the relational similarity, and constructs the relational forest according to the relational similarity and certain rules. After building the relational forest, each task retrieves the relational forest to obtain the corresponding relational objects when applying relational similarity technology. The relational similarity technology is decoupled from its downstream tasks, and the technology is highly cohesive. ① This alleviates the problem of low accuracy of relation extraction in the current relation extraction task due to the similarity of relations. ②In the construction of knowledge graph, the constructed relation forest can be used to retrieve entities with similar relations. (3) In the relational reasoning task, the relational forest is used to obtain similar relations to reason to obtain the next object.

为达到上述目的,本发明提供如下技术方案:To achieve the above object, the present invention provides the following technical solutions:

一种基于关系相似度的上下义关系森林构建方法,包括以下步骤:A method for constructing upper and lower relational forests based on relational similarity, comprising the following steps:

S1:输入所有三元组,通过传统多层感知机获得所有三元组中关系的实体对概率集合 C=(Pθ1,Pθ2,Pθ3,…,Pθn),n为关系个数;S1: Input all triples, and obtain the entity pair probability set C=(P θ1 , P θ2 , P θ3 ,..., P θn ) of relations in all triples through traditional multilayer perceptron, and n is the number of relations;

S2:将集合C与其自身做笛卡尔积得到一个n×n大小的集合,将笛卡尔积中的两个元 素相同的一对删除,剩下n2-n对元素,然后通过公式计算每一对元素的关系相似度量值S(ri,rj),其中i=1,2,…,n,j=1,2,…,n;S2: Do the Cartesian product of the set C with itself to obtain a set of size n×n, delete the same pair of two elements in the Cartesian product, and leave n 2 -n pairs of elements, and then calculate each The relationship similarity measure value S(r i ,r j ) for the elements, where i=1,2,...,n, j=1,2,...,n;

S3:设定门限值ThresholdA,并筛选出大于等于ThresholdA的关系度量值S(ri,rj)作为相 似关系集合F(ri,rj);S3: Set the threshold value ThresholdA, and filter out the relationship metric value S(r i , r j ) that is greater than or equal to ThresholdA as the similar relationship set F(r i , r j );

S4:改进多层感知机模型,分别将关系词向量或实体词向量作为改进的多层感知机模型 的输入,输出为Tmp_rel矩阵或Tmp_ent矩阵;S4: Improve the multi-layer perceptron model, respectively use the relation word vector or the entity word vector as the input of the improved multi-layer perceptron model, and the output is a Tmp_rel matrix or a Tmp_ent matrix;

S5:将关系词向量和实体词向量输入多层感知机后的输出结果进行矩阵相乘即Tmp_rel*Tmp_ent得到Result_tmp矩阵,维度为关系个数×实体个数;S5: Multiply the output result after inputting the relation word vector and the entity word vector into the multi-layer perceptron, that is, Tmp_rel*Tmp_ent to obtain the Result_tmp matrix, and the dimension is the number of relations × the number of entities;

S6:对Result_tmp矩阵的每一行求Softmax,表示对每一个关系对应的实体求得概率; 并设定门限值ThresholdB,筛选出Result_tmp矩阵中概率值大于ThresholdB的实体,得到 Result矩阵,即Result=(soft max(Result_tmp)>=TresholdB);S6: Calculate Softmax for each row of the Result_tmp matrix, indicating that a probability is obtained for the entity corresponding to each relationship; and set a threshold value ThresholdB, screen out the entities whose probability value is greater than ThresholdB in the Result_tmp matrix, and obtain the Result matrix, that is, Result= (soft max(Result_tmp)>=TresholdB);

S7:根据相似关系集合F(ri,rj)和Result矩阵构建多叉树;S7: Construct a polytree according to the similarity relationship set F(r i ,r j ) and the Result matrix;

S8:由多个多叉树构成关系森林。S8: The relational forest is composed of multiple multi-fork trees.

进一步,所述步骤S4中,改进的多层感知机模型的感知机为5层,hidden为1024即每 一层有1024个神经单元;其中,关系词向量维度为关系词个数×神经单元个数,Tmp_rel矩 阵维度为神经单元个数×关系个数,Tmp_ent矩阵维度为神经单元个数×实体词个数。Further, in the step S4, the perceptron of the improved multi-layer perceptron model is 5 layers, and the hidden value is 1024, that is, each layer has 1024 neural units; wherein, the dimension of the relation word vector is the number of relation words × the number of neural units The dimension of the Tmp_rel matrix is the number of neural units × the number of relations, and the dimension of the Tmp_ent matrix is the number of neural units × the number of entity words.

进一步,所述步骤S7中,构建多叉树具体包括:统计Result矩阵中每一行的数量rn_num, 结合相似关系集合F(ri,rj);相似关系中,假设对应实体数数量大的则作为上义关系,对应实 体数量少的则作为下义关系;Further, in the step S7, constructing a polytree specifically includes: counting the number r n _num of each row in the Result matrix, and combining the similarity relationship set F(r i ,r j ); in the similarity relationship, it is assumed that the number of corresponding entities is large If the number of corresponding entities is small, it will be regarded as a hyponymous relationship;

情况1:若F(ri,rj)中存在相似关系S(r1,r2),S(r1,r3),则根据r1,r2,r3找出在Result矩阵中对 应的r1_num,r2_num,r3_num;Case 1: If there is a similarity relationship S(r 1 , r 2 ), S(r 1 , r 3 ) in F(r i , r j ), then find out in the Result matrix according to r 1 , r 2 , r 3 corresponding r 1 _num, r 2 _num, r 3 _num;

情况2:若F(ri,rj)中存在相似关系S(r1,r2),S(r1,r3),S(r1,r4),S(r3,r5),则根据r1,r2,r3找出在 Result矩阵中对应的r1_num,r2_num,r3_num,r4_num,r5_num;Case 2: If there is a similarity relationship S(r 1 , r 2 ), S(r 1 , r 3 ), S(r 1 , r 4 ), S(r 3 , r 5 ) in F(r i , r j ) ), then find out the corresponding r 1 _num, r 2 _num, r 3 _num, r 4 _num, r 5 _num in the Result matrix according to r 1 , r 2 , r 3 ;

以此类推,根据相似关系集合F(ri,rj)和Result矩阵构建多叉树。By analogy, a polytree is constructed according to the similarity relation set F(r i ,r j ) and the Result matrix.

更进一步,情况1中,R1=softmax(r1_num),R2=softmax(r2_num),R3=softmax(r3_num), 取根节点Root=max(R1,R2,R3),其余为子节点。Further, in case 1, R1=softmax(r 1 _num), R2=softmax(r 2 _num), R3=softmax(r 3 _num), take the root node Root=max(R1, R2, R3), and the rest are child node.

更进一步,情况2中,R1=softmax(r1_num),R2=softmax(r2_num),R3=softmax(r3_num), R4=softmax(r4_num),R5=softmax(r5_num),取根节点Root=max(R1,R2,R3,R4,R5),其余 为子节点。Furthermore, in case 2, R1=softmax(r 1 _num), R2=softmax(r 2 _num), R3=softmax(r 3 _num), R4=softmax(r 4 _num), R5=softmax(r 5 _num) ), take the root node Root=max(R1, R2, R3, R4, R5), and the rest are child nodes.

本发明的有益效果在于:The beneficial effects of the present invention are:

1)现有关系抽取技术中由于存在关系相似的问题导致关系抽取精度不高,利用本发明关 系森林进行关系去冗可以有效缓解关系抽取精度不高的问题。1) In the existing relationship extraction technology, due to the problem of similar relationships, the relationship extraction accuracy is not high. Using the relationship forest of the present invention to remove redundant relationships can effectively alleviate the problem of low relationship extraction accuracy.

2)在推荐系统任务中,利用本发明关系森林,有效发现相似关系的推荐内容,提高推荐 精度。2) In the task of the recommendation system, the relationship forest of the present invention is used to effectively discover the recommended content of the similar relationship and improve the recommendation accuracy.

3)在关系推理任务中,在关系森林中检索相似关系,提高关系推理的深度和广度。3) In relational reasoning tasks, similar relations are retrieved in relational forests to improve the depth and breadth of relational reasoning.

本发明的其他优点、目标和特征在某种程度上将在随后的说明书中进行阐述,并且在某 种程度上,基于对下文的考察研究对本领域技术人员而言将是显而易见的,或者可以从本发 明的实践中得到教导。本发明的目标和其他优点可以通过下面的说明书来实现和获得。Other advantages, objects, and features of the present invention will be set forth in the description that follows, and will be apparent to those skilled in the art based on a study of the following, to the extent that is taught in the practice of the present invention. The objectives and other advantages of the present invention may be realized and attained by the following description.

附图说明Description of drawings

为了使本发明的目的、技术方案和优点更加清楚,下面将结合附图对本发明作优选的详 细描述,其中:In order to make the purpose, technical scheme and advantages of the present invention clearer, the present invention will be preferably described in detail below in conjunction with the accompanying drawings, wherein:

图1为传统多层原感知机模型结构示意图;Figure 1 is a schematic structural diagram of a traditional multi-layer original perceptron model;

图2为本发明的上下义关系森林构建方法流程图;Fig. 2 is the flow chart of the construction method of the upper and lower sense relation forest of the present invention;

图3为本发明改进的多层感知机模型结构示意图,(a)为输入为关系词向量的内部结构 示意图,(b)为输入为实体词向量的内部结构示意图;Fig. 3 is the improved multilayer perceptron model structure schematic diagram of the present invention, (a) is the internal structure schematic diagram that input is relation word vector, (b) is the internal structure schematic diagram that input is entity word vector;

图4为本发明构建多叉树中情况1的节点关系示意图;Fig. 4 is the node relation schematic diagram of situation 1 in the present invention constructs multi-fork tree;

图5为本发明构建多叉树中情况2的节点关系示意图;5 is a schematic diagram of the node relationship of the situation 2 in the construction of a multi-node tree according to the present invention;

图6为由多个多叉树构成的关系森林示意图;Fig. 6 is the relational forest schematic diagram that is formed by a plurality of multi-fork trees;

图7为本发明构建的关系森林在推荐系统中应用到搜索引擎的流程图。FIG. 7 is a flow chart of applying the relational forest constructed by the present invention to a search engine in a recommendation system.

具体实施方式Detailed ways

以下通过特定的具体实例说明本发明的实施方式,本领域技术人员可由本说明书所揭露 的内容轻易地了解本发明的其他优点与功效。本发明还可以通过另外不同的具体实施方式加 以实施或应用,本说明书中的各项细节也可以基于不同观点与应用,在没有背离本发明的精 神下进行各种修饰或改变。需要说明的是,以下实施例中所提供的图示仅以示意方式说明本 发明的基本构想,在不冲突的情况下,以下实施例及实施例中的特征可以相互组合。The embodiments of the present invention are described below through specific specific examples, and those skilled in the art can easily understand other advantages and effects of the present invention from the contents disclosed in this specification. The present invention can also be implemented or applied through other different specific embodiments, and various details in this specification can also be modified or changed based on different viewpoints and applications without departing from the spirit of the present invention. It should be noted that the drawings provided in the following embodiments are only used to illustrate the basic idea of the present invention in a schematic manner, and the following embodiments and features in the embodiments can be combined with each other without conflict.

请参阅图1~图6,图2为一种基于关系相似度的上下义关系森林构建方法流程图,该方 法具体包括以下步骤:Please refer to Fig. 1 to Fig. 6, Fig. 2 is a flowchart of a method for constructing an upper-lower relationship forest based on relationship similarity, and the method specifically includes the following steps:

步骤一:本发明实施例的语料来自知识库中抽取的三元组(实体1,关系,实体2),根 据图1所示的多层感知机模型获得关系1的实体对概率Pθ1=(h,t|r1)和关系2的实体对概率 Pθ2=(h,t|r2),其中,h:(head)实体1,t:(tail)实体2,r:(relation)关系然后通过公式(2)和公式(3)得到关系1与关系2的关系相似度Simr1,r2Step 1: The corpus of the embodiment of the present invention comes from triples (entity 1, relationship, entity 2) extracted from the knowledge base, and the entity pair probability P θ1 =( h,t|r 1 ) and relation 2 entity pair probability P θ2 =(h,t|r 2 ), where h: (head) entity 1, t: (tail) entity 2, r: (relation) relation Then, the relationship similarity Sim r1,r2 of the relationship 1 and the relationship 2 is obtained through the formula (2) and the formula (3).

步骤二:通过多层感知机获得所有三元组中关系的实体对概率集合 C=(Pθ1,Pθ2,Pθ3,…,Pθn)(n为关系个数),将集合C与其自身做笛卡尔积得到一个n×n大小 的集合。Step 2: Obtain the entity pair probability set C=(P θ1 ,P θ2 ,P θ3 ,...,P θn ) (n is the number of relations) in all triples through the multi-layer perceptron, and compare the set C with itself Do a Cartesian product to get a set of size n×n.

将表1中笛卡尔积中的两个元素相同的一对删除,剩下n2-n对元素每一对元素用公式(2) 和公式(3)得出一个关系相似度量值S(ri,rj)(i=1,2,…,n,j=1,2,…,n)。Delete the same pair of two elements in the Cartesian product in Table 1, leaving n 2 -n pairs of elements. Each pair of elements uses formula (2) and formula (3) to obtain a relationship similarity measure S(r i , r j ) (i=1,2,...,n, j=1,2,...,n).

表1 C×C(笛卡尔积)Table 1 C×C (Cartesian product)

P<sub>θ1</sub> P<sub>θ1</sub>P<sub>θ1</sub> P<sub>θ1</sub> P<sub>θ1</sub> P<sub>θ2</sub>P<sub>θ1</sub> P<sub>θ2</sub> P<sub>θ1</sub> P<sub>θ3</sub>P<sub>θ1</sub> P<sub>θ3</sub> ……... P<sub>θ1</sub> P<sub>θn</sub>P<sub>θ1</sub> P<sub>θn</sub> P<sub>θ2</sub> P<sub>θ1</sub>P<sub>θ2</sub> P<sub>θ1</sub> P<sub>θ2</sub> P<sub>θ2</sub>P<sub>θ2</sub> P<sub>θ2</sub> P<sub>θ2</sub> P<sub>θ3</sub>P<sub>θ2</sub> P<sub>θ3</sub> ……... P<sub>θ2</sub> P<sub>θn</sub>P<sub>θ2</sub> P<sub>θn</sub> P<sub>θ3</sub> P<sub>θ1</sub>P<sub>θ3</sub> P<sub>θ1</sub> P<sub>θ3</sub> P<sub>θ2</sub>P<sub>θ3</sub> P<sub>θ2</sub> P<sub>θ3</sub> P<sub>θ3</sub>P<sub>θ3</sub> P<sub>θ3</sub> ……... P<sub>θ3</sub> P<sub>θn</sub>P<sub>θ3</sub> P<sub>θn</sub> ……... ……... ……... ……... ……... P<sub>θn</sub> P<sub>θ1</sub>P<sub>θn</sub> P<sub>θ1</sub> P<sub>θn</sub> P<sub>θ2</sub>P<sub>θn</sub> P<sub>θ2</sub> P<sub>θn</sub> P<sub>θ3</sub>P<sub>θn</sub> P<sub>θ3</sub> ……... P<sub>θn</sub> P<sub>θn</sub>P<sub>θn</sub> P<sub>θn</sub>

步骤三:取门限值ThresholdA=0.7,取表2中所有S(ri,rj)>=ThresholdA的关系度量值, 经过此条件过滤后的S(ri,rj)值称为相似关系。从表2提取出相似关系后的关系集合为F(ri,rj)。Step 3: Take the threshold value ThresholdA=0.7, take all the relational metric values of S(r i ,r j )>=ThresholdA in Table 2, the S(r i ,r j ) value filtered by this condition is called similar relation. The relation set after the similarity relation is extracted from Table 2 is F(r i ,r j ).

表2关系相似度量值矩阵Table 2 Relationship similarity measure matrix

S(r<sub>1</sub>,r<sub>1</sub>)S(r<sub>1</sub>,r<sub>1</sub>) S(r<sub>1</sub>,r<sub>2</sub>)S(r<sub>1</sub>,r<sub>2</sub>) S(r<sub>1</sub>,r<sub>3</sub>)S(r<sub>1</sub>,r<sub>3</sub>) ……... S(r<sub>1</sub>,r<sub>n</sub>)S(r<sub>1</sub>,r<sub>n</sub>) S(r<sub>2</sub>,r<sub>1</sub>)S(r<sub>2</sub>,r<sub>1</sub>) S(r<sub>2</sub>,r<sub>2</sub>)S(r<sub>2</sub>,r<sub>2</sub>) S(r<sub>2</sub>,r<sub>3</sub>)S(r<sub>2</sub>,r<sub>3</sub>) ……... S(r<sub>2</sub>,r<sub>n</sub>)S(r<sub>2</sub>,r<sub>n</sub>) S(r<sub>3</sub>,r<sub>1</sub>)S(r<sub>3</sub>,r<sub>1</sub>) S(r<sub>3</sub>,r<sub>2</sub>)S(r<sub>3</sub>,r<sub>2</sub>) S(r<sub>3</sub>,r<sub>3</sub>)S(r<sub>3</sub>,r<sub>3</sub>) ……... S(r<sub>3</sub>,r<sub>n</sub>)S(r<sub>3</sub>,r<sub>n</sub>) ……... ……... ……... ……... ……... S(r<sub>n</sub>,r<sub>1</sub>)S(r<sub>n</sub>,r<sub>1</sub>) S(r<sub>n</sub>,r<sub>2</sub>)S(r<sub>n</sub>,r<sub>2</sub>) S(r<sub>n</sub>,r<sub>3</sub>)S(r<sub>n</sub>,r<sub>3</sub>) ……... S(r<sub>n</sub>,r<sub>n</sub>) S(r<sub>n</sub>,r<sub>n</sub>)

步骤四:设计新的模型多层感知机,感知机为5层,hidden为1024即每一层有1024个 神经单元。如图3(a)所示,将关系词向量作为多层感知机模型的输入,关系词向量维度为 (关系词个数×神经单元个数);输出为Tmp_rel矩阵,维度为(神经单元个数×关系个数)。 如图3(b)所示,将实体词向量作为多层感知机模型的输入,实体词向量维度为(实体词个 数×神经单元个数);输出为Tmp_ent矩阵,维度为(神经单元个数×实体词个数)。Step 4: Design a new model multi-layer perceptron, the perceptron is 5 layers, and the hidden value is 1024, that is, each layer has 1024 neural units. As shown in Figure 3(a), the relation word vector is used as the input of the multilayer perceptron model, and the dimension of the relation word vector is (the number of relation words × the number of neural units); the output is the Tmp_rel matrix, and the dimension is (the number of neural units) number × number of relationships). As shown in Figure 3(b), the entity word vector is used as the input of the multi-layer perceptron model, and the dimension of the entity word vector is (the number of entity words × the number of neural units); the output is the Tmp_ent matrix, and the dimension is (the number of neural units) number × number of entity words).

步骤五:将关系词向量和实体词向量输入多层感知机后的结果进行矩阵相乘即Tmp_rel*Tmp_ent得到Result_tmp矩阵,矩阵维度为(关系个数×实体个数)。Step 5: Multiply the result of the relationship word vector and the entity word vector into the multi-layer perceptron, that is, Tmp_rel*Tmp_ent to obtain the Result_tmp matrix, and the matrix dimension is (the number of relations × the number of entities).

步骤六:对Result_tmp每一行求Softmax.表示对每一个关系对应的实体求得概率,设门 限值ThresholdB=0.8。对矩阵Result_tmp进行过滤,保留Result矩阵中概率值大于0.8的实 体。Step 6: Calculate Softmax for each row of Result_tmp. It means to obtain the probability for the entity corresponding to each relation, and set the threshold value ThresholdB=0.8. Filter the matrix Result_tmp to retain entities with probability values greater than 0.8 in the Result matrix.

Result_tmp=Tmp_rel*Tmp_ent (4)Result_tmp=Tmp_rel*Tmp_ent (4)

Result=(soft max(Result_tmp)>=TresholdB) (5)Result=(soft max(Result_tmp)>=TresholdB) (5)

此处作假设:在相似关系中,关系对应的实体数量大的则作为上义关系,关系对应实体 数量少的则作为下义关系。The assumption is made here: in the similarity relationship, a relationship with a large number of entities corresponding to the relationship is regarded as a superordinate relationship, and a relationship with a small number of corresponding entities is regarded as a hyponymous relationship.

步骤七:根据相似关系集合F(ri,rj)和Result矩阵构建多叉树;Step 7: Construct a polytree according to the similarity relationship set F(r i ,r j ) and the Result matrix;

统计Result矩阵中每一行的个数,结合相似关系集合F(ri,rj)。相似关系中,对应实体数 数量大的则作为上义关系,对应实体数量少的则作为下义关系。Count the number of each row in the Result matrix, and combine the similarity relationship set F(r i ,r j ). In the similarity relationship, the one with a large number of corresponding entities is regarded as a superordinate relationship, and the one with a small number of corresponding entities is regarded as a subordinate relationship.

情况1:如图4所示,若F(ri,rj)中存在相似关系S(r1,r2),S(r1,r3),则根据r1,r2,r3找出在Result 矩阵中对应的r1_num,r2_num,r3_num。Case 1: As shown in Figure 4, if there is a similarity relationship S(r 1 , r 2 ), S(r 1 , r 3 ) in F(ri , r j ), then according to r 1 , r 2 , r 3 Find the corresponding r 1 _num, r 2 _num, r 3 _num in the Result matrix.

R1=softmax(r1_num) (6)R1=softmax(r 1 _num) (6)

R2=softmax(r2_num) (7)R2=softmax(r 2 _num) (7)

R3=softmax(r3_num) (8)R3=softmax(r 3 _num) (8)

取根节点为:Take the root node as:

Root=max(R1,R2,R3) (9)Root=max(R1,R2,R3) (9)

左右子节点为R2,R3。The left and right child nodes are R2, R3.

情况2:如图5所示,若F(ri,rj)中存在相似关系S(r1,r2),S(r1,r3),S(r1,r4),S(r3,r5),则根据 r1,r2,r3找出在Result矩阵中对应的r1_num,r2_num,r3_num,r4_num,r5_num。Case 2: As shown in Figure 5, if there is a similarity relationship S(r 1 , r 2 ), S(r 1 , r 3 ), S(r 1 , r 4 ), S(r 1 , r 3 ), S(r 1 , r 4 ) in F(r i , r j ) (r 3 , r 5 ), then find out the corresponding r 1 _num, r 2 _num, r 3 _num, r 4 _num, r 5 _num in the Result matrix according to r 1 , r 2 , r 3 .

R1=softmax(r1_num) (10)R1=softmax(r 1 _num) (10)

R2=softmax(r2_num) (11)R2=softmax(r 2 _num) (11)

R3=softmax(r3_num) (12)R3=softmax(r 3 _num) (12)

R4=softmax(r4_num) (13)R4=softmax(r 4 _num) (13)

R5=softmax(r5_num) (14)R5=softmax(r 5 _num) (14)

取根节点为:Take the root node as:

Root=max(R1,R2,R3,R4,R5) (15)Root=max(R1,R2,R3,R4,R5) (15)

孩子节点为R2,R3,R4,且R5为R3的子节点。The child nodes are R2, R3, R4, and R5 is a child node of R3.

以此类推,根据F(ri,rj)相似关系集合和Result矩阵构建多叉树。And so on, according to F(r i , r j ) similar relation set and Result matrix to construct multi-fork tree.

步骤八:由多个多叉树构成关系森林。图6中灰色节点为根节点,由root=max(R1,R2,....,Rn)取得,白色节点为子节点,子节点由公式(6)~(15)取得。图6中例举了8组 相似关系。Step 8: Constitute a relational forest from multiple polytrees. In Fig. 6, the gray node is the root node, which is obtained by root=max(R1, R2, . . . , Rn), the white node is the child node, and the child node is obtained by formulas (6) to (15). Eight sets of similarity relationships are illustrated in Figure 6.

如图7所示,本发明构建的关系森林在实际应用中的应用,如应用到推荐系统中用于搜 索引擎,能够快速精准的获取需要的推荐内容。As shown in Figure 7, the application of the relational forest constructed by the present invention in practical applications, such as applying it to a recommendation system for a search engine, can quickly and accurately obtain the required recommendation content.

最后说明的是,以上实施例仅用以说明本发明的技术方案而非限制,尽管参照较佳实施 例对本发明进行了详细说明,本领域的普通技术人员应当理解,可以对本发明的技术方案进 行修改或者等同替换,而不脱离本技术方案的宗旨和范围,其均应涵盖在本发明的权利要求 范围当中。Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the present invention and not to limit them. Although the present invention has been described in detail with reference to the preferred embodiments, those of ordinary skill in the art should understand that the technical solutions of the present invention can be Modifications or equivalent replacements, without departing from the spirit and scope of the technical solution, should all be included in the scope of the claims of the present invention.

Claims (5)

1.一种基于关系相似度的上下义关系森林构建方法,其特征在于,该方法包括以下步骤:1. a method for constructing a relational forest based on relational similarity, characterized in that the method comprises the following steps: S1:输入所有三元组,通过传统多层感知机获得所有三元组中关系的实体对概率集合C=(Pθ1,Pθ2,Pθ3,…,Pθn),其中,概率Pθi为(实体i1,实体i2|关系),n为关系个数;S1: Input all triples, and obtain the entity pair probability set C=(P θ1 , P θ2 , P θ3 ,..., P θn ) through the traditional multilayer perceptron, where the probability P θi is (entity i 1 , entity i 2 | relationship), n is the number of relationships; S2:将集合C与其自身做笛卡尔积得到一个n×n大小的集合,将笛卡尔积中的两个元素相同的一对删除,剩下n2-n对元素,然后通过公式计算每一对元素的关系相似度量值S(ri,rj),其中i=1,2,…,n,j=1,2,…,n;S2: Do the Cartesian product of the set C with itself to obtain a set of size n×n, delete the same pair of two elements in the Cartesian product, and leave n 2 -n pairs of elements, and then calculate each The relationship similarity measure value S(r i ,r j ) for the elements, where i=1,2,...,n, j=1,2,...,n; S3:设定门限值ThresholdA,并筛选出大于等于ThresholdA的关系度量值S(ri,rj)作为相似关系集合F(ri,rj);S3: Set the threshold value ThresholdA, and filter out the relationship metric value S(r i , r j ) that is greater than or equal to ThresholdA as the similar relationship set F(r i , r j ); S4:改进多层感知机模型,分别将关系词向量或实体词向量作为改进的多层感知机模型的输入,输出为Tmp_rel矩阵或Tmp_ent矩阵;S4: Improve the multi-layer perceptron model, respectively use the relation word vector or the entity word vector as the input of the improved multi-layer perceptron model, and the output is a Tmp_rel matrix or a Tmp_ent matrix; S5:将关系词向量和实体词向量输入多层感知机后的输出结果进行矩阵相乘即Tmp_rel*Tmp_ent得到Result_tmp矩阵,维度为关系个数×实体个数;S5: Multiply the output result after inputting the relation word vector and the entity word vector into the multi-layer perceptron, that is, Tmp_rel*Tmp_ent to obtain the Result_tmp matrix, and the dimension is the number of relations × the number of entities; S6:对Result_tmp矩阵的每一行求Softmax,表示对每一个关系对应的实体求得概率;并设定门限值ThresholdB,筛选出Result_tmp矩阵中概率值大于ThresholdB的实体,得到Result矩阵,即Result=(softmax(Result_tmp)>=TresholdB);S6: Calculate Softmax for each row of the Result_tmp matrix, indicating that the probability is obtained for the entity corresponding to each relationship; and set the threshold value ThresholdB, filter out the entities whose probability value is greater than ThresholdB in the Result_tmp matrix, and obtain the Result matrix, that is, Result= (softmax(Result_tmp)>=TresholdB); S7:根据相似关系集合F(ri,rj)和Result矩阵构建多叉树;S7: Construct a polytree according to the similarity relationship set F(r i ,r j ) and the Result matrix; S8:由多个多叉树构成关系森林。S8: The relational forest is composed of multiple multi-fork trees. 2.根据权利要求1所述的一种基于关系相似度的上下义关系森林构建方法,其特征在于,所述步骤S4中,改进的多层感知机模型的感知机为5层,hidden为1024即每一层有1024个神经单元;其中,关系词向量维度为关系词个数×神经单元个数,Tmp_rel矩阵维度为神经单元个数×关系个数,Tmp_ent矩阵维度为神经单元个数×实体词个数。2. a kind of upper and lower relation forest construction method based on relation similarity according to claim 1, is characterized in that, in described step S4, the perceptron of the improved multilayer perceptron model is 5 layers, and hidden is 1024 That is, each layer has 1024 neural units; among them, the dimension of the relation word vector is the number of relation words × the number of neural units, the dimension of the Tmp_rel matrix is the number of neural units × the number of relations, and the dimension of the Tmp_ent matrix is the number of neural units × the number of entities number of words. 3.根据权利要求1所述的一种基于关系相似度的上下义关系森林构建方法,其特征在于,所述步骤S7中,构建多叉树具体包括:统计Result矩阵中每一行的数量rn_num,结合相似关系集合F(ri,rj);相似关系中,假设对应实体数数量大的则作为上义关系,对应实体数量少的则作为下义关系;3. a kind of upper and lower sense relation forest construction method based on relation similarity according to claim 1, is characterized in that, in described step S7, constructing polynomial tree specifically comprises: the quantity r n of each row in statistical Result matrix _num, combined with the similarity relationship set F(r i , r j ); in the similarity relationship, it is assumed that the number of corresponding entities is large as a superordinate relationship, and the number of corresponding entities is small as a hyponymous relationship; 情况1:若F(ri,rj)中存在相似关系S(r1,r2),S(r1,r3),则根据r1,r2,r3找出在Result矩阵中对应的r1_num,r2_num,r3_num;Case 1: If there is a similarity relationship S(r 1 , r 2 ), S(r 1 , r 3 ) in F(r i , r j ), then find out in the Result matrix according to r 1 , r 2 , r 3 corresponding r 1 _num, r 2 _num, r 3 _num; 情况2:若F(ri,rj)中存在相似关系S(r1,r2),S(r1,r3),S(r1,r4),S(r3,r5),则根据r1,r2,r3找出在Result矩阵中对应的r1_num,r2_num,r3_num,r4_num,r5_num;Case 2: If there is a similarity relationship S(r 1 , r 2 ), S(r 1 , r 3 ), S(r 1 , r 4 ), S(r 3 , r 5 ) in F(r i , r j ) ), then find out the corresponding r 1 _num, r 2 _num, r 3 _num, r 4 _num, r 5 _num in the Result matrix according to r 1 , r 2 , r 3 ; 以此类推,根据相似关系集合F(ri,rj)和Result矩阵构建多叉树。By analogy, a polytree is constructed according to the similarity relation set F(r i ,r j ) and the Result matrix. 4.根据权利要求3所述的一种基于关系相似度的上下义关系森林构建方法,其特征在于,情况1中,R1=softmax(r1_num),R2=softmax(r2_num),R3=softmax(r3_num),取根节点Root=max(R1,R2,R3),其余为子节点。4. The method for constructing a relational forest based on relational similarity according to claim 3, characterized in that, in case 1 , R1=softmax(r1_num), R2=softmax( r2_num ), R3 =softmax(r 3 _num), take the root node Root=max(R1, R2, R3), and the rest are child nodes. 5.根据权利要求3所述的一种基于关系相似度的上下义关系森林构建方法,其特征在于,情况2中,R1=softmax(r1_num),R2=softmax(r2_num),R3=softmax(r3_num),R4=softmax(r4_num),R5=softmax(r5_num),取根节点Root=max(R1,R2,R3,R4,R5),其余为子节点。5. A method for constructing a relational similarity forest based on relational similarity according to claim 3, wherein in case 2, R1=softmax( r1_num ), R2=softmax( r2_num ), R3 =softmax(r 3 _num), R4=softmax(r 4 _num), R5=softmax(r 5 _num), take the root node Root=max(R1, R2, R3, R4, R5), and the rest are child nodes.
CN202010296825.8A 2020-04-15 2020-04-15 Relationship similarity based upper and lower meaning relationship forest construction method Active CN111506706B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010296825.8A CN111506706B (en) 2020-04-15 2020-04-15 Relationship similarity based upper and lower meaning relationship forest construction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010296825.8A CN111506706B (en) 2020-04-15 2020-04-15 Relationship similarity based upper and lower meaning relationship forest construction method

Publications (2)

Publication Number Publication Date
CN111506706A true CN111506706A (en) 2020-08-07
CN111506706B CN111506706B (en) 2022-06-17

Family

ID=71867436

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010296825.8A Active CN111506706B (en) 2020-04-15 2020-04-15 Relationship similarity based upper and lower meaning relationship forest construction method

Country Status (1)

Country Link
CN (1) CN111506706B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107944559A (en) * 2017-11-24 2018-04-20 国家计算机网络与信息安全管理中心 A kind of entity relationship automatic identifying method and system
US20180336183A1 (en) * 2017-05-22 2018-11-22 International Business Machines Corporation Deep Embedding for Natural Language Content Based on Semantic Dependencies
CN109635124A (en) * 2018-11-30 2019-04-16 北京大学 A kind of remote supervisory Relation extraction method of combination background knowledge
CN110222199A (en) * 2019-06-20 2019-09-10 青岛大学 A kind of character relation map construction method based on ontology and a variety of Artificial neural network ensembles
CN110309321A (en) * 2019-07-10 2019-10-08 电子科技大学 A knowledge representation learning method based on graph representation learning
CN110472233A (en) * 2019-07-16 2019-11-19 清华大学 Method and system for measuring relationship similarity based on head-tail entity distribution in knowledge base
CN110555083A (en) * 2019-08-26 2019-12-10 北京工业大学 A zero-shot based unsupervised entity relationship extraction method
CN110570920A (en) * 2019-08-20 2019-12-13 华东理工大学 An entity-relation joint learning method based on a concentrated attention model

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180336183A1 (en) * 2017-05-22 2018-11-22 International Business Machines Corporation Deep Embedding for Natural Language Content Based on Semantic Dependencies
CN107944559A (en) * 2017-11-24 2018-04-20 国家计算机网络与信息安全管理中心 A kind of entity relationship automatic identifying method and system
CN109635124A (en) * 2018-11-30 2019-04-16 北京大学 A kind of remote supervisory Relation extraction method of combination background knowledge
CN110222199A (en) * 2019-06-20 2019-09-10 青岛大学 A kind of character relation map construction method based on ontology and a variety of Artificial neural network ensembles
CN110309321A (en) * 2019-07-10 2019-10-08 电子科技大学 A knowledge representation learning method based on graph representation learning
CN110472233A (en) * 2019-07-16 2019-11-19 清华大学 Method and system for measuring relationship similarity based on head-tail entity distribution in knowledge base
CN110570920A (en) * 2019-08-20 2019-12-13 华东理工大学 An entity-relation joint learning method based on a concentrated attention model
CN110555083A (en) * 2019-08-26 2019-12-10 北京工业大学 A zero-shot based unsupervised entity relationship extraction method

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
ALICIA PÉREZ ET AL.: "Semi-supervised medical entity recognition: A study on Spanish and Swedish clinical corpora", 《JOURNAL OF BIOMEDICAL INFORMATICS》 *
BIFAN WEI ET AL.: "Motif-Based Hyponym Relation Extraction from Wikipedia Hyperlinks", 《IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING》 *
孙洋 等: "基于子语义空间的挖掘短文本策略方法", 《电信科学》 *
张英杰: "基于相似度的关系抽取方法研究与实现", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Also Published As

Publication number Publication date
CN111506706B (en) 2022-06-17

Similar Documents

Publication Publication Date Title
CN109902183B (en) Knowledge graph embedding method based on diverse graph attention machine mechanism
CN112116331B (en) Talent recommendation method and device
WO2021083239A1 (en) Graph data query method and apparatus, and device and storage medium
CN103646032B (en) A kind of based on body with the data base query method of limited natural language processing
CN108763376B (en) A Knowledge Representation Learning Method Integrating Relation Path, Type, and Entity Description Information
CN110275947A (en) Domain-specific knowledge graph natural language query method and device based on named entity recognition
CN107545033B (en) A Computational Method for Knowledge Base Entity Classification Based on Representation Learning
CN108874878A (en) A kind of building system and method for knowledge mapping
CN108920556A (en) Recommendation expert method based on subject knowledge map
CN112434024B (en) Data dictionary generation method, device, equipment and medium for relational database
CN113312342A (en) Scientific and technological resource integration system based on multi-source database
CN105335524B (en) A kind of graph search method applied to extensive irregular eutectic data
CN106844407A (en) Label network production method and system based on data set correlation
CN111651447A (en) An intelligent construction full-life data processing, analysis, management and control system
CN109710621A (en) Keyword Search KSANEW Algorithm Combining Semantic Class Nodes and Edge Weights
CN111428056A (en) Method and device for constructing scientific research personnel cooperative community
CN105045863A (en) Method and system used for entity matching
CN110162642A (en) Patent knowledge map construction method based on object broker database
CN106897408A (en) It is a kind of to realize the method that quick search tree structure data specifies node subordinate
CN115344794A (en) A tourist attraction recommendation method based on knowledge graph semantic embedding
CN105138588A (en) Database overlap mode abstract generating method based on multi-label propagation
CN105912602A (en) True-value finding method based on entity attributes
CN102129457A (en) Method for inquiring large-scale semantic data paths
CN105045806A (en) Dynamic splitting and maintenance method of quantile query oriented summary data
CN117217602A (en) Generalized and automatic unmanned aerial vehicle task efficiency evaluation index construction method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant