技术领域technical field
本发明涉及利用句义结构特征的句子相似度计算方法,属于计算机科学及自然语言处理领域。The invention relates to a sentence similarity calculation method using sentence meaning structure features, and belongs to the fields of computer science and natural language processing.
背景技术Background technique
句子相似度计算用于衡量两个文本内容的语义相似程度,是自然语言处理中信息检索、自动摘要等任务的基础环节。伴随社交网站的快速发展,以微博为代表的社交短文本大量涌现,其篇幅短小、表示方式多样化,由于缺失长文档的结构化信息造成传统的句子相似度计算方法无法直接适用于此类短文本的句子相似度计算。Sentence similarity calculation is used to measure the semantic similarity between two texts, and it is the basic link of information retrieval and automatic summarization in natural language processing. With the rapid development of social networking sites, a large number of social short texts represented by Weibo have emerged, which are short in length and diverse in presentation methods. Due to the lack of structured information of long documents, the traditional sentence similarity calculation method cannot be directly applied to this type of text. Sentence similarity calculation for short texts.
目前,根据对句子语义分析的深度不同,针对社交短文本中句子的相似度计算方法主要包括基于词特征、基于词义特征和基于句法分析特征三类。At present, according to the depth of semantic analysis of sentences, the similarity calculation methods for sentences in short social texts mainly include three categories: word-based features, word-meaning-based features, and syntactic analysis-based features.
基于词特征的方法是早期的句子相似度计算方法,该方法主要是将句子视为词的线性组合,采用统计的手段计算句子的词频、词性、句长、词序等词语表层信息,典型方法包括Jaccard Similarity Coefficient字符串匹配,其通过统计两个句子中包含的相同词数目作为句子的相似度、TF-IDF词频统计方法将句子表征为向量,计算余弦距离作为相似度结果。The method based on word features is an early sentence similarity calculation method. This method mainly regards sentences as a linear combination of words, and uses statistical means to calculate the word frequency, part of speech, sentence length, word order and other word surface information of sentences. Typical methods include Jaccard Similarity Coefficient string matching, which counts the number of identical words contained in two sentences as the similarity of the sentences, TF-IDF word frequency statistics method represents the sentence as a vector, and calculates the cosine distance as the similarity result.
基于词义特征的方法从语义分析的角度,通过借助语义知识资源捕捉词语的语义信息。根据利用的资源不同,分为基于语义词典的方法和基于语料库的方法。基于语义词典的方法主要是借助WordNet、HowNet等基于词义组织词汇信息的词汇数据库,结合词义消歧的技术挖掘句子中词语在给定的上下文语境下所表达的涵义,从而提高整个句子的语义分辨率。基于语料库的方法主要通过引入语言模型框架,以两个词共同出现的概率大小来推断其相似性,常用的技术是利用潜在语言分析(Latent Semantic Analysis,LSA)方法将词-文档矩阵进行奇异值分解实现高维特征表示到低维潜在语义空间表示的空间映射。From the perspective of semantic analysis, the method based on semantic features captures the semantic information of words by using semantic knowledge resources. According to the different resources utilized, it can be divided into methods based on semantic dictionary and methods based on corpus. The method based on the semantic dictionary is mainly to use WordNet, HowNet and other lexical databases that organize vocabulary information based on word meaning, combined with word meaning disambiguation technology to mine the meaning expressed by the words in the sentence in a given context, so as to improve the semantics of the entire sentence resolution. The corpus-based method mainly introduces a language model framework to infer the similarity of two words based on the probability of their co-occurrence. A common technique is to use the Latent Semantic Analysis (LSA) method to perform singular value of the word-document matrix. Decomposition enables spatial mapping from high-dimensional feature representations to low-dimensional latent semantic space representations.
基于句法分析特征的方法通过对句子进行整体的结构分析判断句子的相似性,该方法主张句子中核心动词是支配其它成分的关键,核心动词本身不受任何成分的支配,而其他句子成分受到核心动词的支配,通过分析词语之间的依存关系来挖掘句子的语义信息,实际应用中通常只计算动词、名词和形容词等有效词以及直接依附于有效词所组成的搭配对之间的相似性来估算句子的相似度,以此避免增加噪音数据对结果带来的偏差。The method based on the characteristics of syntactic analysis judges the similarity of sentences by analyzing the overall structure of the sentence. This method claims that the core verb in the sentence is the key to dominate other components. The core verb itself is not dominated by any component, while other sentence components are controlled by the core. The dominance of verbs is to mine the semantic information of sentences by analyzing the dependencies between words. In practical applications, only effective words such as verbs, nouns, and adjectives, and the similarity between collocation pairs that are directly attached to effective words are usually calculated. Estimate the similarity of sentences to avoid biasing the results by adding noise data.
上述各类方法虽然从不同的分析层面计算句子的相似度,但社交短文本的实词较少,不加任何句法结构分析和对句子语义信息的挖掘,仅靠对词频、词形等表层信息的统计无法辨别词语的深层信息。基于词义特征的方法虽然考虑了词语的语义信息,但该方法受限于外部语义资源,社交短文本包含大量的未登录词、内容时效性强,由于字典的不全面和特征稀疏的影响往往造成语义信息不明确的问题。基于句法特征的方法受当前句法分析技术不成熟的限制,没有考虑句子上下文信息和深层语义信息,信息的缺失对相似度计算结果的准确性带来无法预估的影响。Although the above-mentioned methods calculate the similarity of sentences from different levels of analysis, the content words in short social texts are less, without any syntactic structure analysis and mining of sentence semantic information, only relying on surface information such as word frequency and word form. Statistics cannot distinguish the deep information of words. Although the semantic feature-based method considers the semantic information of words, this method is limited by external semantic resources. Social short texts contain a large number of unregistered words, and the content is time-sensitive. Due to the influence of incomplete dictionaries and sparse features, it often leads to The problem of ambiguous semantic information. The method based on syntactic features is limited by the immaturity of the current syntactic analysis technology. It does not consider the sentence context information and deep semantic information. The lack of information will have an unpredictable impact on the accuracy of the similarity calculation results.
发明内容Contents of the invention
本发明为解决社交短文本句子相似度计算的特征稀疏和没有考虑深层语义信息的问题,提出利用句义结构特征的句子相似度计算方法。在综合考虑句子的语义信息和上下文信息的前提下,将多元信息加权融合,使句子信息更加全面,通过深度挖掘句子语义信息,使句子相似度计算结果不受表述方式的影响,更加全面、准确地计算句子语义的关联程度。In order to solve the problem of feature sparseness and lack of consideration of deep semantic information in the calculation of social short text sentence similarity, the present invention proposes a sentence similarity calculation method using sentence structure features. On the premise of comprehensively considering the semantic information and context information of the sentence, the multi-information is weighted and fused to make the sentence information more comprehensive. By deeply mining the semantic information of the sentence, the sentence similarity calculation result is not affected by the expression method and is more comprehensive and accurate. Calculate the relevance degree of sentence semantics accurately.
本发明的设计原理为:1)基于句义结构模型(Chinese Semantic StructureModel,CSM)分析句子语义,提取句义成分,利用Latent Dirichlet Allocation(LDA)主题模型挖掘潜在的主题知识,根据知识库扩充句义成分对应维度的特征1112,得到基于句子本身语义信息的句子向量;2)引入Paragraph Vector(PV)深度学习模型自适应地学习文本特征,得到基于句子上下文信息的句子向量;3)利用两种句子向量分别计算得到句子间的相似度,并进行线性加权,通过网格法调节优化系数,使句子相似度计算结果更加准确。The design principles of the present invention are: 1) analyze the sentence semantics based on the Chinese Semantic Structure Model (CSM), extract the sentence components, utilize the Latent Dirichlet Allocation (LDA) topic model to mine potential topic knowledge, and expand the sentence according to the knowledge base The feature 1112 of the dimension corresponding to the semantic component, and obtain the sentence vector based on the semantic information of the sentence itself; 2) introduce the Paragraph Vector (PV) deep learning model to learn text features adaptively, and obtain the sentence vector based on the sentence context information; 3) use two Sentence vectors are calculated separately to obtain the similarity between sentences, and linearly weighted, and the optimization coefficient is adjusted through the grid method to make the sentence similarity calculation result more accurate.
具体步骤如下:Specific steps are as follows:
步骤1,对社交短文本集进行预处理,先进行分句,然后进行分词和词性标注,去停用词。Step 1. Preprocess the social short text set, first segment sentences, and then perform word segmentation and part-of-speech tagging to remove stop words.
步骤2,基于利用CSM对每条句子的句义结构分析结果和利用LDA主题模型对短文本集进行分析得到的主题和词语分布,对句子进行特征扩充,并计算句子相似度。Step 2. Based on the analysis results of the semantic structure of each sentence by using CSM and the topic and word distribution obtained by analyzing the short text set by using the LDA topic model, the feature expansion of the sentence is carried out, and the sentence similarity is calculated.
步骤2.1,在步骤1的基础上,对每条句子进行句义结构分析,提取句子的话题、述题、基本项、一般项。CSM将整个句义的语义表示为结构树的形式,具体表示为句型层、描述层、对象层和细节层四个层次。句型层指明句子的句义类型,包括简单句义、复杂句义、复合句义、多重句义四中类型;描述层中包含话题和述题,话题和述题是对句义的初步划分,是句义结构中的基本句义成分,话题定义为句义中的被描述对象,述题定义为句义中的话题的描述内容;对象层中包含谓词、基本项、一般项、语义格,语义格是对词语的语义标注,包括7种基本格和12种一般格,基本项定义为句义中与谓词具有直接联系的成分,构成一个句子语义的主干,其对应的语义格为基本格,一般项定义为句义中的修饰成分,其对应的语义格为一般格;细节层中包含句子的引申含义。Step 2.1, on the basis of step 1, analyze the sentence structure of each sentence, and extract the topic, statement, basic items, and general items of the sentence. CSM expresses the semantics of the entire sentence as a structure tree, which is specifically expressed as four levels: sentence type layer, description layer, object layer and detail layer. The sentence type layer indicates the sentence meaning type of the sentence, including four types: simple sentence meaning, complex sentence meaning, compound sentence meaning, and multiple sentence meaning; the description layer includes topic and topic, which are the preliminary division of sentence meaning , is the basic sentence meaning component in the sentence meaning structure, the topic is defined as the described object in the sentence meaning, and the topic is defined as the description content of the topic in the sentence meaning; the object layer includes predicates, basic items, general items, semantic case , the semantic case is the semantic labeling of words, including 7 basic cases and 12 general cases. The general item is defined as the modifier in the sentence meaning, and its corresponding semantic case is the general case; the detail layer contains the extended meaning of the sentence.
步骤2.2,利用LDA主题模型对短文本集进行分析,挖掘文本中的潜在主题知识,提取文本中的主题和主题下的词语分布,得到文本-主题矩阵和主题-词语矩阵。LDA主题模型可以获取文本中的主题,可以用来对文本中的词语进行划分,同一主题下的词语具有相同或相似的语义。Step 2.2: Use the LDA topic model to analyze the short text set, mine the potential topic knowledge in the text, extract the topics in the text and the word distribution under the topic, and obtain the text-topic matrix and topic-word matrix. The LDA topic model can obtain the topics in the text and can be used to divide the words in the text. The words under the same topic have the same or similar semantics.
步骤2.3,根据话题对句子进行特征扩充,得到基于话题的句子向量。如果两个相同的词语在句子中分别充当话题和述题的一部分,则认为这两个词语具有不同的语义,定义这两个词语为不同的词语,根据此定义,对句子进行特征扩充时,应分别根据话题和述题部分对句子进行特征扩充。句子的话题部分的特征扩充具体方法为:首先提取话题下的基本项和一般项对应的词语,然后根据步骤2.2中得到的主题-词语矩阵,比较词语在不同主题下的概率,选取概率最高的主题,将该主题下的其它词语补充到句子中,作为句子的一部分,最后使用句子的所有词语作为特征,构建特征向量表示句子,其中句中原有词语所对应的维度上的取值为词语的在句中的出现次数,而扩充的词语所对应的维度上的取值按公式(1)进行计算,In step 2.3, the feature of the sentence is expanded according to the topic, and the sentence vector based on the topic is obtained. If two identical words serve as part of the topic and the topic respectively in the sentence, the two words are considered to have different semantics, and these two words are defined as different words. According to this definition, when the sentence is extended with features, Sentences should be feature-expanded according to topic and recitation parts respectively. The specific method of feature expansion of the topic part of the sentence is as follows: first extract the words corresponding to the basic items and general items under the topic, and then compare the probabilities of words under different topics according to the topic-word matrix obtained in step 2.2, and select the one with the highest probability Topic, add other words under the topic to the sentence as part of the sentence, and finally use all the words in the sentence as features to construct a feature vector to represent the sentence, where the value of the dimension corresponding to the original word in the sentence is the value of the word The number of occurrences in the sentence, and the value on the dimension corresponding to the expanded word is calculated according to formula (1),
V=n*w (1)V=n*w (1)
V是扩充词语对应维度上的取值,n是扩充词语在句子中出现的次数,w为扩充词语在对应主题下的概率值。V is the value of the corresponding dimension of the extended word, n is the number of times the extended word appears in the sentence, and w is the probability value of the extended word under the corresponding topic.
步骤2.4,按步骤2.3的方法,根据述题对句子进行特征扩充,得到基于述题的句子向量。In step 2.4, according to the method of step 2.3, the feature of the sentence is expanded according to the question, and the sentence vector based on the question is obtained.
步骤2.5,分别基于步骤2.3和2.4得到的两种句子向量计算句子相似度,对两个相似度值进行加权,得到句子间的最终相似度值,具体计算公式如下,In step 2.5, the sentence similarity is calculated based on the two sentence vectors obtained in steps 2.3 and 2.4 respectively, and the two similarity values are weighted to obtain the final similarity value between sentences. The specific calculation formula is as follows,
其中,SA和SB代表任意两个句子,sim1(SA,SB)表示两个句子的相似度值,和分别表示句子SA和SB的基于话题的句子向量,和分别表示句子SA和SB的表示基于述题的句子向量,ω为可调参数,取值范围为[0,1],用来调整两种相似度的加权系数。Among them, SA and SB represent any two sentences, sim1(SA , SB ) represents the similarity value of the two sentences, and denote the topic-based sentence vectors of sentences SA and SB respectively, and Sentences SA and SB respectively represent the sentence vector based on the question, and ω is an adjustable parameter with a value range of [0,1], which is used to adjust the weighting coefficient of the two similarities.
步骤3,将经过步骤1预处理后的所有句子输入到PV深度学习模型,利用PV模型学习文本特征,得到句子向量,并基于该句子向量计算句子间的余弦距离作为句子间的相似度,计算公式如下,Step 3, input all the sentences preprocessed in step 1 into the PV deep learning model, use the PV model to learn text features, obtain sentence vectors, and calculate the cosine distance between sentences based on the sentence vectors as the similarity between sentences, calculate The formula is as follows,
其中,SA和SB代表任意两个句子,sim2(SA,SB)表示两个句子的相似度值,和分别表示用PV模型学习得到的句子向量。PV模型是一种非监督的学习方式,输入是任意长度的文本(这里文本可以是文章、段落、句子等任意形式,统称为文本),输出则是对应文本的连续分布式向量表示,同word2vec词向量的原理类似,在保留语义和词序信息的基础上通过对特征学习得到有效的句子或者篇章的向量表示,PV模型能有效解决词袋模型没有考虑词义和词序的问题,生成的向量维度稠密,也可有效克服短文本的句子表示的特征稀疏问题。Among them, SA and SB represent any two sentences, sim2(SA , SB ) represents the similarity value of the two sentences, and Represent the sentence vectors learned by the PV model, respectively. The PV model is an unsupervised learning method. The input is a text of any length (here the text can be in any form such as articles, paragraphs, sentences, etc., collectively referred to as text), and the output is a continuous distributed vector representation of the corresponding text, similar to word2vec The principle of word vector is similar. On the basis of retaining semantic and word order information, effective sentence or chapter vector representation can be obtained through feature learning. PV model can effectively solve the problem that word bag model does not consider word meaning and word order, and the generated vector dimension is dense. , can also effectively overcome the feature sparsity problem of short text sentence representation.
步骤4,将步骤2和步骤3得到的句子间的相似度值进行线性加权,通过网格法调节参数,找到一组最优的参数取值,输出最终的句子对之间的相似度值,计算公式如下,Step 4: Carry out linear weighting on the similarity value between the sentences obtained in step 2 and step 3, adjust the parameters through the grid method, find a set of optimal parameter values, and output the final similarity value between the sentence pairs, Calculated as follows,
sim(SA,SB)=θ*sim1(SA,SB)+(1-θ)*sim2(SA,SB) (4)sim(SA ,SB )=θ*sim1(SA ,SB )+(1-θ)*sim2(SA ,SB ) (4)
其中,SA和SB代表任意两个句子,sim(SA,SB)表示两个句子的相似度值,θ为可调参数,取值范围为[0,1],sim1(SA,SB)和sim2(SA,SB)分别按公式(2)和(3)计算得到。根据公式(4),结合公式(2)和(3),完整的句子相似度计算公式为:Among them, SA and SB represent any two sentences, sim(SA , SB ) represents the similarity value of two sentences, θ is an adjustable parameter, and the value range is [0,1], sim1(SA ,SB ) and sim2(SA ,SB ) are calculated according to formulas (2) and (3) respectively. According to formula (4), combined with formulas (2) and (3), the complete sentence similarity calculation formula is:
ω和θ都是可调参数,取值范围都是[0,1],利用网格法根据句子相似度的计算或应用结果对参数进行调优,取最优参数值。Both ω and θ are adjustable parameters, and the value range is [0,1]. The grid method is used to optimize the parameters according to the calculation or application results of sentence similarity, and the optimal parameter value is taken.
有益效果Beneficial effect
本发明的句子相似度计算方法有效地减少了语义信息的丢失,更加全面、准确地刻画了句子之间的内在联系,通过深度挖掘句子的上下文和内在语义结构特征使句子相似度计算不直接依赖于句子的表述方式,提高了计算结果的准确率。The sentence similarity calculation method of the present invention effectively reduces the loss of semantic information, more comprehensively and accurately depicts the internal relationship between sentences, and makes the sentence similarity calculation not directly dependent on Based on the expression of the sentence, the accuracy of the calculation result is improved.
具体实施方式detailed description
为了更好的说明本发明的目的和优点,下面结合具体实例对本发明方法的实施方式做进一步详细说明。In order to better illustrate the purpose and advantages of the present invention, the implementation of the method of the present invention will be further described in detail below in conjunction with specific examples.
实验采用NLP&&CC会议2013年的面向中文微博观点要素抽取评测任务所公开的语料。从中随机挑选了5个话题,共10896个句子作为短文集,采用将句子相似度计算应用到短文本聚类并评价聚类效果的方式,对句子相似度计算的效果进行评价。对于聚类效果的评价,采用轮廓系数(Silhouette Coefficient)指标来衡量,轮廓系数这一概念最早由PeterJ.Pousseeuw在1986年提出,它结合内聚度和分离度两种因素来判断聚类效果。The experiment uses the corpus disclosed in the 2013 NLP&&CC Conference oriented Chinese microblog viewpoint element extraction evaluation task. Randomly selected 5 topics and a total of 10,896 sentences as a collection of short texts, and evaluated the effect of sentence similarity calculation by applying sentence similarity calculation to short text clustering and evaluating the clustering effect. For the evaluation of clustering effect, the index of Silhouette Coefficient is used to measure. The concept of Silhouette Coefficient was first proposed by PeterJ.Pousseeuw in 1986. It combines two factors of cohesion and separation to judge the clustering effect.
轮廓系数的计算步骤如下:The calculation steps of the silhouette coefficient are as follows:
(1)对于第i个对象,计算其到所属簇中其他对象的平均距离,记为ai。(1) For the i-th object, calculate its average distance to other objects in its cluster, denoted as ai .
(2)对于第i个对象,计算该对象到到不包含该对象的任意簇中所有对象的平均距离,找出每个簇中的最小值记为bi。(2) For the i-th object, calculate the average distance from this object to all objects in any cluster that does not contain this object, and find the minimum value in each cluster and record it as bi .
(3)对于第i个对象,轮廓系数记为si,计算方法如公式(6)所示。(3) For the i-th object, the silhouette coefficient is recorded as si , and the calculation method is shown in formula (6).
轮廓系数取值范围为[-1,1],从公式(6)可以看出,如果si<0,表明第i个对象和同一个簇内部的元素之间的平均距离小于其他的簇,聚类效果不准确。如果ai的值趋于0,或者bi足够大,那么si取值越接近于1,说明聚类后簇内数据越紧密、簇间越分离差异明显,聚类效果越好。The value range of the silhouette coefficient is [-1,1]. It can be seen from the formula (6) that if si <0, it indicates that the average distance between the i-th object and the elements inside the same cluster is smaller than other clusters, The clustering effect is not accurate. If the value of ai tends to 0, or if the value of bi is large enough, then the closer the value ofsi is to 1, the closer the data in the cluster after clustering, the more separated the clusters, the better the clustering effect.
具体实施步骤为:The specific implementation steps are:
步骤1,对社交短文本集进行分句,然后对每一条句子利用ICTCLAS2015进行分词和词性标注,根据从Internet下载的停用词表,去除文本中的停用词。Step 1. Segment the social short text set, and then use ICTCLAS2015 for word segmentation and part-of-speech tagging for each sentence, and remove the stop words in the text according to the stop word list downloaded from the Internet.
步骤2,利用CSM对短文本集中的每条句子进行句义结构分析,并利用LDA主题模型对短文本集进行分析,获取短文本的主题和词语分布,对句子进行特征丰富,并计算句子相似度。Step 2, use CSM to analyze the sentence structure of each sentence in the short text set, and use the LDA topic model to analyze the short text set, obtain the topic and word distribution of the short text, enrich the features of the sentence, and calculate the sentence similarity Spend.
步骤2.1,在步骤1的基础上,对每条句子进行句义结构分析,提取句子的话题、述题、基本项、一般项。Step 2.1, on the basis of step 1, analyze the sentence structure of each sentence, and extract the topic, statement, basic items, and general items of the sentence.
步骤2.2,利用LDA主题模型对短文本集进行分析,提取文本中的主题和主题下的词语分布,得到主题-词语矩阵。In step 2.2, the LDA topic model is used to analyze the short text set, extract the topics in the text and the word distribution under the topic, and obtain the topic-word matrix.
步骤2.3,根据话题对句子进行特征扩充,得到基于话题的句子向量。具体方法为:首先提取话题下的基本项和一般项对应的词语,然后根据步骤2.2中得到的主题-词语矩阵,比较词语在不同主题下的概率,选取概率最高的主题,将该主题下的其它词语补充到句子中,作为句子的一部分,最后使用句子的所有词语作为特征,构建特征向量表示句子,其中句中原有词语所对应的维度上的取值为词语的在句中的出现次数,而扩充的词语所对应的维度上的取值按公式(1)进行计算,In step 2.3, the feature of the sentence is expanded according to the topic, and the sentence vector based on the topic is obtained. The specific method is: first extract the words corresponding to the basic items and general items under the topic, and then compare the probabilities of the words under different topics according to the topic-word matrix obtained in step 2.2, select the topic with the highest probability, and use the topic under the topic Add other words to the sentence as part of the sentence, and finally use all the words in the sentence as features to construct a feature vector to represent the sentence, where the value of the dimension corresponding to the original word in the sentence is the number of occurrences of the word in the sentence, And the value on the dimension corresponding to the expanded word is calculated according to the formula (1),
步骤2.4,按步骤2.3的方法,根据述题对句子进行特征扩充,得到基于述题的句子向量。In step 2.4, according to the method of step 2.3, the feature of the sentence is expanded according to the question, and the sentence vector based on the question is obtained.
步骤2.5,分别基于步骤2.3和2.4得到的两种句子向量计算句子相似度,对两个相似度值进行加权,按公式(2)得到句子间的最终相似度值。In step 2.5, the sentence similarity is calculated based on the two sentence vectors obtained in steps 2.3 and 2.4 respectively, and the two similarity values are weighted, and the final similarity value between sentences is obtained according to formula (2).
步骤3,将经过步骤1预处理后的所有句子输入到PV深度学习模型,利用PV模型学习文本特征,得到句子向量,并基于该句子向量计算句子间的余弦距离作为句子间的相似度,其中PV模型里的参数均采用工具中的默认值。Step 3, input all the sentences preprocessed in step 1 into the PV deep learning model, use the PV model to learn text features, obtain sentence vectors, and calculate the cosine distance between sentences based on the sentence vectors as the similarity between sentences, where The parameters in the PV model adopt the default values in the tool.
步骤4,将步骤2和步骤3得到的句子间的相似度值进行线性加权,通过网格法调节参数ω和θ,选择最优的一组参数。Step 4: Perform linear weighting on the similarity values between sentences obtained in Step 2 and Step 3, adjust the parameters ω and θ through the grid method, and select the optimal set of parameters.
对5个话题的聚类效果,在PV模型中的向量长度size=100,窗口长度window=5,ω取0.33时,θ取0.25,轮廓系数达到最优效果0.45;当θ取0时,即只考虑基于CSM句义结构分析得到的相似度计算结果,轮廓系数达到0.42;当θ取1时,即只考虑依赖PV分析得到的句子相似度结果,轮廓系数达到0.31。实验结果表明利用CSM得到的句子向量能够包含更深层次的句子内部语义信息,PV模型使句子向量能够获取丰富的上下文信息,既考虑自身语义信息同时还包含上下文信息的句子相似度计算方法更能准确衡量句子间的相似程度。For the clustering effect of 5 topics, in the PV model, the vector length size=100, the window length window=5, when ω is 0.33, θ is 0.25, and the silhouette coefficient reaches the optimal effect of 0.45; when θ is 0, that is Only considering the similarity calculation results based on the CSM sentence structure analysis, the contour coefficient reaches 0.42; when θ is 1, that is, only considers the sentence similarity results obtained by PV analysis, the contour coefficient reaches 0.31. The experimental results show that the sentence vector obtained by using CSM can contain deeper semantic information inside the sentence, the PV model enables the sentence vector to obtain rich context information, and the sentence similarity calculation method that not only considers its own semantic information but also contains context information is more accurate measure the similarity between sentences.
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201610867254.2ACN106445920A (en) | 2016-09-29 | 2016-09-29 | Sentence similarity calculation method based on sentence meaning structure characteristics |
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201610867254.2ACN106445920A (en) | 2016-09-29 | 2016-09-29 | Sentence similarity calculation method based on sentence meaning structure characteristics |
| Publication Number | Publication Date |
|---|---|
| CN106445920Atrue CN106445920A (en) | 2017-02-22 |
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201610867254.2APendingCN106445920A (en) | 2016-09-29 | 2016-09-29 | Sentence similarity calculation method based on sentence meaning structure characteristics |
| Country | Link |
|---|---|
| CN (1) | CN106445920A (en) |
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107273474A (en)* | 2017-06-08 | 2017-10-20 | 成都数联铭品科技有限公司 | Autoabstract abstracting method and system based on latent semantic analysis |
| CN108009152A (en)* | 2017-12-04 | 2018-05-08 | 陕西识代运筹信息科技股份有限公司 | A kind of data processing method and device of the text similarity analysis based on Spark-Streaming |
| CN108287824A (en)* | 2018-03-07 | 2018-07-17 | 北京云知声信息技术有限公司 | Semantic similarity calculation method and device |
| CN108363816A (en)* | 2018-03-21 | 2018-08-03 | 北京理工大学 | Open entity relation extraction method based on sentence justice structural model |
| CN108509408A (en)* | 2017-02-27 | 2018-09-07 | 芋头科技(杭州)有限公司 | A kind of sentence similarity judgment method |
| CN109101489A (en)* | 2018-07-18 | 2018-12-28 | 武汉数博科技有限责任公司 | A kind of text automatic abstracting method, device and a kind of electronic equipment |
| CN109145299A (en)* | 2018-08-16 | 2019-01-04 | 北京金山安全软件有限公司 | Text similarity determination method, device, equipment and storage medium |
| CN109857990A (en)* | 2018-12-18 | 2019-06-07 | 重庆邮电大学 | A kind of financial class notice information abstracting method based on file structure and deep learning |
| CN110008465A (en)* | 2019-01-25 | 2019-07-12 | 网经科技(苏州)有限公司 | The measure of sentence semantics distance |
| CN110020421A (en)* | 2018-01-10 | 2019-07-16 | 北京京东尚科信息技术有限公司 | The session information method of abstracting and system of communication software, equipment and storage medium |
| CN110287291A (en)* | 2019-07-03 | 2019-09-27 | 桂林电子科技大学 | An Unsupervised Method for Sentence Off-topic Analysis of English Short Texts |
| CN110348133A (en)* | 2019-07-15 | 2019-10-18 | 西南交通大学 | A kind of bullet train three-dimensional objects structure technology effect figure building system and method |
| CN110413761A (en)* | 2019-08-06 | 2019-11-05 | 浩鲸云计算科技股份有限公司 | A kind of method that the territoriality in knowledge based library is individually talked with |
| CN110765360A (en)* | 2019-11-01 | 2020-02-07 | 新华网股份有限公司 | Text topic processing method and device, electronic equipment and computer storage medium |
| CN110895656A (en)* | 2018-09-13 | 2020-03-20 | 武汉斗鱼网络科技有限公司 | Text similarity calculation method and device, electronic equipment and storage medium |
| CN110990451A (en)* | 2019-11-15 | 2020-04-10 | 浙江大华技术股份有限公司 | Data mining method, device and equipment based on sentence embedding and storage device |
| CN110990537A (en)* | 2019-12-11 | 2020-04-10 | 中山大学 | A sentence similarity calculation method based on edge information and semantic information |
| CN111008783A (en)* | 2019-12-05 | 2020-04-14 | 浙江工业大学 | Factory processing flow recommendation method based on singular value decomposition |
| CN111078849A (en)* | 2019-12-02 | 2020-04-28 | 百度在线网络技术(北京)有限公司 | Method and apparatus for outputting information |
| CN111125301A (en)* | 2019-11-22 | 2020-05-08 | 泰康保险集团股份有限公司 | Text method and device, electronic equipment and computer readable storage medium |
| CN111209375A (en)* | 2020-01-13 | 2020-05-29 | 中国科学院信息工程研究所 | Universal clause and document matching method |
| CN111368532A (en)* | 2020-03-18 | 2020-07-03 | 昆明理工大学 | A Disambiguation Method and System for Subject Word Embedding Based on LDA |
| CN112650836A (en)* | 2020-12-28 | 2021-04-13 | 成都网安科技发展有限公司 | Text analysis method and device based on syntax structure element semantics and computing terminal |
| CN112686025A (en)* | 2021-01-27 | 2021-04-20 | 浙江工商大学 | Chinese choice question interference item generation method based on free text |
| CN113536907A (en)* | 2021-06-06 | 2021-10-22 | 南京理工大学 | A method and system for social relationship recognition based on deep supervised feature selection |
| CN114861631A (en)* | 2022-05-19 | 2022-08-05 | 昆明理工大学 | Context-based Chinese cross-language sentence embedding method |
| CN115935943A (en)* | 2022-10-28 | 2023-04-07 | 北京语言大学 | An Analysis Framework Supporting Computation of Natural Language Structures |
| CN116701574A (en)* | 2023-06-09 | 2023-09-05 | 北京海卓飞网络科技有限公司 | Text semantic similarity calculation method, device, equipment and storage medium |
| CN116756347A (en)* | 2023-08-21 | 2023-09-15 | 中国标准化研究院 | A semantic information retrieval method based on big data |
| CN119538935A (en)* | 2024-10-24 | 2025-02-28 | 中国电力科学研究院有限公司 | A text semantic representation method and device based on information compression and spatial decomposition |
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104778256A (en)* | 2015-04-20 | 2015-07-15 | 江苏科技大学 | Rapid incremental clustering method for domain question-answering system consultations |
| CN104834747A (en)* | 2015-05-25 | 2015-08-12 | 中国科学院自动化研究所 | Short text classification method based on convolution neutral network |
| CN105573985A (en)* | 2016-03-04 | 2016-05-11 | 北京理工大学 | Sentence expression method based on Chinese sentence meaning structural model and topic model |
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104778256A (en)* | 2015-04-20 | 2015-07-15 | 江苏科技大学 | Rapid incremental clustering method for domain question-answering system consultations |
| CN104834747A (en)* | 2015-05-25 | 2015-08-12 | 中国科学院自动化研究所 | Short text classification method based on convolution neutral network |
| CN105573985A (en)* | 2016-03-04 | 2016-05-11 | 北京理工大学 | Sentence expression method based on Chinese sentence meaning structural model and topic model |
| Title |
|---|
| TOMAS等: "UWB at SemEval-2016 Task 1:Semantic Textual Similarity using Lexical,Syntactic,and Semantic Information", 《PROCEEDINGS OF SEMEVAL-2016》* |
| YUHUA LI 等: "Sentence Similarity Based on Semantic Nets and Corpus Statistics", 《IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING》* |
| 林萌 等: "融合句义结构模型的微博话题摘要算法", 《浙江大学学报(工学版)》* |
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN108509408A (en)* | 2017-02-27 | 2018-09-07 | 芋头科技(杭州)有限公司 | A kind of sentence similarity judgment method |
| CN108509408B (en)* | 2017-02-27 | 2019-11-22 | 芋头科技(杭州)有限公司 | A Method for Sentence Similarity Judgment |
| CN107273474A (en)* | 2017-06-08 | 2017-10-20 | 成都数联铭品科技有限公司 | Autoabstract abstracting method and system based on latent semantic analysis |
| CN108009152A (en)* | 2017-12-04 | 2018-05-08 | 陕西识代运筹信息科技股份有限公司 | A kind of data processing method and device of the text similarity analysis based on Spark-Streaming |
| CN110020421A (en)* | 2018-01-10 | 2019-07-16 | 北京京东尚科信息技术有限公司 | The session information method of abstracting and system of communication software, equipment and storage medium |
| CN108287824A (en)* | 2018-03-07 | 2018-07-17 | 北京云知声信息技术有限公司 | Semantic similarity calculation method and device |
| CN108363816A (en)* | 2018-03-21 | 2018-08-03 | 北京理工大学 | Open entity relation extraction method based on sentence justice structural model |
| CN109101489A (en)* | 2018-07-18 | 2018-12-28 | 武汉数博科技有限责任公司 | A kind of text automatic abstracting method, device and a kind of electronic equipment |
| CN109101489B (en)* | 2018-07-18 | 2022-05-20 | 武汉数博科技有限责任公司 | Text automatic summarization method and device and electronic equipment |
| CN109145299A (en)* | 2018-08-16 | 2019-01-04 | 北京金山安全软件有限公司 | Text similarity determination method, device, equipment and storage medium |
| CN109145299B (en)* | 2018-08-16 | 2022-06-21 | 北京金山安全软件有限公司 | Text similarity determination method, device, equipment and storage medium |
| CN110895656B (en)* | 2018-09-13 | 2023-12-29 | 北京橙果转话科技有限公司 | Text similarity calculation method and device, electronic equipment and storage medium |
| CN110895656A (en)* | 2018-09-13 | 2020-03-20 | 武汉斗鱼网络科技有限公司 | Text similarity calculation method and device, electronic equipment and storage medium |
| CN109857990B (en)* | 2018-12-18 | 2022-11-25 | 重庆邮电大学 | A financial announcement information extraction method based on document structure and deep learning |
| CN109857990A (en)* | 2018-12-18 | 2019-06-07 | 重庆邮电大学 | A kind of financial class notice information abstracting method based on file structure and deep learning |
| CN110008465A (en)* | 2019-01-25 | 2019-07-12 | 网经科技(苏州)有限公司 | The measure of sentence semantics distance |
| CN110287291B (en)* | 2019-07-03 | 2021-11-02 | 桂林电子科技大学 | An unsupervised method for analyzing off-topic sentences in short English texts |
| CN110287291A (en)* | 2019-07-03 | 2019-09-27 | 桂林电子科技大学 | An Unsupervised Method for Sentence Off-topic Analysis of English Short Texts |
| CN110348133A (en)* | 2019-07-15 | 2019-10-18 | 西南交通大学 | A kind of bullet train three-dimensional objects structure technology effect figure building system and method |
| CN110348133B (en)* | 2019-07-15 | 2022-08-19 | 西南交通大学 | System and method for constructing high-speed train three-dimensional product structure technical effect diagram |
| CN110413761A (en)* | 2019-08-06 | 2019-11-05 | 浩鲸云计算科技股份有限公司 | A kind of method that the territoriality in knowledge based library is individually talked with |
| CN110765360A (en)* | 2019-11-01 | 2020-02-07 | 新华网股份有限公司 | Text topic processing method and device, electronic equipment and computer storage medium |
| CN110765360B (en)* | 2019-11-01 | 2022-08-02 | 新华网股份有限公司 | Text topic processing method and device, electronic equipment and computer storage medium |
| CN110990451A (en)* | 2019-11-15 | 2020-04-10 | 浙江大华技术股份有限公司 | Data mining method, device and equipment based on sentence embedding and storage device |
| CN110990451B (en)* | 2019-11-15 | 2023-05-12 | 浙江大华技术股份有限公司 | Sentence embedding-based data mining method, device, equipment and storage device |
| CN111125301A (en)* | 2019-11-22 | 2020-05-08 | 泰康保险集团股份有限公司 | Text method and device, electronic equipment and computer readable storage medium |
| CN111078849A (en)* | 2019-12-02 | 2020-04-28 | 百度在线网络技术(北京)有限公司 | Method and apparatus for outputting information |
| CN111008783B (en)* | 2019-12-05 | 2022-03-18 | 浙江工业大学 | Factory processing flow recommendation method based on singular value decomposition |
| CN111008783A (en)* | 2019-12-05 | 2020-04-14 | 浙江工业大学 | Factory processing flow recommendation method based on singular value decomposition |
| CN110990537A (en)* | 2019-12-11 | 2020-04-10 | 中山大学 | A sentence similarity calculation method based on edge information and semantic information |
| CN110990537B (en)* | 2019-12-11 | 2023-06-27 | 中山大学 | Sentence similarity calculation method based on edge information and semantic information |
| CN111209375B (en)* | 2020-01-13 | 2023-01-17 | 中国科学院信息工程研究所 | A General Method for Matching Items and Documents |
| CN111209375A (en)* | 2020-01-13 | 2020-05-29 | 中国科学院信息工程研究所 | Universal clause and document matching method |
| CN111368532A (en)* | 2020-03-18 | 2020-07-03 | 昆明理工大学 | A Disambiguation Method and System for Subject Word Embedding Based on LDA |
| CN111368532B (en)* | 2020-03-18 | 2022-12-09 | 昆明理工大学 | Topic word embedding disambiguation method and system based on LDA |
| CN112650836A (en)* | 2020-12-28 | 2021-04-13 | 成都网安科技发展有限公司 | Text analysis method and device based on syntax structure element semantics and computing terminal |
| CN112686025A (en)* | 2021-01-27 | 2021-04-20 | 浙江工商大学 | Chinese choice question interference item generation method based on free text |
| CN112686025B (en)* | 2021-01-27 | 2023-09-19 | 浙江工商大学 | Chinese choice question interference item generation method based on free text |
| CN113536907A (en)* | 2021-06-06 | 2021-10-22 | 南京理工大学 | A method and system for social relationship recognition based on deep supervised feature selection |
| CN113536907B (en)* | 2021-06-06 | 2024-09-06 | 南京理工大学 | Social relationship identification method and system based on depth supervised feature selection |
| CN114861631A (en)* | 2022-05-19 | 2022-08-05 | 昆明理工大学 | Context-based Chinese cross-language sentence embedding method |
| CN115935943A (en)* | 2022-10-28 | 2023-04-07 | 北京语言大学 | An Analysis Framework Supporting Computation of Natural Language Structures |
| CN116701574A (en)* | 2023-06-09 | 2023-09-05 | 北京海卓飞网络科技有限公司 | Text semantic similarity calculation method, device, equipment and storage medium |
| CN116756347A (en)* | 2023-08-21 | 2023-09-15 | 中国标准化研究院 | A semantic information retrieval method based on big data |
| CN116756347B (en)* | 2023-08-21 | 2023-10-27 | 中国标准化研究院 | Semantic information retrieval method based on big data |
| CN119538935A (en)* | 2024-10-24 | 2025-02-28 | 中国电力科学研究院有限公司 | A text semantic representation method and device based on information compression and spatial decomposition |
| Publication | Publication Date | Title |
|---|---|---|
| CN106445920A (en) | Sentence similarity calculation method based on sentence meaning structure characteristics | |
| CN108052593B (en) | A topic keyword extraction method based on topic word vector and network structure | |
| Kamkarhaghighi et al. | Content tree word embedding for document representation | |
| CN103226580B (en) | A kind of topic detection method of interaction text | |
| CN109858028A (en) | A kind of short text similarity calculating method based on probabilistic model | |
| CN103324700B (en) | Noumenon concept attribute learning method based on Web information | |
| CN107315734B (en) | A kind of method and system to be standardized based on time window and semantic variant word | |
| CN106547739A (en) | A kind of text semantic similarity analysis method | |
| CN110020189A (en) | A kind of article recommended method based on Chinese Similarity measures | |
| CN103455562A (en) | Text orientation analysis method and product review orientation discriminator on basis of same | |
| CN106372061A (en) | Short text similarity calculation method based on semantics | |
| CN102866989A (en) | Viewpoint extracting method based on word dependence relationship | |
| CN107423282A (en) | Semantic Coherence Sexual Themes and the concurrent extracting method of term vector in text based on composite character | |
| CN104008090A (en) | Multi-subject extraction method based on concept vector model | |
| Sadr et al. | Unified topic-based semantic models: a study in computing the semantic relatedness of geographic terms | |
| CN103970730A (en) | Method for extracting multiple subject terms from single Chinese text | |
| CN102609427A (en) | Public opinion vertical search analysis system and method | |
| Aggarwal et al. | Wikipedia-Based Distributional Semantics for Entity Relatedness. | |
| CN111782810A (en) | Text abstract generation method based on theme enhancement | |
| CN108399238A (en) | A kind of viewpoint searching system and method for fusing text generalities and network representation | |
| Jiang et al. | Word network topic model based on Word2Vector | |
| CN105573985A (en) | Sentence expression method based on Chinese sentence meaning structural model and topic model | |
| CN113157857A (en) | Hot topic detection method, device and equipment for news | |
| Shen et al. | Emotional tendency dictionary construction for college teaching evaluation | |
| CN106776569A (en) | Tourist hot spot and its Feature Extraction Method and system in mass text |
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| WD01 | Invention patent application deemed withdrawn after publication | ||
| WD01 | Invention patent application deemed withdrawn after publication | Application publication date:20170222 |