Movatterモバイル変換


[0]ホーム

URL:


CN112860959B - Entity analysis method based on random forest improvement - Google Patents

Entity analysis method based on random forest improvement
Download PDF

Info

Publication number
CN112860959B
CN112860959BCN202110160938.XACN202110160938ACN112860959BCN 112860959 BCN112860959 BCN 112860959BCN 202110160938 ACN202110160938 ACN 202110160938ACN 112860959 BCN112860959 BCN 112860959B
Authority
CN
China
Prior art keywords
random forest
trees
decision trees
entity
method based
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110160938.XA
Other languages
Chinese (zh)
Other versions
CN112860959A (en
Inventor
巩建光
刘凌灼
黄若文
吴昊
王福焱
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tongpu Information Technology Co ltd
Original Assignee
Harbin Engineering University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Engineering UniversityfiledCriticalHarbin Engineering University
Priority to CN202110160938.XApriorityCriticalpatent/CN112860959B/en
Publication of CN112860959ApublicationCriticalpatent/CN112860959A/en
Application grantedgrantedCritical
Publication of CN112860959BpublicationCriticalpatent/CN112860959B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Classifications

Landscapes

Abstract

Translated fromChinese

本发明提供了一种基于随机森林改进的实体解析方法,包括以下步骤:S1:提供一个包括k个决策树的随机森林F,提供若干个字符串Bi;S2:执行修剪步骤包括:S2.1:从k个决策树中提取m个决策树Tm,分别使用Tm执行每一个字符串Bi,得到输出Cm;S2.2:建立集合I=C1∩C2∩...∩Cm;S3:执行验证步骤包括:S3.1:建立集合J=(C1∪C2∪...∪Cm)\(C1∩C2∩...∩Cm);S3.2:从随机森林F中提取n个决策树Rn,使用Rn执行集合J,以生成集合Kn;S4:随机森林F输出结果为I∪K1∪K2∪...∪Kn。本发明通过将执行每一个决策树分解为在修剪步骤中执行树的子集,然后在验证步骤中执行剩余的树,通过树的重用计算简化执行决策树集,以大幅缩短时间。

Figure 202110160938

The present invention provides an improved entity parsing method based on random forest, including the following steps: S1: provide a random forest F including k decision trees, and provide several character strings Bi ; S2: performing the pruning step includes: S2. 1: Extract m decision trees Tm from k decision trees, use Tm to execute each string Bi respectively, and obtain the output Cm ; S2.2: Establish a set I=C1 ∩ C2 ∩... ∩Cm ; S3: Execute the verification step including: S3.1: Establish a set J=(C1 ∪C2 ∪...∪Cm )\(C1 ∩C2 ∩...∩Cm ); S3 .2: Extract n decision trees Rn from random forest F, use Rn to execute set J to generate set Kn ; S4: Random forest F output result is I∪K1 ∪K2 ∪...∪Kn . The present invention greatly shortens the time by decomposing each decision tree for execution into a subset of trees that are executed in the pruning step, and then executing the remaining trees in the verification step, and simplifying the execution of the decision tree set by reusing the trees.

Figure 202110160938

Description

Entity analysis method based on random forest improvement
Technical Field
The invention relates to the technical field of data processing, in particular to an entity analysis method based on random forest improvement.
Background
In a dataset, objects in the real world, to which data is directed, are generally referred to as entities. There may be many different representations or descriptions of the same entity in different or even the same data set, and when data sets from different sources are combined for analysis, the descriptions of the same entity may be mixed together to cause some degree of duplication. Entity resolution is the process of identifying and linking multiple different descriptions in a data set, and determining which descriptions map to the same entity in the real world. Entity analysis is an important step in the data preprocessing process and is mainly used for solving the quality problems of repeated redundancy and the like of data.
The current entity analysis means that different data may have different descriptions (the descriptions include data formats, representation methods, and the like) for the same thing, that is, an entity, but they may often have errors such as typesetting or wrongly-written characters in the description storage process, which increases the time for data processing analysis and easily causes that matching redundancy cannot accurately obtain a data set that we want.
Disclosure of Invention
The invention aims to provide an entity analysis method based on random forest improvement, which can perform similarity connection on matching of a character string and an entity through random forest, improve accuracy and efficiency of matching a data set and overcome the defects of the existing entity analysis technology.
The invention provides an entity analysis method based on random forest improvement, which comprises the following steps:
s1: providing a random forest F comprising k decision trees, wherein k is 1, 2.. N; providing a plurality of character strings BiWherein i is 1, 2.. N; and performing the following training steps:
s1.1: given a number of sample data tables Ai, where i ═ 1, 2.. N;
s1.2: randomly selecting a group of Xp tuples from an Ap table, randomly selecting a group of Xq tuples from an Aq table that may match the Ap table, pairing the Xp with the Xq to form a sample S, wherein: p belongs to i, q belongs to i, and p is not equal to q;
s1.3: examining patterns of the Ap table and the Aq table, creating a set of characteristics, and converting tuple pairs in the sample S into feature vectors by using the characteristics;
s1.4: training the random forest F by using the feature vectors in S1.3;
s2: performing a trimming step, the trimming step comprising:
s2.1: extracting m decision trees T from the k decision trees1,T2...TmRespectively using said T1,T2...TmExecuting each of the character strings BiTo obtain an output C1,C2...CmM is the minimum number of decision trees required for correct analysis, and the correct analysis is that the random forest F uses the character string BiCorrectly resolving into an entity;
s2.2: establishing a set I ═ C1∩C2∩...∩Cm
S3: performing a verification step, the verification step comprising:
s3.1: set up J ═ C1∪C2∪...∪Cm)\(C1∩C2∩...∩Cm);
S3.2: extracting n decision trees R from the random forest F1,R2...RnUsing said R1,R2...RnExecuting the set J to generate a set K1,K2...KnAnd wherein
Figure GDA0003276480150000021
S4: the random forest F outputs an entity analysis result as I U K1∪K2∪...∪Kn
Further, in S3.2, (R)1,R2...Rn)∪(T1,T2...Tm) A random forest F.
Further, in S1.2, in the AqConstructing a reverse index in a table, and quickly searching the A by using the reverse indexqIn table with said XpTuples with tuples sharing X symbols, constituting XqTuple, where x ≧ 2.
Further, in S2, the k decision trees are pruned before execution.
The invention reduces the number of candidate pairs to be matched, reduces the sample, trains the random forest through the reduced sample to shorten the time, decomposes each executed decision tree into subsets of executed trees in the pruning step, executes the rest trees in the verification step, and simplifies the executed decision tree set through the reuse calculation of the trees to shorten the time again, simplify the data and accurately analyze the result.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and other drawings can be obtained by those skilled in the art without creative efforts.
FIG. 1 is a schematic diagram of an entity resolution process according to the present invention;
FIG. 2 is a schematic diagram of a conventional entity resolution process;
Detailed Description
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application. As used herein, the singular forms also include the plural forms unless the context clearly dictates otherwise, and further, it is understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of the stated features, steps, operations, devices, components, and/or combinations thereof.
The technical solutions of the present invention will be described clearly and completely with reference to the following embodiments, and it should be understood that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example 1:
the invention provides an entity analysis method based on random forest improvement, which comprises the following steps:
s1: providing a random forest F comprising k decision trees, wherein k is 1, 2.. N; providing a plurality of character strings BiWherein i is 1, 2.. N;
training the random forest F;
s1.1: given a number of sample data tables AiWherein i is 1, 2.. N;
s1.2: from ApRandomly selecting a set of X in the tablepTuple, from AqRandom selection of and A in the tablepSet of possible table matchesqTuple, XpAnd XqThe pairings constitute a sample S, where: p belongs to i, q belongs to i, and p is not equal to q; in S1.2, in AqConstructing an inverse index in the table, and quickly searching A by using the inverse indexqIn table with XpTuples with tuples sharing X symbols, constituting XqTuple, where x ≧ 2.
S1.3: examination ApWatch and AqA table pattern, creating a set of characteristics, and converting tuple pairs in the sample S into feature vectors by using the characteristics;
s1.4: the random forest F is trained using the feature vectors in S1.3.
For example, in S1.1, two sample data tables A are given1And A2In S1.2, from A1Randomly selecting a set of X in the table1Tuple, from A2Random selection of and A in the table1Set of possible matches of X in the table2Tuple, X1And X2The pairings constitute sample S. S1.3, by checking A1Watch and A2Schema of a table, create a set of properties, the way to create the properties, for example, if it is detected that property city belongs to string type, then many properties using string similarity measures are created, for example: edit dist (A)1.city,A2.city),jaccard 2gram(A1.city,A2.city)。
S2: performing a trimming step, the trimming step comprising:
s2.1: extracting m decision trees T from k decision trees1,T2...TmUsing T separately1,T2...TmExecuting each character string BiTo obtain an output C1,C2...Cm
S2.2: establishing a set I ═ C1∩C2∩...∩Cm
S3: performing a verification step, the verification step comprising:
s3.1: set up J ═ C1∪C2∪...∪Cm)\(C1∩C2∩...∩Cm);
S3.2: extracting n decision trees R from random forest F1,R2...RnUsing R1,R2...RnExecuting the set J to generate a set K1,K2...KnAnd wherein
Figure GDA0003276480150000051
S4: the random forest F outputs an entity analysis result as I U K1∪K2∪...∪Kn
As shown in fig. 1: for example, in S1, the random forest F includes 3 decision trees T1、T2And R1And in S2, two character strings B are provided1And B2In B1And B2Upper execution of T1、T2To obtain an output C1And C2Then, in S3, set I ═ C is established1∩C2Then I is represented by T1And T2Predicting all pair components of a match, which may be part of the random forest F output, then in S4, set J ═ C is established1∪C2)\(C1∩C2) Then, in S5, useR1Set J is executed to produce set K, so obviously K also matches random forest F, so the output of random forest F is IuU K, and any other pair (neither in I nor J) cannot be T1And T2A match is predicted and therefore cannot be matched to the random forest F.
In S2, since the set J is often small, the tree T is divided3Application to J is often more than application to the original string B1And B2The aggregation is much faster. When the F is large, such as 10 trees, the time saved is considerable. Assuming that in this case we need at least 5 trees to match F, then we can match B1And B2Apply 6 trees to get sets I and J, and then apply the remaining 4 trees to the relatively smaller set J, the former being pruning of the trees and the latter being verification of the trees.
The method specifically comprises the following steps: consider a forest F consisting of 10 trees, of which at least 5 must match to match the F. The pruning step then executes 6 trees to produce a set of J, taking into account that a pair of p1 ∈ J matches 4 trees during pruning. Then, as soon as one of the remaining 4 trees matches p1, we can declare p1 as a match; considering that a pair of p2 ∈ J only matches one tree at pruning, we can declare a p2 mismatch when one of the predicted p2 of the remaining 4 trees does not match.
Therefore, at S2In m decision trees T1,T2...TmThe execution process is a trimming process; in S5, R1,R2...RmThe execution process of (2) is a verification process, and the verification process can simplify the time for data processing and analysis.
In S3.2, (R)1,R2...Rn)∪(T1,T2...Tm) A random forest F. And all the trees in the random forest F are fully utilized, so that the analysis result is more accurate.
In S2.1, m is the minimum number of decision trees required by correct analysis, and the character string B is obtained by the random forest F through the correct analysisiCorrectly resolved into entities. When used in the trimming stepThe minimum number m of decision trees for correct parsing can be guaranteed, and the remaining number n of decision trees is used in the verification step, so that the obvious effect of shortening the parsing time is the maximum, because the pruning step needs to execute all the character strings, and the verification step only needs to execute the pruned set J.
In S2, k decision trees are pruned before execution. A subset of the tree is applied for pruning, which is done to avoid overfitting, and then verified with the remaining tree application J.
Example 2:
in this example, consider matching two sets of names, both long (e.g., Graphene Nanospheres) and short (e.g., Golf Ball). Two sets of strings B by performing pruning and validation of random forest F1And B2And matching to obtain the desired character string.
Figure GDA0003276480150000061
Figure GDA0003276480150000062
Figure DA00032764801540680861
To match two sets of character strings B1And B2We learn the random forest F, then at B1And B2F is performed. The execution process of the invention is that the execution of F is divided into two steps: pruning and verifying. The following describes how to perform these two steps efficiently.
Assume that the random forest F has k trees, of which at least a dk/2e tree is needed for F matching. It is clear that when the pruning step is performed for at least (bk/2c +1) trees, any string pairs not output by this step cannot be matched.
As shown in FIG. 2, assume that the random forest F has at least three treesTwo trees output the same pair before outputting a pair (i.e., declaring it matching). Simply put, we can be in two sets of strings B1And B2Is to execute F by executing B1And B2Each tree T of (1)iTo obtain an output CiThen all pairs appearing in the output of at least two trees are output (see fig. 2), which is a time consuming approach.
As shown in FIG. 1, the present invention has the advantage of only performing two trees, such as at B1And B2Upper execution of T1、T2To obtain an output C1And C2(see FIG. 1). Set I ═ C1∩C2From T1、T2All pairs of the predicted matches are composed and therefore can be immediately output as part of the random forest F output.
The implementation process of the trimming step is as follows: set J ═ C1∪C2)\(C1∩C2) Composed of only one tree (T)1Or T2) All pairs of matching are predicted to constitute. It can be easily seen that we only need to leave the remaining trees R1Applied to the set J, let K be R in J1The set of pairs that are predicted to match, it is clear that any such pair is also a match of a random forest F, since it consists of exactly two trees (T)1Or T2And R1) And (6) matching. Thus, the output of random forest F is I @ (see FIG. 1). None of the other pairings (i.e. pairings in I and J) can be represented by T1And T2A match is predicted and therefore cannot be matched with F.
The implementation process of the verification step is as follows: the set J tends to be relatively small, and therefore the tree R1Application to J tends to be more than application to string B1And B2Much faster, when F is large (e.g., 10 trees), this time saving is very significant; suppose in this case we need at least five trees to match F. We can then apply six trees to B1And B2To obtain sets I and J (i.e., a pruning step), and then apply the remaining four trees to the relatively smaller set J (i.e., a verification step).
In the step of verification, the verification step,suppose that the pruning T is performed in a preceding pruning stepmTrees produce a set of pairs J, then it is necessary to consider how to execute the remaining trees on J: let U be the set of remaining trees that are executed on set J, similar to the way trees are executed in the pruning step, where the above optimization procedure can simply be used to generate a plan P that executes all the trees in U in a combined manner (i.e. reuse calculations), but a better solution is to apply the trees in order to avoid applying all the trees in U to all pairs in J.
In summary, the invention trains a random forest comprising k decision trees, inputs character strings in a reference set, executes the decision trees in the random forest to construct a new set I, J, constructs the output of the random forest by using the set I, J, completes the matching of the reference set and the real entity, and can verify the matching by using the constructed set J in the execution process of the random forest.
Finally, it should be noted that: the above embodiments are only used to illustrate the technical solution of the present invention, and not to limit the same; while the invention has been described in detail and with reference to the foregoing embodiments, it will be understood by those skilled in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present invention.

Claims (4)

Translated fromChinese
1.一种基于随机森林改进的实体解析方法,其特征在于,包括以下步骤:1. an improved entity parsing method based on random forest, is characterized in that, comprises the following steps:S1:提供一个包括k个决策树的随机森林F,其中k=1,2...N;提供若干个字符串Bi,其中i=1,2...N;并执行如下训练步骤:S1: Provide a random forest F including k decision trees, where k=1, 2...N; provide several strings Bi , where i=1, 2...N; and perform the following training steps:S1.1:给定若干个样本数据表Ai,其中i=1,2...N;S1.1: Given several sample data tables Ai, where i=1, 2...N;S1.2:从Ap表中随机选择一组Xp元组,从Aq表中随机选择与所述Ap表可能匹配的一组Xq元组,将所述Xp与所述Xq配对组成样本S,其中:p∈i,q∈i,p≠q;S1.2: Randomly select a set of Xp tuples from the Ap table, randomly select a set of Xq tuples that may match the Ap table from the Aq table, and pair the Xp and the Xq to form a sample S, wherein : p∈i, q∈i, p≠q;S1.3:检查所述Ap表与所述Aq表的模式,创建一组特性,使用所述特性将所述样本S中的元组对转换为特征向量;S1.3: Check the schema of the Ap table and the Aq table, create a set of characteristics, and use the characteristics to convert the tuple pairs in the sample S into feature vectors;S1.4:使用S1.3中的所述特征向量训练所述随机森林F;S1.4: Use the feature vector in S1.3 to train the random forest F;S2:执行修剪步骤,所述修剪步骤包括:S2: perform a trimming step, the trimming step includes:S2.1:从所述k个决策树中提取m个决策树T1,T2...Tm,分别使用所述T1,T2...Tm执行每一个所述字符串Bi,得到输出C1,C2...Cm,所述m为正确解析所需要的最小的决策树数量,所述正确解析为所述随机森林F将所述字符串Bi正确解析为实体;S2.1: Extract m decision trees T1 , T2 . . . Tm from the k decision trees, and use the T1 , T2 . . . Tm to execute each of the character strings B respectivelyi , the outputs C1 , C2 . . . Cm are obtained, where m is the minimum number of decision trees required for correct parsing. The correct parsing is that the random forest F correctly parses the string Bi as entity;S2.2:建立集合I=C1∩C2∩...∩CmS2.2: Create a set I=C1 ∩C2 ∩...∩Cm ;S3:执行验证步骤,所述验证步骤包括:S3: Execute a verification step, the verification step includes:S3.1:建立集合J=(C1∪C2∪...∪Cm)\(C1∩C2∩...∩Cm);S3.1: Establish set J=(C1 ∪C2 ∪...∪Cm )\(C1 ∩C2 ∩...∩Cm );S3.2:从所述随机森林F中提取n个决策树R1,R2...Rn,使用所述R1,R2...Rn执行所述集合J,以生成集合K1,K2...Kn,且其中S3.2: Extract n decision trees R1 , R2 . . . Rn from the random forest F, and execute the set J using the R1 , R2 . . . Rn to generate a set K1 , K2 ...Kn , and where
Figure FDA0003276480140000011
Figure FDA0003276480140000011
S4:所述随机森林F输出实体解析结果为I∪K1∪K2∪...∪KnS4: The random forest F outputs the entity analysis result as I∪K1 ∪K2 ∪...∪Kn .2.根据权利要求1所述的一种基于随机森林改进的实体解析方法,其特征在于,S3.2中,(R1,R2...Rn)∪(T1,T2...Tm)=随机森林F。2. An improved entity parsing method based on random forest according to claim 1, characterized in that, in S3.2, (R1 , R2 . . . Rn )∪(T1 , T2 .. .Tm ) = random forest F.3.根据权利要求1所述的一种基于随机森林改进的实体解析方法,其特征在于,S1.2中,在所述Aq表中构建反向索引,使用所述反向索引快速查找所述Aq表中与所述Xp元组共享x个符号的元组,组成Xq元组,其中x≥2。3. The improved entity parsing method based on random forest according to claim 1, wherein in S1.2, an inverted index is constructed in theAq table, and the inverted index is used to quickly find all the objects. The tuples in the Aq table that share x symbols with the Xp tuples form Xq tuples, where x≥2.4.根据权利要求2所述的一种基于随机森林改进的实体解析方法,其特征在于,S2中,在执行前对所述k个决策树进行修剪。4 . The improved entity parsing method based on random forest according to claim 2 , wherein, in S2 , the k decision trees are pruned before execution. 5 .
CN202110160938.XA2021-02-052021-02-05Entity analysis method based on random forest improvementActiveCN112860959B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202110160938.XACN112860959B (en)2021-02-052021-02-05Entity analysis method based on random forest improvement

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202110160938.XACN112860959B (en)2021-02-052021-02-05Entity analysis method based on random forest improvement

Publications (2)

Publication NumberPublication Date
CN112860959A CN112860959A (en)2021-05-28
CN112860959Btrue CN112860959B (en)2021-11-05

Family

ID=75988512

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202110160938.XAActiveCN112860959B (en)2021-02-052021-02-05Entity analysis method based on random forest improvement

Country Status (1)

CountryLink
CN (1)CN112860959B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN105844300A (en)*2016-03-242016-08-10河南师范大学Optimized classification method and optimized classification device based on random forest algorithm
CN107657063A (en)*2017-10-302018-02-02合肥工业大学The construction method and device of medical knowledge collection of illustrative plates

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US9904916B2 (en)*2015-07-012018-02-27Klarna AbIncremental login and authentication to user portal without username/password
US10162967B1 (en)*2016-08-172018-12-25Trend Micro IncorporatedMethods and systems for identifying legitimate computer files
RU2637992C1 (en)*2016-08-252017-12-08Общество с ограниченной ответственностью "Аби Продакшн"Method of extracting facts from texts on natural language
CN107403067A (en)*2017-07-312017-11-28京东方科技集团股份有限公司Intelligence based on medical knowledge base point examines server, terminal and system
CN108959577B (en)*2018-07-062021-12-07中国民航大学Entity matching method and computer program based on non-dominant attribute outlier detection
CN109829471B (en)*2018-12-192021-10-15东软集团股份有限公司Training method and device for random forest, storage medium and electronic equipment
CN112699793A (en)*2020-12-292021-04-23长安大学Fatigue driving detection optimization identification method based on random forest

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN105844300A (en)*2016-03-242016-08-10河南师范大学Optimized classification method and optimized classification device based on random forest algorithm
CN107657063A (en)*2017-10-302018-02-02合肥工业大学The construction method and device of medical knowledge collection of illustrative plates

Also Published As

Publication numberPublication date
CN112860959A (en)2021-05-28

Similar Documents

PublicationPublication DateTitle
CN109829155B (en)Keyword determination method, automatic scoring method, device, equipment and medium
US20230205610A1 (en)Systems and methods for removing identifiable information
CN108038234B (en) A method and device for automatically generating question template
CN113947084B (en) Question-answering knowledge retrieval method, device and equipment based on graph embedding
CN112257446B (en)Named entity recognition method, named entity recognition device, named entity recognition computer equipment and named entity recognition readable storage medium
US9536444B2 (en)Evaluating expert opinions in a question and answer system
CN114386048B (en) Sorting-based open source software security vulnerability patch location method
CN111178085B (en)Text translator training method, and professional field text semantic parsing method and device
CN113010679A (en)Question and answer pair generation method, device and equipment and computer readable storage medium
CN114238645A (en) A Relation Selection Method Based on BERT Siamese Attention Network and Fusion Graph Embedding Features
CN118897886B (en)Question-answering method, system, equipment and medium based on specific domain knowledge graph
CN118761417B (en) A method for improving large model knowledge question answering using triple proofreading mechanism
CN112365139A (en)Crowd danger degree analysis method under graph convolution neural network
CN111339258A (en) Recommendation method for college computer basic exercises based on knowledge graph
CN115048490A (en)Deep learning-based cloud manufacturing service flow recommendation method
CN116432125B (en) Code Classification Method Based on Hash Algorithm
CN112860873B (en) Intelligent response method, device and storage medium
CN114969294A (en) An Extension Method for Proximity Sensitive Words
CN115221292A (en) A generative knowledge question answering method and device
CN116610815A (en) A relationship prediction method, device, equipment and storage medium for knowledge graph
CN119961628A (en) Model hallucination detection method and device, storage medium and electronic device
CN112860959B (en)Entity analysis method based on random forest improvement
JagadambaOnline subjective answer verifying system using artificial intelligence
CN114780589A (en) Multi-table connection query method, device, device and storage medium
CN118779414A (en) A test question intelligent generation method and system based on large language model

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant
TR01Transfer of patent right

Effective date of registration:20220606

Address after:No. 50, Daqing sub Road, Xiangfang District, Harbin, Heilongjiang 150046

Patentee after:Tongpu Information Technology Co.,Ltd.

Address before:150001 Harbin Engineering University, 145 Nantong street, Harbin City, Heilongjiang Province

Patentee before:HARBIN ENGINEERING University

TR01Transfer of patent right
PE01Entry into force of the registration of the contract for pledge of patent right

Denomination of invention:An improved entity analysis method based on random forest

Effective date of registration:20230112

Granted publication date:20211105

Pledgee:Heilongjiang Xinzheng financing guarantee Group Co.,Ltd.

Pledgor:Tongpu Information Technology Co.,Ltd.

Registration number:Y2023230000015

PE01Entry into force of the registration of the contract for pledge of patent right

[8]ページ先頭

©2009-2025 Movatter.jp