Movatterモバイル変換


[0]ホーム

URL:


CN111496784B - A space environment recognition method and system for robot intelligent service - Google Patents

A space environment recognition method and system for robot intelligent service
Download PDF

Info

Publication number
CN111496784B
CN111496784BCN202010228789.1ACN202010228789ACN111496784BCN 111496784 BCN111496784 BCN 111496784BCN 202010228789 ACN202010228789 ACN 202010228789ACN 111496784 BCN111496784 BCN 111496784B
Authority
CN
China
Prior art keywords
article
relationship
image
visual
spatial
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010228789.1A
Other languages
Chinese (zh)
Other versions
CN111496784A (en
Inventor
吴皓
马庆
焦梦林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong University
Original Assignee
Shandong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong UniversityfiledCriticalShandong University
Priority to CN202010228789.1ApriorityCriticalpatent/CN111496784B/en
Publication of CN111496784ApublicationCriticalpatent/CN111496784A/en
Application grantedgrantedCritical
Publication of CN111496784BpublicationCriticalpatent/CN111496784B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Classifications

Landscapes

Abstract

Translated fromChinese

本公开提供了一种用于机器人智能化服务的空间环境识别方法及系统,属于机器人技术领域,获取待识别空间环境区域的至少一幅图像;根据获取的图像,采用预设的视觉关系获取模型,得到图像中物品的视觉关系三元组;根据获得的物品特征,利用预设属性获取模型,得到各个物品的属性;根据得到的各个物品的视觉关系三元组和各个物品的属性,得到当前图像内的物品多关系图;对待识别环境区域的每幅图像进行处理得到物品多关系图,根据得到的每幅图像的物品多关系图构建机器人语义视觉空间,实现对空间环境的识别;本公开通过构建关系三元组,得到了空间中各个物品之间以及物品和属性之间的关系,极大的提高了对空间的环境识别能力。

Figure 202010228789

The present disclosure provides a space environment recognition method and system for robot intelligent service, belonging to the field of robotics technology. At least one image of a space environment area to be recognized is obtained; according to the obtained image, a preset visual relationship is used to obtain a model. , obtain the visual relationship triplet of the items in the image; according to the obtained item features, use the preset attribute to obtain the model to obtain the attributes of each item; according to the obtained visual relationship triplet of each item and the attributes of each item, get the current A multi-relationship diagram of items in an image; processing each image of an environment area to be recognized to obtain a multi-relationship diagram of items, and constructing a robot semantic visual space according to the multi-relationship diagram of items of each image obtained to realize the recognition of the space environment; the present disclosure By constructing relation triples, the relationships between items in the space and between items and attributes are obtained, which greatly improves the ability to recognize the environment in the space.

Figure 202010228789

Description

Space environment identification method and system for robot intelligent service
Technical Field
The disclosure relates to the technical field of robots, in particular to a space environment identification method and system for robot intelligent service.
Background
The statements in this section merely provide background information related to the present disclosure and may not necessarily constitute prior art.
When the robot is in intelligent service, the environment needs to be understood at a semantic level, particularly, the environment can be mapped to a semantic space in the aspect of vision, the cognitive intelligence of the robot is enhanced, and tasks are better analyzed.
The inventor of the present disclosure finds that (1) most of the existing space identification methods simply identify images without considering the attributes of the articles in the images, which results in inaccurate space identification and can not define the relationship of each article in the space well; (2) the existing recognition algorithm mainly adopts a neural network algorithm, realizes the feature extraction of an image so as to realize the recognition of articles in the image, has less research on the relative position relationship in the specific image, and cannot effectively express the relative position relationship among the articles in the space.
Disclosure of Invention
In order to solve the defects of the prior art, the space environment identification method and system for robot intelligent service are provided by the disclosure, the relation among all articles in the space and the relation among the articles and the attributes are obtained by constructing the relation triple, and the environment identification capability of the space is greatly improved.
In order to achieve the purpose, the following technical scheme is adopted in the disclosure:
the first aspect of the disclosure provides a space environment identification method for robot intelligent service.
A space environment identification method for robot intelligent service comprises the following steps:
acquiring at least one image of a spatial environment region to be identified;
acquiring a model by adopting a preset visual relationship according to the acquired image, acquiring article characteristics in the image and relationship characteristics among articles, and further acquiring a visual relationship triple of the articles in the image;
obtaining a model by using a preset attribute according to the obtained article characteristics to obtain the attribute of each article;
obtaining an article multiple relation graph in the current image according to the obtained visual relation triples of each article and the attributes of each article;
and processing each acquired image of the environment area to be identified to obtain an article multiple relation graph, and constructing a robot semantic visual space according to the article multiple relation graph of each acquired image to realize identification of the space environment.
A second aspect of the present disclosure provides a spatial environment recognition system for a robot-intelligent service.
A spatial environment recognition system for robotic intelligence services, comprising:
a data acquisition module configured to: acquiring at least one image of a spatial environment region to be identified;
a visual relationship triplet acquisition module configured to: acquiring a model by adopting a preset visual relationship according to the acquired image, acquiring article characteristics in the image and relationship characteristics among articles, and further acquiring a visual relationship triple of the articles in the image;
an item attribute acquisition module configured to: obtaining a model by using a preset attribute according to the obtained article characteristics to obtain the attribute of each article;
an item multiple relationship diagram acquisition module configured to: obtaining an article multiple relation graph in the current image according to the obtained visual relation triples of each article and the attributes of each article;
an environment identification module configured to: and processing each image of the environment area to be identified to obtain an article multiple relation graph, and constructing a robot semantic visual space according to the obtained article multiple relation graph of each image to realize the identification of the environment.
A third aspect of the present disclosure provides a medium on which a program is stored, the program implementing, when executed by a processor, the steps in the spatial environment recognition method for robot intelligence service according to the first aspect of the present disclosure.
A fourth aspect of the present disclosure provides an electronic device, including a memory, a processor, and a program stored on the memory and executable on the processor, where the processor executes the program to implement the steps in the method for identifying a spatial environment for robot intelligent services according to the first aspect of the present disclosure.
Compared with the prior art, the beneficial effect of this disclosure is:
1. according to the method, the system, the medium and the electronic equipment, the preset visual relationship obtaining model is adopted, the object features in the image and the relationship features between the objects are obtained, the triple features of the objects in the image are obtained, the visual relationship triple of each object in the image is further obtained, the identification of the relative position relationship of each object in the image is achieved, and the identification capacity of the space is improved.
2. According to the method, the system, the medium and the electronic equipment, the model is obtained by using the preset attribute according to the obtained article characteristics, the attribute of each article is obtained, the article multi-relation graph in the current image is obtained according to the obtained visual relation triple of each article and the attribute of each article, the visual relation of each article and the attribute of each article are effectively combined, and the space environment is more efficiently and accurately identified.
Drawings
Fig. 1 is a schematic diagram of a visual relationship acquisition framework provided in embodiment 1 of the present disclosure.
Fig. 2 is a schematic diagram of a BRNN framework in the semantic visual space construction provided in embodiment 1 of the present disclosure.
Fig. 3 is a multiple relation diagram in a semantic visual space provided in embodiment 1 of the present disclosure.
Detailed Description
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present disclosure. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
The embodiments and features of the embodiments in the present disclosure may be combined with each other without conflict.
Example 1:
as shown in fig. 1 to 3, embodiment 1 of the present disclosure provides a spatial environment recognition method for robot intelligent service, including the following steps:
acquiring at least one image of a spatial environment region to be identified;
according to the obtained image, a preset visual relationship obtaining model is adopted, the object features in the image and the relationship features between the objects are obtained, the triple features of the objects in the image are obtained, and then the visual relationship triple of the objects in the image is obtained;
obtaining a model by using a preset attribute according to the obtained article characteristics to obtain the attribute of each article;
obtaining an article multiple relation graph in the current image according to the obtained visual relation triples of each article and the attributes of each article;
and processing each acquired image of the environment area to be identified to obtain an article multiple relation graph, and constructing a robot semantic visual space according to the article multiple relation graph of each acquired image to realize identification of the space environment.
The visual space is constructed by the following steps:
firstly, the object relation in the environment is mapped to a semantic level, and the cognition of the robot to the service environment space relation is enhanced. Next, in order to further understand the intrinsic logical relationship of the item, a method of obtaining the item attribute from the network is given. And finally, storing the obtained environmental knowledge in a form of a multi-relation graph to complete the construction of the visual space.
The method specifically comprises the following steps:
(1) acquisition of visual relationships
In the semantic visual space construction process, firstly, the object visual relationships in the scene, such as "Computer on task", are obtained, and these relationships generally express < subject, predictor, object > in the form of triples. The visual relation realizes higher-level understanding of semantic level and is an important way for realizing the increase from perception intelligence to cognitive intelligence. In the construction work of the semantic visual space, the visual relation directly provides the leading knowledge in the environment. Fig. 1 provides a method for acquiring a visual relationship, which combines an image and semantic information to acquire a visual relationship.
In the method, firstly, a fast RCNN (Faster convolutional neural network based on regions) technology is used for target detection, and the semantics of a target object and the coordinate information of a detection frame of the target object are obtained. And performing feature coding on the semantics of the target object, and acquiring the relationship features between the objects according to the detection frame information to complete the construction of the triple features. Inputting the triple characteristics into a Bi-RNN (bidirectional recurrent neural network) network to construct a relation triple. In particular, considering that the existence of the dynamic object can influence the robustness of the constructed visual relationship, in the framework, the dynamic article filtering module is used at the same time, so that the influence of the dynamic object on the visual relationship library is reduced.
(1-1) triple feature construction
The triple features include word vectors for constructing entity targets, and entity relationship coding. A BERT (Bidirectional Encoder representation of transducers) model is adopted for constructing word vectors of an entity target, semantic labels in a data set are used as training data and input into the BERT model, fine adjustment of the model is carried out, parameters obtained by the model are used as characteristics, and the characteristics are the word vectors.
When obtaining the relational features, the spatial relationships between the targets are utilized to model the relational features. Relationship between objects uses [ x ]1,x2,x3]Denotes x1Represents object 1, x3Represents object 3, x2Representing a relationship. The spatial relationship s is represented as follows:
Figure BDA0002428639560000061
in the formula 1, the first and second groups of the compound,
Figure BDA0002428639560000062
a prediction block (i ═ 1,3) representing a target object;
Figure BDA0002428639560000063
refers to the center coordinates, w, of the prediction boxiAnd hiRepresenting the width and height of the prediction box, respectively. W and H represent the width and height of the intersection of the target 1 and target 3 prediction boxes;
Figure BDA0002428639560000064
are codes for relevant spatial relationships, and the codes are particularly important for representing complex spatial relationships. After the spatial relation s is obtained, the s is input into a multilayer perceptron (MLP) to obtain a feature expression x with the same dimension as the word vector2And thus obtain the feature expression of the triplet.
(1-2) construction of visual relationship
And after the triple features are obtained, inputting the features into the Bi-RNN for visual relation prediction. The BRNN (bidirectional recurrent neural network) is an efficient natural language processing model, which is formed by superposing two RNN networks up and down, and the output is determined by the states of the two RNN (convolutional neural networks). In the visual relationship acquisition model, the output of the Bi-RNN is designed to be one rather than multiple outputs, and the input is a feature of a triplet, as shown in fig. 2.
In the framework shown in fig. 2, forward propagation is included
Figure BDA0002428639560000065
And backward propagation
Figure BDA0002428639560000066
The framework is mainly divided into an input layer, a hidden layer and an output layer. Wherein the formulas for the hidden layer and the output layer are defined as follows:
Figure BDA0002428639560000067
Figure BDA0002428639560000068
Figure BDA0002428639560000069
in the context of these formulas, the expression,
Figure BDA0002428639560000071
and
Figure BDA0002428639560000072
the parameter matrixes respectively represent an input layer-hidden layer and a hidden layer-hidden layer in the forward propagation process.
Figure BDA0002428639560000073
And
Figure BDA0002428639560000074
respectively, the hidden layer-output layer parameter matrix and the offset vector in the forward propagation process. The meaning of the parameters in the back propagation process can be obtained in the same way. x is the number oftRepresenting the input, y the output, f the activation function, byIs an offset value.
From the framework shown in FIG. 2 and equation (4), the predicted relationship output y can be obtained:
Figure BDA0002428639560000075
wherein h is1,h2,h3The values in the forward propagation and the backward propagation can be obtained by formula (2) and formula (3), and y represents the predicted relationship, such as "while", "on", etc. And after the word vectors and the relation characteristics are output through Bi-RNN training, the visual relation triple can be obtained.
(1-3) dynamic object Filtering
In the process of acquiring the visual relationship, the service robot pays attention to the relationship among the articles in the environment, and the relationship among the articles is always kept unchanged for a certain time. However, the existence of dynamic objects in the environment directly affects the timeliness of the visual relationship. For example, when detecting visual relationships, relationships between people and items are often extracted, and these relationships may change rapidly. If the relation is added into the relation library, inconsistency with reality is generated when the visual relation is searched in subsequent services, meanwhile, the redundancy degree of the relation library is increased, and the service efficiency of the robot is reduced. Therefore, the obtained triple relationship containing the dynamic object needs to be filtered.
The dynamic filtering module mainly provides a dynamic object list, and when detecting that the visual relation triplets contain dynamic objects, the triplets are filtered. The objects in the list may be added according to common sense, e.g. "people". The objects can also be added according to experience in actual operation, and through multiple operation observation, if the visual relationship of a certain type of objects changes rapidly, the objects are added. Thereby finally completing the acquisition of the visual relationship.
(2) Attribute knowledge acquisition
When the robot executes the intelligent service, not only the physical relationship between the objects needs to be understood, but also the deep-level attributes of the objects need to be further understood, so that the relationship of the objects on the logical level can be better understood. Therefore, when building the semantic visual space, the attributes of the item also need to be added. Due to the diversity of items and related attributes in the environment, it is desirable to obtain the related attributes of the items from the network.
In the process of acquiring the related attributes of the article, firstly, the article semantics are used as key words to search related information on the network, and after the search is carried out, a series of result character strings are obtained. Attributes of the item are extracted from the result strings.
And extracting the attribute of the article from the obtained result character string, and identifying the attribute key words in the result by adopting an NER (named entity identification) model. In the process of entity identification, a BERT-CRF method is adopted to carry out named entity identification. In the method, text information is extracted by using a BERT model, characteristic coding is carried out, the coding is input into a CRF layer to carry out information decoding, and a labeling sequence is obtained. CRF refers to conditional random fields, a discriminative probabilistic undirected graph model for labeling and partitioning sequence data. BERT-CRF is an improvement on Bi-LSTM-CRF, and a BERT model is used for replacing Bi-LSTM, and the BERT model and the Bi-LSTM have similar training principles.
Input sequence X ═ X1,X2,…,Xn) Obtaining a predicted tag sequence y (y) through output of a BERT-CRF model1,y2,..,yn)。
The scores defining the predicted sequences are as follows:
Figure BDA0002428639560000081
wherein
Figure BDA0002428639560000082
Indicating that the output result at the ith position is yiThe probability of (c).
Figure BDA0002428639560000083
Representing the transition probability from state i to i + 1. The scoring function takes the previous states into account and the obtained result is more consistent with the actual output result.
During training, for each sample X, all possible sequences y are scored as s (X, y), and the total score is normalized:
Figure BDA0002428639560000084
in equation 7, Y represents the correct annotation sequence and Y represents all possible annotation sequences. The loss function can be obtained by equation 7, as follows:
Figure BDA0002428639560000091
through continuous iterative training of the loss function in the model, the loss function is reduced to the minimum value, and therefore an ideal entity labeling effect is obtained. The attribute keywords in the result character string are extracted through the model, and the relationship is established with the objects in the environment, so that the object semantics in the environment are enriched, and the knowledge system in the semantic visual space is further improved.
(3) Storage of environmental knowledge
After the visual relationship and the attribute of the article in the environment are acquired, the semantic visual space is constructed to finish data preparation, and the connection form of the data knowledge is diversified, so that the diversified information needs to be efficiently organized and expressed in a multi-relationship graph form. The basic construction form of the multi-relationship graph is nodes and edges, wherein the nodes store entity semantic and item attribute information, and the edges represent the relationship between entities and the relationship between the entities and the attributes. Unlike a general graph, a multi-relationship graph contains multiple types of nodes and multiple types of edges. FIG. 3 is an example of a multiple relationship graph store.
In fig. 3, the gray boxes represent entity information in the environment and the white boxes represent attributes of the item. It can be seen that the visual relationship and the article attribute in the environment are combined in the form of a multiple relationship diagram, and through the combination, the robot can understand the article relationship in the environment not only at the semantic level, but also increase the knowledge of the article at the logical level, thereby forming a uniform and standard expression.
And storing the acquired environmental knowledge through a multi-relation graph to finally form a semantic visual space suitable for the intelligent service of the robot.
Example 2:
an embodiment 2 of the present disclosure provides a spatial environment recognition system for robot intelligent service, including:
a data acquisition module configured to: acquiring at least one image of a spatial environment region to be identified;
a visual relationship triplet acquisition module configured to: according to the obtained image, a preset visual relationship obtaining model is adopted, the object features in the image and the relationship features between the objects are obtained, the triple features of the objects in the image are obtained, and then the visual relationship triple of the objects in the image is obtained;
an item attribute acquisition module configured to: obtaining a model by using a preset attribute according to the obtained article characteristics to obtain the attribute of each article;
an item multiple relationship diagram acquisition module configured to: obtaining an article multiple relation graph in the current image according to the obtained visual relation triples of each article and the attributes of each article;
an environment identification module configured to: and processing each image of the environment area to be identified to obtain an article multiple relation graph, and constructing a robot semantic visual space according to the obtained article multiple relation graph of each image to realize the identification of the environment.
The working method of the system described in this embodiment is the same as the spatial environment recognition method for the robot intelligent service in embodiment 1, and details are not repeated here.
Example 3:
the embodiment 3 of the present disclosure provides a medium on which a program is stored, which when executed by a processor, implements the steps in the spatial environment recognition method for robot intelligent service according to the embodiment 1 of the present disclosure.
Example 4:
the embodiment 4 of the present disclosure provides an electronic device, which includes a memory, a processor, and a program stored in the memory and executable on the processor, and when the processor executes the program, the steps in the method for identifying a spatial environment for robot intelligent services according to embodiment 1 of the present disclosure are implemented.
As will be appreciated by one skilled in the art, embodiments of the present disclosure may be provided as a method, system, or computer program product. Accordingly, the present disclosure may take the form of a hardware embodiment, a software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present disclosure may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, optical storage, and the like) having computer-usable program code embodied therein.
The present disclosure is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the disclosure. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by a computer program, which can be stored in a computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. The storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), or the like.
The above description is only a preferred embodiment of the present disclosure and is not intended to limit the present disclosure, and various modifications and changes may be made to the present disclosure by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present disclosure should be included in the protection scope of the present disclosure.

Claims (9)

1. A space environment identification method for robot intelligent service is characterized by comprising the following steps:
acquiring at least one image of a spatial environment region to be identified;
acquiring a model by adopting a preset visual relationship according to the acquired image, acquiring article characteristics in the image and relationship characteristics among articles, and further acquiring a visual relationship triple of the articles in the image;
the relation characteristics among the objects are characterized by using the spatial relation among the objects, and specifically comprise the following steps:
the triple features comprise word vectors of objects in the image, and the relation between the targets uses [ x ]1,x2,x3]Denotes x1Represents object 1, x3Represents object 3, x2Representing the relationship, then the spatial relationship s is represented as follows:
Figure FDA0002993952510000011
wherein,
Figure FDA0002993952510000012
a prediction block representing a target object, wherein i-1 or 3,
Figure FDA0002993952510000013
to predict the center coordinates of the box, wiAnd hiRespectively representing the width and height of the prediction box, W and H representing the width and height of the intersection of the target 1 and target 3 prediction boxes,
Figure FDA0002993952510000014
is the encoding of the relevant spatial relationships;
after the spatial relation s is obtained, the s is input into a multilayer perceptron, and a feature representation x with the same dimension as the word vector is obtained2Thereby obtaining a feature expression of the triplet;
obtaining a model by using a preset attribute according to the obtained article characteristics to obtain the attribute of each article;
obtaining an article multiple relation graph in the current image according to the obtained visual relation triples of each article and the attributes of each article;
and processing each acquired image of the environment area to be identified to obtain an article multiple relation graph, and constructing a robot semantic visual space according to the article multiple relation graph of each acquired image to realize identification of the space environment.
2. The spatial environment recognition method for robot intelligent services according to claim 1, wherein the preset visual relationship acquisition model comprises a convolutional neural network and a bidirectional cyclic neural network for target detection, and specifically comprises:
performing target detection on the obtained image according to the convolutional neural network, acquiring the semantics of a target article and the coordinate information of a detection frame thereof, performing feature coding on the semantics of the target article and a non-target article, and acquiring the relationship features between the articles according to the coordinate information of the detection frame to obtain triple features;
and inputting the obtained triple features into a bidirectional cyclic neural network for visual relationship prediction to obtain the visual relationship triples of each article in the image.
3. The spatial environment recognition method for robot intelligent services according to claim 2, wherein the preset visual relationship obtaining model further includes a dynamic filtering module, and the dynamic filtering module filters out the visual relationship triples when detecting that the visual relationship triples contain dynamic objects through a preset dynamic object list.
4. The method of claim 3, wherein if the speed of change of the visual relationship of a certain type of objects exceeds a preset threshold, adding the objects into the dynamic object list to filter out the corresponding visual relationship triples.
5. The spatial environment recognition method for robot intelligent services according to claim 1, wherein the attributes of each item are obtained using a preset attribute acquisition model, specifically:
searching article correlation in a preset database by taking the semantics of the articles as keywords to obtain a plurality of result character strings;
according to the obtained result character string, identifying attribute keywords in the result by adopting a BERT-CRF model, which specifically comprises the following steps:
semantic text information of the articles is extracted by using a BERT model, characteristic coding is carried out, the codes are input into a CRF layer for information decoding, and the labeling sequences of the articles are obtained.
6. The spatial environment recognition method for robotic intelligence services as claimed in claim 1 wherein the item multiple relationship graph includes a plurality of nodes for storing entity semantics and item attribute information and a plurality of edges representing relationships between entities and attributes.
7. A spatial environment recognition system for robotic intelligence services, comprising:
a data acquisition module configured to: acquiring at least one image of a spatial environment region to be identified;
a visual relationship triplet acquisition module configured to: acquiring a model by adopting a preset visual relationship according to the acquired image, acquiring article characteristics in the image and relationship characteristics among articles, and further acquiring a visual relationship triple of the articles in the image;
the relation characteristics among the objects are characterized by using the spatial relation among the objects, and specifically comprise the following steps:
the triple features comprise word vectors of objects in the image, and the relation between the targets uses [ x ]1,x2,x3]Denotes x1Represents object 1, x3Represents object 3, x2Representing the relationship, then the spatial relationship s is represented as follows:
Figure FDA0002993952510000031
wherein,
Figure FDA0002993952510000032
a prediction block representing a target object, wherein i-1 or 3,
Figure FDA0002993952510000033
to predict the center coordinates of the box, wiAnd hiRespectively representing the width and height of the prediction box, W and H representing the width and height of the intersection of the target 1 and target 3 prediction boxes,
Figure FDA0002993952510000034
is the encoding of the relevant spatial relationships;
after the spatial relation s is obtained, the s is input into a multilayer perceptron, and a feature representation x with the same dimension as the word vector is obtained2Thereby obtaining a feature expression of the triplet;
an item attribute acquisition module configured to: obtaining a model by using a preset attribute according to the obtained article characteristics to obtain the attribute of each article;
an item multiple relationship diagram acquisition module configured to: obtaining an article multiple relation graph in the current image according to the obtained visual relation triples of each article and the attributes of each article;
an environment identification module configured to: and processing each image of the environment area to be identified to obtain an article multiple relation graph, and constructing a robot semantic visual space according to the obtained article multiple relation graph of each image to realize the identification of the environment.
8. A medium on which a program is stored, characterized in that the program, when being executed by a processor, realizes the steps in the method for spatial environment recognition for robot intelligence services according to any of claims 1-6.
9. An electronic device comprising a memory, a processor and a program stored on the memory and executable on the processor, wherein the processor implements the steps of the method for spatial environment recognition for robot intelligence services according to any of claims 1-6 when executing the program.
CN202010228789.1A2020-03-272020-03-27 A space environment recognition method and system for robot intelligent serviceActiveCN111496784B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202010228789.1ACN111496784B (en)2020-03-272020-03-27 A space environment recognition method and system for robot intelligent service

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202010228789.1ACN111496784B (en)2020-03-272020-03-27 A space environment recognition method and system for robot intelligent service

Publications (2)

Publication NumberPublication Date
CN111496784A CN111496784A (en)2020-08-07
CN111496784Btrue CN111496784B (en)2021-05-07

Family

ID=71867152

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202010228789.1AActiveCN111496784B (en)2020-03-272020-03-27 A space environment recognition method and system for robot intelligent service

Country Status (1)

CountryLink
CN (1)CN111496784B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN114630280B (en)*2020-12-142024-06-04曲阜师范大学 Indoor intelligent positioning system and control method thereof
CN112765992A (en)*2021-01-142021-05-07深圳市人马互动科技有限公司Training data construction method and device, computer equipment and storage medium

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN107516127B (en)*2017-08-212020-06-30山东大学 A method and system for a service robot to autonomously acquire the attribution semantics of items carried by people
CN107807986B (en)*2017-10-312019-12-17中南大学 A Method for Intelligent Understanding of Remote Sensing Image Using Semantics to Describe the Spatial Relationship of Objects
CN110851609A (en)*2018-07-242020-02-28华为技术有限公司 Representation learning method and device
CN109816686A (en)*2019-01-152019-05-28山东大学 Robot semantic SLAM method, processor and robot based on object instance matching
CN110334219B (en)*2019-07-122023-05-09电子科技大学Knowledge graph representation learning method based on attention mechanism integrated with text semantic features
CN110489395B (en)*2019-07-272022-07-29西南电子技术研究所(中国电子科技集团公司第十研究所)Method for automatically acquiring knowledge of multi-source heterogeneous data
CN110717052A (en)*2019-10-152020-01-21山东大学 Environment representation method in intelligent service of service robot
CN110825829B (en)*2019-10-162023-05-26华南理工大学 A Method for Autonomous Robot Navigation Based on Natural Language and Semantic Maps
CN110866190B (en)*2019-11-182021-05-14支付宝(杭州)信息技术有限公司Method and device for training neural network model for representing knowledge graph
CN110781254A (en)*2020-01-022020-02-11四川大学Automatic case knowledge graph construction method, system, equipment and medium

Also Published As

Publication numberPublication date
CN111496784A (en)2020-08-07

Similar Documents

PublicationPublication DateTitle
Rukhovich et al.Iterdet: iterative scheme for object detection in crowded environments
CN113221787B (en)Pedestrian multi-target tracking method based on multi-element difference fusion
CN109344285B (en)Monitoring-oriented video map construction and mining method and equipment
CN109993102A (en)Similar face retrieval method, apparatus and storage medium
Dutta et al.Predicting human actions taking into account object affordances
CN110146080B (en)SLAM loop detection method and device based on mobile robot
CN111496784B (en) A space environment recognition method and system for robot intelligent service
CN117423032B (en) Spatiotemporal fine-grained human motion timing segmentation method, electronic device and computer-readable storage medium
CN112884802A (en) A Generative Adversarial Attack Method
AFFES et al.Comparison of YOLOV5, YOLOV6, YOLOV7 and YOLOV8 for Intelligent Video Surveillance.
Korrapati et al.Vision-based sparse topological mapping
CN116912579A (en) Scene graph generation method based on multi-level attention mechanism
EP1801731A1 (en)Adaptive scene dependent filters in online learning environments
CN118351138B (en) A general category multi-target tracking method integrating prototype learning mechanism
Gupta et al.Traffic sign sensing: A deep learning approach for enhanced road safety
Sanchez-Matilla et al.Motion prediction for first-person vision multi-object tracking
CN116597419B (en)Vehicle height limiting scene identification method based on parameterized mutual neighbors
Tao et al.Semi-supervised online learning for efficient classification of objects in 3d data streams
ChallaArtificial intelligence for object detection and its metadata
CN110765809A (en) A facial expression classification method, device and emotionally intelligent robot
Ravipati et al.Object-oriented material classification and 3d clustering for improved semantic perception and mapping in mobile robots
Ahn et al.Interpretable Information Visualization for Enhanced Temporal Action Detection in Videos
JokelaPerson counter using real-time object detection and a small neural network
Naushad et al.An Investigation of Context-Aware Object Detection based on Scene Recognition
Rios-Cabrera et al.Dynamic categorization of 3D objects for mobile service robots

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp