Movatterモバイル変換


[0]ホーム

URL:


US20210334629A1 - Hybrid neural network architecture within cascading pipelines - Google Patents

Hybrid neural network architecture within cascading pipelines
Download PDF

Info

Publication number
US20210334629A1
US20210334629A1US17/116,229US202017116229AUS2021334629A1US 20210334629 A1US20210334629 A1US 20210334629A1US 202017116229 AUS202017116229 AUS 202017116229AUS 2021334629 A1US2021334629 A1US 2021334629A1
Authority
US
United States
Prior art keywords
inferencing
deep learning
metadata
data
processing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/116,229
Inventor
Wind Yuan
Kaustubh Purandare
Bhushan Rupde
Shaunak Gupte
Farzin Aghdasi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nvidia Corp
Original Assignee
Nvidia Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nvidia CorpfiledCriticalNvidia Corp
Priority to US17/116,229priorityCriticalpatent/US20210334629A1/en
Assigned to NVIDIA CORPORATIONreassignmentNVIDIA CORPORATIONASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: GUPTE, SHAUNAK, RUPDE, BHUSHAN, YUAN, WIND, AGHDASI, FARZIN, PURANDARE, KAUSTUBH
Publication of US20210334629A1publicationCriticalpatent/US20210334629A1/en
Pendinglegal-statusCriticalCurrent

Links

Images

Classifications

Definitions

Landscapes

Abstract

A multi-stage multimedia inferencing pipeline may be set up and executed using configuration data including information used to set up each stage by deploying the specified or desired models and/or other pipeline components into a repository (e.g., a shared folder in a repository). The configuration data may also include information a central inference server library uses to manage and set parameters for these components with respect to a variety of inference frameworks that may be incorporated into the pipeline. The configuration data can define a pipeline that encompasses stages for video decoding, video transform, cascade inferencing on different frameworks, metadata filtering and exchange between models and display. The entire pipeline can be efficiently hardware-accelerated using parallel processing circuits (e.g., one or more GPUs, CPUs, DPUs, or TPUs). Embodiments of the present disclosure can integrate an entire video/audio analytics pipeline into an embedded platform in real time.

Description

Claims (21)

What is claimed is:
1. A method comprising:
accessing configuration data corresponding to an inferencing pipeline, the inferencing pipeline comprising at least one pre-processing stage, at least one inferencing stage, and at least one post-processing stage;
pre-processing multimedia data during the at least one pre-processing stage;
providing the multimedia data to one or more deep learning models during the at least one inferencing stage, the one or more deep learning models including at least a first deep learning model associated with a first framework and a second deep learning model associated with a second framework that is a different framework from than the first framework;
post-processing, during the at least one post-processing stage, output generated during the inferencing stage , the inferencing performed on the multimedia data using the plurality of deep learning models; and
providing the post-processed output for display by an on-screen display.
2. The method ofclaim 1, wherein the first deep learning model is configured according to a first configuration file, the second deep learning model is configured according to a second configuration file, and the inferencing pipeline is configured according to a third configuration file.
3. The method ofclaim 1, wherein the pre-processing of the multimedia data comprises batch processing a plurality of multimedia streams concurrently.
4. The method ofclaim 1, wherein the inferencing, the pre-processing, and the post-processing are performed on at least one of: a cloud-based server or an edge device.
5. The method ofclaim 1, wherein the inferencing operates the first deep learning model and the second deep learning model in parallel.
6. The method ofclaim 1, wherein the pre-processing extracts metadata from the multimedia data, and the inferencing pipeline filters the metadata to generate a first input to the first deep learning model and a second input to the second deep learning model, wherein the inferencing uses the first input and the second input.
7. The method ofclaim 1, wherein the pre-processing is hardware-accelerated and comprises at least one of:
decoding of the multimedia data;
converting the multimedia data from a first multimedia format to a second multimedia format; or
resizing one or more units of the multimedia data.
8. The method ofclaim 1, wherein one or more stages of the inferencing pipeline are performed, at least partially, on at least one of: one or more Virtual Machines (VMs) or one or more containerized applications.
9. The method ofclaim 1, wherein the post-processed output comprises metadata, and the post-processing comprises batch post-processing the output from the first deep learning model and the second deep learning model separately and respectively to generate the metadata.
10. The method ofclaim 1, wherein the post-processing comprises at least one of:
performing object detection;
performing object classification;
performing class segmentation;
performing super resolution processing; or
performing language processing of audio data.
11. The method ofclaim 1, wherein the at least one pre-processing stage comprises performing primary inferencing and the at least one inferencing stage comprises performing secondary inferencing.
12. A method comprising:
pre-processing multimedia data to extract metadata using at least a first stage of an inferencing pipeline;
providing the multimedia data and the metadata to a plurality of deep learning models of at least a second stage of the inferencing pipeline, the plurality of deep learning models including at least a first deep learning model associated with a first framework and a second deep learning model associated with a second framework;
generating post-processed output of inferencing using at least a third stage of the inferencing pipeline, the inferencing performed on the multimedia data using the plurality of deep learning models and the metadata; and
providing the post-processed output for display by an on-screen display.
13. The method ofclaim 12, wherein the providing the metadata to the plurality of deep learning models comprises providing at least the metadata to a backend using one or more Application Programming Interfaces (APIs).
14. The method ofclaim 12, wherein the metadata comprises data corresponding to at least one of:
one or more class-identifiers;
one or more labels;
display information;
one or more filtered objects;
one or more segmentation maps;
network information; or one or more tensors representing raw sensor output.
15. The method ofclaim 12, wherein the providing the multimedia data and the metadata comprises filtering the metadata to generate a first input to the first deep learning model and a second input to the second deep learning model, wherein the inferencing uses the first input and the second input.
16. The method ofclaim 12, comprising accessing configuration data that defines at least the first stage, the second stage, and the third stage of the inferencing pipeline.
17. A system comprising:
one or more processing devices and one or more memory devices communicatively coupled to the one or more processing devices storing programmed instructions thereon, which when executed by the one or more processing devices causes performance of an inferencing pipeline by the one or more processing devices, the performance comprising:
determining first metadata from multimedia data using at least one deep learning model corresponding to a first runtime environment;
sending the first metadata to a backend server library using one or more Application Programming Interfaces (APIs), the backend server library executing a plurality of deep learning models including at least a first deep learning model of a first framework and corresponding to a second runtime environment, and a second deep learning model of a second framework and corresponding to a third runtime environment;
receiving, using the one or more APIs, inferencing output generated using the multimedia data, the plurality of deep learning models, and the first metadata;
generating second metadata from at least a first portion of the output of the second runtime environment and a second portion of the output from the third runtime environment; and
providing the second metadata to one or more downstream components.
18. The system ofclaim 17, wherein the first runtime environment corresponds to a third framework that is different than the first framework and the second framework.
19. The system ofclaim 17, wherein the at least one deep learning model corresponds to an object detector used to detect objects and the first deep learning model includes an object classifier used to classify one or more of the objects.
20. The system ofclaim 17, wherein the at least one deep learning model corresponds to an object detector to generate object detections and the first metadata is generated using an object tracker that operates on the object detections.
21. The system ofclaim 17, wherein the at least one deep learning model corresponds to at least one of:
a control system for an autonomous or semi-autonomous machine;
a perception system for an autonomous or semi-autonomous machine;
a system for performing simulation operations;
a system for performing deep learning operations;
a system implemented using an edge device;
a system implemented using a robot;
a system incorporating one or more Virtual Machines (VMs);
a system implemented at least partially in a data center; or
a system implemented at least partially using cloud computing resources.
US17/116,2292020-04-252020-12-09Hybrid neural network architecture within cascading pipelinesPendingUS20210334629A1 (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
US17/116,229US20210334629A1 (en)2020-04-252020-12-09Hybrid neural network architecture within cascading pipelines

Applications Claiming Priority (2)

Application NumberPriority DateFiling DateTitle
US202063015486P2020-04-252020-04-25
US17/116,229US20210334629A1 (en)2020-04-252020-12-09Hybrid neural network architecture within cascading pipelines

Publications (1)

Publication NumberPublication Date
US20210334629A1true US20210334629A1 (en)2021-10-28

Family

ID=78222424

Family Applications (1)

Application NumberTitlePriority DateFiling Date
US17/116,229PendingUS20210334629A1 (en)2020-04-252020-12-09Hybrid neural network architecture within cascading pipelines

Country Status (1)

CountryLink
US (1)US20210334629A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN114140311A (en)*2021-11-042022-03-04北京激浊扬清文化科技有限公司 Delay control method for streaming media AI reasoning based on gstreamer
CN114550175A (en)*2022-02-172022-05-27Oppo广东移动通信有限公司 Image processing method, apparatus, electronic device, and computer-readable storage medium
CN115662518A (en)*2022-12-272023-01-31四川大学华西医院Gene sequencing and storage cooperation system, method and computer readable storage medium
US20230214640A1 (en)*2021-12-312023-07-06Discover Financial ServicesBuilding and executing deep learning-based data pipelines
US20230230351A1 (en)*2022-01-142023-07-20Microsoft Technology Licensing, LlcConstructing processing pipeline at edge computing device
WO2023177106A1 (en)*2022-03-182023-09-21한국과학기술원Microservice-based edge device architecture for image analysis
EP4276624A1 (en)*2022-05-132023-11-15Siemens AktiengesellschaftA method and a server for installation and execution of an ai model in an industrial execution environment
WO2023249764A1 (en)*2022-06-212023-12-28Microsoft Technology Licensing, LlcOn-device artificial intelligence processing in-browser

Citations (20)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20190141340A1 (en)*2018-04-022019-05-09Intel CorporationVideo analytics encoding for improved efficiency of video processing and compression
US20190324745A1 (en)*2018-04-202019-10-24International Business Machines CorporationMapping components of a non-distributed environment to a distributed environment
US20190392345A1 (en)*2018-06-262019-12-26Holland & Hart LLPAutomatic action loading based on a data context and a machine-learned model
US20200021873A1 (en)*2018-07-112020-01-16Adobe Inc.Utilizing artificial intelligence to generate enhanced digital content and improve digital content campaign design
US20200219291A1 (en)*2019-01-032020-07-09Samsung Electronics Co., Ltd.Display apparatus, image providing apparatus, and methods of controlling the same
US20200225317A1 (en)*2020-03-272020-07-16Chulong ChenApparatus, system and method of generating radar perception data
US20200319919A1 (en)*2019-04-052020-10-08Alibaba Group Holding LimitedSystems and methods for scheduling neural networks by varying batch sizes
US10810445B1 (en)*2018-06-292020-10-20Zoox, Inc.Pipeline with point cloud filtering
US20200342290A1 (en)*2018-06-152020-10-29Deep Insight Solutions, Inc. d/b/a/ LucdOptimistic Execution of Reservoir Computing Machine Learning Models
US10834127B1 (en)*2018-04-242020-11-10Trend Micro IncorporatedDetection of business email compromise attacks
US20200403912A1 (en)*2019-06-242020-12-24Amazon Technologies, Inc.Serverless packet processing service with configurable exception paths
US20210012236A1 (en)*2019-07-102021-01-14Sony Interactive Entertainment LLCDual machine learning pipelines for transforming data and optimizing data transformation
US20210064431A1 (en)*2019-08-302021-03-04Microstrategy IncorporatedEnvironment monitoring and management
US20210074425A1 (en)*2019-09-052021-03-11Pearl Inc.Systems and methods for automated medical image analysis
US20210096264A1 (en)*2019-09-302021-04-01Zoox, Inc.Complex ground profile estimation
US20210201128A1 (en)*2019-12-272021-07-01Clari Inc.System and method for generating scores for predicting probabilities of task completion
US20220053071A1 (en)*2019-03-142022-02-17Hewlett-Packard Development Company, L.P.Preprocessing sensor data for machine learning
US20220138596A1 (en)*2020-11-022022-05-05Adobe Inc.Increasing efficiency of inferencing digital videos utilizing machine-learning models
US20230105476A1 (en)*2020-03-062023-04-06Google LlcDistributed computing pipeline processing
US11797875B1 (en)*2018-03-282023-10-24Meta Platforms, Inc.Model compression for selecting content

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US11797875B1 (en)*2018-03-282023-10-24Meta Platforms, Inc.Model compression for selecting content
US20190141340A1 (en)*2018-04-022019-05-09Intel CorporationVideo analytics encoding for improved efficiency of video processing and compression
US20190324745A1 (en)*2018-04-202019-10-24International Business Machines CorporationMapping components of a non-distributed environment to a distributed environment
US10834127B1 (en)*2018-04-242020-11-10Trend Micro IncorporatedDetection of business email compromise attacks
US20200342290A1 (en)*2018-06-152020-10-29Deep Insight Solutions, Inc. d/b/a/ LucdOptimistic Execution of Reservoir Computing Machine Learning Models
US20190392345A1 (en)*2018-06-262019-12-26Holland & Hart LLPAutomatic action loading based on a data context and a machine-learned model
US10810445B1 (en)*2018-06-292020-10-20Zoox, Inc.Pipeline with point cloud filtering
US20200021873A1 (en)*2018-07-112020-01-16Adobe Inc.Utilizing artificial intelligence to generate enhanced digital content and improve digital content campaign design
US20200219291A1 (en)*2019-01-032020-07-09Samsung Electronics Co., Ltd.Display apparatus, image providing apparatus, and methods of controlling the same
US20220053071A1 (en)*2019-03-142022-02-17Hewlett-Packard Development Company, L.P.Preprocessing sensor data for machine learning
US20200319919A1 (en)*2019-04-052020-10-08Alibaba Group Holding LimitedSystems and methods for scheduling neural networks by varying batch sizes
US20200403912A1 (en)*2019-06-242020-12-24Amazon Technologies, Inc.Serverless packet processing service with configurable exception paths
US20210012236A1 (en)*2019-07-102021-01-14Sony Interactive Entertainment LLCDual machine learning pipelines for transforming data and optimizing data transformation
US20210064431A1 (en)*2019-08-302021-03-04Microstrategy IncorporatedEnvironment monitoring and management
US20210074425A1 (en)*2019-09-052021-03-11Pearl Inc.Systems and methods for automated medical image analysis
US20210096264A1 (en)*2019-09-302021-04-01Zoox, Inc.Complex ground profile estimation
US20210201128A1 (en)*2019-12-272021-07-01Clari Inc.System and method for generating scores for predicting probabilities of task completion
US20230105476A1 (en)*2020-03-062023-04-06Google LlcDistributed computing pipeline processing
US20200225317A1 (en)*2020-03-272020-07-16Chulong ChenApparatus, system and method of generating radar perception data
US20220138596A1 (en)*2020-11-022022-05-05Adobe Inc.Increasing efficiency of inferencing digital videos utilizing machine-learning models

Cited By (11)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN114140311A (en)*2021-11-042022-03-04北京激浊扬清文化科技有限公司 Delay control method for streaming media AI reasoning based on gstreamer
US20230214640A1 (en)*2021-12-312023-07-06Discover Financial ServicesBuilding and executing deep learning-based data pipelines
WO2023129402A1 (en)*2021-12-312023-07-06Discover Financial ServicesBuilding and executing deep learning-based data pipelines
US20230230351A1 (en)*2022-01-142023-07-20Microsoft Technology Licensing, LlcConstructing processing pipeline at edge computing device
US12211248B2 (en)*2022-01-142025-01-28Microsoft Technology Licensing, LlcConstructing processing pipeline at edge computing device
CN114550175A (en)*2022-02-172022-05-27Oppo广东移动通信有限公司 Image processing method, apparatus, electronic device, and computer-readable storage medium
WO2023177106A1 (en)*2022-03-182023-09-21한국과학기술원Microservice-based edge device architecture for image analysis
EP4276624A1 (en)*2022-05-132023-11-15Siemens AktiengesellschaftA method and a server for installation and execution of an ai model in an industrial execution environment
WO2023217485A1 (en)2022-05-132023-11-16Siemens AktiengesellschaftA method and a server for installation and execution of an ai model in an industrial execution environment
WO2023249764A1 (en)*2022-06-212023-12-28Microsoft Technology Licensing, LlcOn-device artificial intelligence processing in-browser
CN115662518A (en)*2022-12-272023-01-31四川大学华西医院Gene sequencing and storage cooperation system, method and computer readable storage medium

Similar Documents

PublicationPublication DateTitle
US20210334629A1 (en)Hybrid neural network architecture within cascading pipelines
JP7690296B2 (en) Unsupervised learning of scene structure for synthetic data generation
US11995883B2 (en)Scene graph generation for unlabeled data
US12288277B2 (en)High-precision semantic image editing using neural networks for synthetic data generation systems and applications
CN114365156A (en) Transfer Learning for Neural Networks
WO2021178936A1 (en)Neural rendering for inverse graphics generation
US20240111894A1 (en)Generative machine learning models for privacy preserving synthetic data generation using diffusion
US20250182404A1 (en)Four-dimensional object and scene model synthesis using generative models
US11922558B2 (en)Hybrid differentiable rendering for light transport simulation systems and applications
WO2022251672A1 (en)Processing ultrahyperbolic representations using neural networks
US20240256831A1 (en)Unsupervised pre-training of neural networks using generative models
US12243152B2 (en)Hybrid differentiable rendering for light transport simulation systems and applications
US20250014571A1 (en)Joint training of speech recognition and speech synthesis models for conversational ai systems and applications
CN118196277A (en) Replicate physical environments and generate 3D assets for synthetic scene generation
US20250232471A1 (en)Visual localization of image viewpoints in 3d scenes using neural representations
US20220383073A1 (en)Domain adaptation using domain-adversarial learning in synthetic data systems and applications
US20250061612A1 (en)Neural networks for synthetic data generation with discrete and continuous variable features
CN119494786A (en) Context Preservation for Synthetic Image Enhancement Using Diffusion
US12400395B2 (en)Rendering and composition of neural 3D objects with non-neural assets in content generation systems and applications
WO2022251661A1 (en)Domain adaptation using domain-adversarial learning in synthetic data systems and applications
CN116766173A (en) Interpret discrete tasks from complex instructions for robotic systems and applications
US20250029409A1 (en)Auto-labeling systems and applications for open-set and out-of-domain segmentation
US20250131680A1 (en)Feature extraction with three-dimensional information
US12339899B2 (en)Content generation systems and applications using octree-based spatial databases
US20240221288A1 (en)Selecting representative image views for 3d object models in synthetic content creation systems and applications

Legal Events

DateCodeTitleDescription
ASAssignment

Owner name:NVIDIA CORPORATION, CALIFORNIA

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YUAN, WIND;PURANDARE, KAUSTUBH;RUPDE, BHUSHAN;AND OTHERS;SIGNING DATES FROM 20201208 TO 20201209;REEL/FRAME:054593/0006

STPPInformation on status: patent application and granting procedure in general

Free format text:DOCKETED NEW CASE - READY FOR EXAMINATION

STPPInformation on status: patent application and granting procedure in general

Free format text:NON FINAL ACTION MAILED

STPPInformation on status: patent application and granting procedure in general

Free format text:RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPPInformation on status: patent application and granting procedure in general

Free format text:FINAL REJECTION MAILED

STPPInformation on status: patent application and granting procedure in general

Free format text:DOCKETED NEW CASE - READY FOR EXAMINATION

STPPInformation on status: patent application and granting procedure in general

Free format text:NON FINAL ACTION MAILED

STPPInformation on status: patent application and granting procedure in general

Free format text:RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPPInformation on status: patent application and granting procedure in general

Free format text:DOCKETED NEW CASE - READY FOR EXAMINATION


[8]ページ先頭

©2009-2025 Movatter.jp