Movatterモバイル変換


[0]ホーム

URL:


US20160358015A1 - Detection of cast members in video content - Google Patents

Detection of cast members in video content
Download PDF

Info

Publication number
US20160358015A1
US20160358015A1US15/238,413US201615238413AUS2016358015A1US 20160358015 A1US20160358015 A1US 20160358015A1US 201615238413 AUS201615238413 AUS 201615238413AUS 2016358015 A1US2016358015 A1US 2016358015A1
Authority
US
United States
Prior art keywords
frame
face
computing device
scene
video program
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/238,413
Inventor
Arnab Sanat Kumar Dhua
Gautam Bhargava
Douglas Ryan Gray
Sunil Ramesh
Colin Jon Taylor
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
A9 com Inc
Original Assignee
A9 com Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by A9 com IncfiledCriticalA9 com Inc
Priority to US15/238,413priorityCriticalpatent/US20160358015A1/en
Publication of US20160358015A1publicationCriticalpatent/US20160358015A1/en
Abandonedlegal-statusCriticalCurrent

Links

Images

Classifications

Definitions

Landscapes

Abstract

Disclosed are various embodiments for detection of cast members in video content such as movies, television shows, and other programs. Data indicating cast members who appear in a video program is obtained. Each cast member is associated with a reference image depicting a face of the cast member. A frame is obtained from the video program, and a face is detected in the frame. The frame can correspond to a scene in the video program. The detected face in the frame is recognized as being a particular cast member based at least in part on the reference image depicting the cast member. An association between the cast member and the frame is generated in response to the detected face in the frame being recognized as the cast member.

Description

Claims (20)

Therefore, the following is claimed:
1. A non-transitory computer-readable medium embodying a program that, when executed by at least one computing device, causes the at least one computing device to at least:
sample a first frame from a scene in a video program;
detect a face in the first frame of the scene;
recognize the face in the first frame as being one of a plurality of cast members known to appear in the video program based at least in part on a plurality of reference images corresponding to the plurality of cast members;
generate a first association between the one of the plurality of cast members and the first frame when the face is recognized as being the one of the plurality of cast members; and
generate a second association between the one of the plurality of cast members and a second frame from the scene in the video program based at least in part on the first association, a third frame of the video program, and a temporal smoothing factor.
2. The non-transitory computer-readable medium ofclaim 1, wherein the program further causes the at least one computing device to at least generate a third association between the one of the plurality of cast members and the scene based at least in part on the first association and the second association.
3. The non-transitory computer-readable medium ofclaim 1, wherein the program further causes the at least one computing device to at least generate a third association between the one of the plurality of cast members and a fourth frame from the scene in the video program, the third association indicating the one of the plurality of cast members is predicted to appear in the fourth frame.
4. The non-transitory computer-readable medium ofclaim 1, wherein the program further causes the at least one computing device to at least detect an ending frame of the scene based at least in part on a scene break.
5. The non-transitory computer-readable medium ofclaim 4, wherein the program further causes the at least one computing device to at least identify the scene break by detecting a change in contrast between at least two frames of the video program.
6. The non-transitory computer-readable medium ofclaim 1, wherein the scene corresponds to a plurality of consecutive frames that comprises the first frame and the second frame.
7. The non-transitory computer-readable medium ofclaim 1, wherein the scene corresponds to a distinct plot element of the video program.
8. A system, comprising:
a data store; and
at least one computing device comprising a hardware processor, the at least one computing device being in communication with the data store, the at least one computing device being configured to at least:
receive data indicating a plurality of persons that are known to appear in a video program, an individual person of the plurality of persons being associated with a reference image depicting a respective face of the individual person;
identify a frame from the video program;
detect a face in the frame;
recognize the face in the frame as being the individual person based at least in part on the reference image depicting the respective face of the individual person; and
generate an association between the individual person and the frame in response to the face in the frame being recognized as being the individual person.
9. The system ofclaim 8, wherein the at least one computing device is further configured to at least:
submit the face in the frame for a manual review; and
in response to the manual review indicating that the face matches the reference image, update the data indicating the plurality of persons to improve accuracy of another recognition in another frame.
10. The system ofclaim 9, wherein the other frame corresponds to a second video program.
11. The system ofclaim 8, wherein the at least one computing device is further configured to at least:
detect a second face in a second frame;
recognize the second face in the second frame as being the individual person based at least in part on the reference image depicting the respective face of the individual person; and
generate a second association between the individual person and the second frame.
12. The system ofclaim 11, wherein the at least one computing device is further configured to at least generate a third association between the individual person and a third frame based at least in part on a temporal smoothing factor, wherein the individual person is unrecognized in the third frame, and the third frame is located chronologically between the frame and the second frame.
13. The system ofclaim 12, wherein a count of a number of frames between the frame and the second frame is at or below a threshold count of video frames in which a person may be unrecognized and be associated with a particular frame.
14. A method, comprising:
obtaining, by at least one computing device, a frame from a scene of a video program;
identifying, by the at least one computing device, a face in the frame;
recognizing, by the at least one computing device, the face in the frame as one of a plurality of persons based at least in part on a reference image depicting a person; and
generating, by the at least one computing device, an association between the person and the scene in response to the face in the frame being recognized as being the person.
15. The method ofclaim 14, further comprising:
obtaining, by the at least one computing device, a second frame from video program;
detecting, by the at least one computing device, a second face in the second frame;
recognizing, by the at least one computing device, the second face in the frame as the one of the plurality of persons; and
determining, by the at least one computing device, a trajectory of the one of the plurality of persons in the video program based at least in part on the frame and the second frame.
16. The method ofclaim 15, further comprising:
obtaining, by the at least one computing device, a third frame from video program; and
detecting, by the at least one computing device, a third face in the third frame based at least in part on the trajectory of the one of the plurality of persons.
17. The method ofclaim 14, further comprising:
obtaining, by the at least one computing device, a subset of a plurality of frames of the video program by selecting a single respective frame of the plurality of frames for an individual predefined time interval;
identifying, by the at least one computing device, a respective face in individual ones of the subset of the plurality of frames; and
recognizing, by the at least one computing device, the respective face in the individual ones of the subset of the plurality of frames as a respective one of the plurality of persons based at least in part on a plurality of reference images depicting a particular person known to appear in the video program.
18. The method ofclaim 14, further comprising:
Identifying, by the at least one computing device, a scene break in the scene by detecting a change in contrast between at least two frames of the scene; and
detecting, by the at least one computing device, an ending frame of the scene based at least in part on the scene break.
19. The method ofclaim 14, further comprising:
providing, by the at least one computing device, the face in the frame for a manual review;
receiving, by the at least one computing device, an indication that the face matches the reference image based at least in part on the manual review;
updating, by the at least one computing device, data comprising the reference image depicting the person based at least in part on the frame; and
recognizing, by the at least one computing device, another face in another video program based at least in part on the updated data.
20. The method ofclaim 14, wherein recognizing the face in the frame as the one of the plurality of persons further comprises:
identifying, by the at least one computing device, a portion of the frame based at least in part on the face identified in the frame; and
performing, by the at least one computing device, a facial recognition on the portion of the frame.
US15/238,4132013-04-102016-08-16Detection of cast members in video contentAbandonedUS20160358015A1 (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
US15/238,413US20160358015A1 (en)2013-04-102016-08-16Detection of cast members in video content

Applications Claiming Priority (2)

Application NumberPriority DateFiling DateTitle
US13/860,347US9449216B1 (en)2013-04-102013-04-10Detection of cast members in video content
US15/238,413US20160358015A1 (en)2013-04-102016-08-16Detection of cast members in video content

Related Parent Applications (1)

Application NumberTitlePriority DateFiling Date
US13/860,347ContinuationUS9449216B1 (en)2013-04-102013-04-10Detection of cast members in video content

Publications (1)

Publication NumberPublication Date
US20160358015A1true US20160358015A1 (en)2016-12-08

Family

ID=56896150

Family Applications (2)

Application NumberTitlePriority DateFiling Date
US13/860,347Active2033-12-03US9449216B1 (en)2013-04-102013-04-10Detection of cast members in video content
US15/238,413AbandonedUS20160358015A1 (en)2013-04-102016-08-16Detection of cast members in video content

Family Applications Before (1)

Application NumberTitlePriority DateFiling Date
US13/860,347Active2033-12-03US9449216B1 (en)2013-04-102013-04-10Detection of cast members in video content

Country Status (1)

CountryLink
US (2)US9449216B1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2020235724A1 (en)*2019-05-232020-11-26엘지전자 주식회사Display device
US20220269800A1 (en)*2019-07-302022-08-25Huawei Technologies Co., Ltd.Privacy protection method for electronic device and electronic device
US20220405689A1 (en)*2019-10-302022-12-22Sony Group CorporationInformation processing apparatus, information processing method, and program

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US9762950B1 (en)2013-09-172017-09-12Amazon Technologies, Inc.Automatic generation of network pages from extracted media content
JP6717606B2 (en)*2016-01-282020-07-01株式会社ブロードリーフ Work analysis support device, work analysis support method, and computer program
US11080316B1 (en)*2017-05-262021-08-03Amazon Technologies, Inc.Context-inclusive face clustering
CN109034040B (en)*2018-07-192021-11-23北京影谱科技股份有限公司Character recognition method, device, equipment and medium based on cast
SG10201807663PA (en)*2018-09-062020-04-29Nec Asia Pacific Pte LtdSystem and method for multi-layer potential associates discovery
CN113302620B (en)2018-11-132024-09-10辉达公司 Use machine learning models to determine associations between objects and people
CN110427816B (en)*2019-06-252023-09-08平安科技(深圳)有限公司Object detection method, device, computer equipment and storage medium
CN111444817B (en)*2020-03-242023-07-07咪咕文化科技有限公司Character image recognition method and device, electronic equipment and storage medium
WO2021240671A1 (en)*2020-05-272021-12-02三菱電機株式会社Gesture detection device and gesture detection method
US11695993B1 (en)*2020-10-052023-07-04America's Collectibles Network, Inc.System and method for creating and organizing content
US12347231B1 (en)*2022-03-312025-07-01Amazon Technologies, Inc.Headshot extraction and curation

Citations (17)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6055327A (en)*1997-07-172000-04-25Aragon; David BradburnMethod of detecting data entry errors by sorting amounts and verifying amount order
US6754389B1 (en)*1999-12-012004-06-22Koninklijke Philips Electronics N.V.Program classification using object tracking
US20050105806A1 (en)*2003-11-142005-05-19Yasuhiko NagaokaMethod and apparatus for organizing digital media based on face recognition
US20050129311A1 (en)*2003-12-112005-06-16Haynes Simon D.Object detection
US20100246999A1 (en)*2007-09-202010-09-30Michael TillbergMethod and Apparatus for Editing Large Quantities of Data Extracted from Documents
US20110064381A1 (en)*2009-09-152011-03-17Apple Inc.Method and apparatus for identifying video transitions
US7962467B2 (en)*2002-10-112011-06-14L-1 Secure Credentialing, Inc.Systems and methods for recognition of individuals using multiple biometric searches
US8077930B2 (en)*2007-04-132011-12-13Atg Advanced Swiss Technology Group AgMethod for recognizing content in an image sequence
US20120106806A1 (en)*2010-11-012012-05-03Microsoft CorporationFace Recognition in Video Content
US20120213490A1 (en)*2011-02-182012-08-23Google Inc.Facial detection, recognition and bookmarking in videos
US20130110870A1 (en)*2005-12-232013-05-02Digimarc CorporationMethods for identifying audio or video content
US20130251217A1 (en)*2008-04-022013-09-26Google Inc.Method and Apparatus to Incorporate Automatic Face Recognition in Digital Image Collections
US20130254816A1 (en)*2012-03-212013-09-26Sony CorporationTemporal video tagging and distribution
US20130322765A1 (en)*2012-06-042013-12-05Comcast Cable Communications, LlcData Recognition in Content
US8689255B1 (en)*2011-09-072014-04-01Imdb.Com, Inc.Synchronizing video content with extrinsic data
US20140270407A1 (en)*2013-03-142014-09-18Microsoft CorporationAssociating metadata with images in a personal image collection
US9218364B1 (en)*2011-01-282015-12-22Yahoo! Inc.Monitoring an any-image labeling engine

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US8392183B2 (en)*2006-04-252013-03-05Frank Elmo WeberCharacter-based automated media summarization
KR100804678B1 (en)*2007-01-042008-02-20삼성전자주식회사 How to classify scenes by video figures
JP2009048490A (en)*2007-08-212009-03-05Toshiba Corp Similar shot detection apparatus, program and method
US8705810B2 (en)*2007-12-282014-04-22Intel CorporationDetecting and indexing characters of videos by NCuts and page ranking
NO331287B1 (en)*2008-12-152011-11-14Cisco Systems Int Sarl Method and apparatus for recognizing faces in a video stream
JP2011019192A (en)*2009-07-102011-01-27Toshiba CorpImage display
US9271035B2 (en)*2011-04-122016-02-23Microsoft Technology Licensing, LlcDetecting key roles and their relationships from video
JP5857450B2 (en)*2011-05-302016-02-10ソニー株式会社 Information processing apparatus, information processing method, and program
US9179201B2 (en)*2011-08-262015-11-03Cyberlink Corp.Systems and methods of detecting significant faces in video streams

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6055327A (en)*1997-07-172000-04-25Aragon; David BradburnMethod of detecting data entry errors by sorting amounts and verifying amount order
US6754389B1 (en)*1999-12-012004-06-22Koninklijke Philips Electronics N.V.Program classification using object tracking
US7962467B2 (en)*2002-10-112011-06-14L-1 Secure Credentialing, Inc.Systems and methods for recognition of individuals using multiple biometric searches
US20050105806A1 (en)*2003-11-142005-05-19Yasuhiko NagaokaMethod and apparatus for organizing digital media based on face recognition
US20050129311A1 (en)*2003-12-112005-06-16Haynes Simon D.Object detection
US20130110870A1 (en)*2005-12-232013-05-02Digimarc CorporationMethods for identifying audio or video content
US8077930B2 (en)*2007-04-132011-12-13Atg Advanced Swiss Technology Group AgMethod for recognizing content in an image sequence
US20100246999A1 (en)*2007-09-202010-09-30Michael TillbergMethod and Apparatus for Editing Large Quantities of Data Extracted from Documents
US20130251217A1 (en)*2008-04-022013-09-26Google Inc.Method and Apparatus to Incorporate Automatic Face Recognition in Digital Image Collections
US20110064381A1 (en)*2009-09-152011-03-17Apple Inc.Method and apparatus for identifying video transitions
US20120106806A1 (en)*2010-11-012012-05-03Microsoft CorporationFace Recognition in Video Content
US9218364B1 (en)*2011-01-282015-12-22Yahoo! Inc.Monitoring an any-image labeling engine
US20120213490A1 (en)*2011-02-182012-08-23Google Inc.Facial detection, recognition and bookmarking in videos
US8689255B1 (en)*2011-09-072014-04-01Imdb.Com, Inc.Synchronizing video content with extrinsic data
US20130254816A1 (en)*2012-03-212013-09-26Sony CorporationTemporal video tagging and distribution
US20130322765A1 (en)*2012-06-042013-12-05Comcast Cable Communications, LlcData Recognition in Content
US20140270407A1 (en)*2013-03-142014-09-18Microsoft CorporationAssociating metadata with images in a personal image collection

Cited By (5)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2020235724A1 (en)*2019-05-232020-11-26엘지전자 주식회사Display device
US12238378B2 (en)2019-05-232025-02-25Lg Electronics Inc.Display device
US20220269800A1 (en)*2019-07-302022-08-25Huawei Technologies Co., Ltd.Privacy protection method for electronic device and electronic device
US12166916B2 (en)*2019-07-302024-12-10Huawei Technologies Co., Ltd.Privacy protection method for electronic device and electronic device
US20220405689A1 (en)*2019-10-302022-12-22Sony Group CorporationInformation processing apparatus, information processing method, and program

Also Published As

Publication numberPublication date
US9449216B1 (en)2016-09-20

Similar Documents

PublicationPublication DateTitle
US9449216B1 (en)Detection of cast members in video content
US10785510B2 (en)Automatic recognition of entities in media-captured events
US10580453B1 (en)Cataloging video and creating video summaries
US10970334B2 (en)Navigating video scenes using cognitive insights
US9154761B2 (en)Content-based video segmentation
CN111295708B (en) Voice recognition device and method of operating the same
US9934423B2 (en)Computerized prominent character recognition in videos
US20200265238A1 (en)Methods and Systems for Identification and Augmentation of Video Content
US10387556B2 (en)Displaying supplemental information about selected e-book objects
JP2014535122A (en) Face recognition using social networking information
US11894021B2 (en)Data processing method and system, storage medium, and computing device
JP6323202B2 (en) System, method and program for acquiring video
CN112118395B (en)Video processing method, terminal and computer readable storage medium
KR101986307B1 (en)Method and system of attention memory for locating an object through visual dialogue
CN115278326B (en) Video display method, device, computer readable medium and electronic device
US10015445B1 (en)Room conferencing system with heat map annotation of documents
US20240214640A1 (en)Systems and methods for displaying subjects of a video portion of content
US10949705B2 (en)Focalized behavioral measurements in a video stream
CN113190107B (en) Gesture recognition method, device and electronic device
US20230215170A1 (en)System and method for generating scores and assigning quality index to videos on digital platform
US11099811B2 (en)Systems and methods for displaying subjects of an audio portion of content and displaying autocomplete suggestions for a search related to a subject of the audio portion
CN112307823B (en)Method and device for labeling objects in video
CN110874554A (en)Action recognition method, terminal device, server, system and storage medium
CN103984415A (en)Information processing method and electronic equipment
US20210089781A1 (en)Systems and methods for displaying subjects of a video portion of content and displaying autocomplete suggestions for a search related to a subject of the video portion

Legal Events

DateCodeTitleDescription
STCBInformation on status: application discontinuation

Free format text:ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION


[8]ページ先頭

©2009-2025 Movatter.jp