Movatterモバイル変換


[0]ホーム

URL:


US20070126926A1 - Hybrid-captioning system - Google Patents

Hybrid-captioning system
Download PDF

Info

Publication number
US20070126926A1
US20070126926A1US11/294,234US29423405AUS2007126926A1US 20070126926 A1US20070126926 A1US 20070126926A1US 29423405 AUS29423405 AUS 29423405AUS 2007126926 A1US2007126926 A1US 2007126926A1
Authority
US
United States
Prior art keywords
captions
video
subsystem
spoken utterances
line
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/294,234
Inventor
Kohtaroh Miyamoto
Noriko Negishi
Kenichi Arakawa
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by IndividualfiledCriticalIndividual
Priority to US11/294,234priorityCriticalpatent/US20070126926A1/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATIONreassignmentINTERNATIONAL BUSINESS MACHINES CORPORATIONASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: ARAKAWA, KENICHI, MIYAMOTO, KOHTAROH, NEGISHI, NORIKO
Publication of US20070126926A1publicationCriticalpatent/US20070126926A1/en
Priority to US12/172,260prioritypatent/US8311832B2/en
Abandonedlegal-statusCriticalCurrent

Links

Images

Classifications

Definitions

Landscapes

Abstract

A hybrid-captioning system for editing captions for spoken utterances within video includes an editor-type caption-editing subsystem, a line-based caption-editing subsystem, and a mechanism. The editor-type subsystem is that in which captions are edited for spoken utterances within the video on a groups-of-line basis without respect to particular lines of the captions and without respect to temporal positioning of the captions in relation to the spoken utterances. The line-based subsystem is that in which captions are edited for spoken utterances within the video on a line-by-line basis with respect to particular lines of the captions and with respect to temporal positioning of the captions in relation to the spoken utterances. For each section of spoken utterances within the video, the mechanism is to select the editor-type or the line-based subsystem to provide captions for the section of spoken utterances in accordance with a predetermined criteria.

Description

Claims (20)

1. A hybrid-captioning system to edit captions for spoken utterances within video comprising:
an editor-type caption-editing subsystem in which captions are edited for spoken utterances within the video on a groups-of-lines basis without respect to particular lines of the captions and without respect to temporal positioning of the captions in relation to the spoken utterances;
a line-based caption-editing subsystem in which captions are edited for spoken utterances within the video on a line-by-line basis with respect to particular lines of the captions and with respect to temporal positioning of the captions in relation to the spoken utterances; and,
a mechanism to, for each section of spoken utterances within the video, select the editor-type caption-editing subsystem or the line-based caption-editing subsystem to provide captions for the section of spoken utterances in accordance with a predetermined criteria.
9. A method comprising:
in relation to video for which captions are to be edited, receiving user input as to a current section of the video for which captions have been generated, within an editor-type caption-editing subsystem in which captions are edited for spoken utterances within the video on a groups-of-lines basis without respect to particular lines of the captions and without respect to temporal positioning of the captions in relation to the spoken utterances;
where the user input corresponds to termination of the editor-type caption editing subsystem,
transmitting the captions generated for the current section to a general-matching subsystem;
the general-matching subsystem transmitting the captions generated for the current section to a line-based caption-editing subsystem in which captions are edited for spoken utterances within the video on a line-by-line basis with respect to particular lines of the captions and with respect to temporal positioning of the captions in relation to the spoken utterances;
otherwise,
transmitting the captions generated for the current section to a particular-matching subsystem; and,
the particular-matching subsystem transmitting the captions generated for the current section back to the editor-type caption-editing subsystem.
15. The method ofclaim 9, further comprising, after the captions generated for the current section have been transmitted to the particular-matching subsystem, the particular-matching subsystem performing:
determining a voice-recognition rate of the captions that have been generated for the current section of the video;
where the voice-recognition rate is greater than a threshold, returning the captions as one or more predicted character strings;
otherwise, where the voice-recognition rate is not greater than the threshold,
performing phoneme-based character matching as to the captions that have been generated for the current section;
where the phoneme-based character matching results in a match value greater than a predetermined threshold,
returning the captions as the one or more predicted character strings; and,
otherwise, where the match value is not greater than the predetermined threshold, returning indication that no matching has occurred.
19. An article of manufacture comprising:
a tangible recordable data storage medium; and,
means in the medium for selecting an editor-type caption-editing subsystem or a line-based caption-editing subsystem to provide captions for each of a plurality of sections of spoken utterances of video, in accordance with a predetermined criteria,
wherein the editor-type caption-editing subsystem is that in which captions are edited for spoken utterances within the video on a groups-of-lines basis without respect to particular lines of the captions and without respect to temporal positioning of the captions in relation to the spoken utterances, and
where the line-based caption-editing subsystem is that in which captions are edited for spoken utterances within the video on a line-by-line basis with respect to particular lines of the captions and with respect to temporal positioning of the captions in relation to the spoken utterances.
US11/294,2342005-12-042005-12-04Hybrid-captioning systemAbandonedUS20070126926A1 (en)

Priority Applications (2)

Application NumberPriority DateFiling DateTitle
US11/294,234US20070126926A1 (en)2005-12-042005-12-04Hybrid-captioning system
US12/172,260US8311832B2 (en)2005-12-042008-07-13Hybrid-captioning system

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
US11/294,234US20070126926A1 (en)2005-12-042005-12-04Hybrid-captioning system

Related Child Applications (1)

Application NumberTitlePriority DateFiling Date
US12/172,260ContinuationUS8311832B2 (en)2005-12-042008-07-13Hybrid-captioning system

Publications (1)

Publication NumberPublication Date
US20070126926A1true US20070126926A1 (en)2007-06-07

Family

ID=38118345

Family Applications (2)

Application NumberTitlePriority DateFiling Date
US11/294,234AbandonedUS20070126926A1 (en)2005-12-042005-12-04Hybrid-captioning system
US12/172,260Expired - Fee RelatedUS8311832B2 (en)2005-12-042008-07-13Hybrid-captioning system

Family Applications After (1)

Application NumberTitlePriority DateFiling Date
US12/172,260Expired - Fee RelatedUS8311832B2 (en)2005-12-042008-07-13Hybrid-captioning system

Country Status (1)

CountryLink
US (2)US20070126926A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20070048715A1 (en)*2004-12-212007-03-01International Business Machines CorporationSubtitle generation and retrieval combining document processing with voice processing
US20080151111A1 (en)*2006-12-212008-06-26Samsung Electronics Co., Ltd.Broadcast receiving apparatus and method for storing open caption information
US20130080163A1 (en)*2011-09-262013-03-28Kabushiki Kaisha ToshibaInformation processing apparatus, information processing method and computer program product
US20180144747A1 (en)*2016-11-182018-05-24Microsoft Technology Licensing, LlcReal-time caption correction by moderator
US10192554B1 (en)*2018-02-262019-01-29Sorenson Ip Holdings, LlcTranscription of communications using multiple speech recognition systems
US20220014623A1 (en)*2014-02-282022-01-13Ultratec, Inc.Semiautomated relay method and apparatus
US11627221B2 (en)2014-02-282023-04-11Ultratec, Inc.Semiautomated relay method and apparatus
US11741963B2 (en)2014-02-282023-08-29Ultratec, Inc.Semiautomated relay method and apparatus
US12035070B2 (en)2020-02-212024-07-09Ultratec, Inc.Caption modification and augmentation systems and methods for use by hearing assisted user

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US8345159B2 (en)*2007-04-162013-01-01Caption Colorado L.L.C.Captioning evaluation system
US8707381B2 (en)2009-09-222014-04-22Caption Colorado L.L.C.Caption and/or metadata synchronization for replay of previously or simultaneously recorded live programs
US8364463B2 (en)*2009-09-252013-01-29International Business Machines CorporationOptimizing a language/media translation map
US8826354B2 (en)*2010-12-012014-09-02At&T Intellectual Property I, L.P.Method and system for testing closed caption content of video assets
US9471334B2 (en)*2013-03-082016-10-18Intel CorporationContent presentation with enhanced closed caption and/or skip back
US9788071B2 (en)2014-11-032017-10-10Microsoft Technology Licensing, LlcAnnotating and indexing broadcast video for searchability
US9922095B2 (en)2015-06-022018-03-20Microsoft Technology Licensing, LlcAutomated closed captioning using temporal data
JP6605995B2 (en)2016-03-162019-11-13株式会社東芝 Speech recognition error correction apparatus, method and program
JP6678545B2 (en)2016-09-122020-04-08株式会社東芝 Correction system, correction method and program
JP6672209B2 (en)2017-03-212020-03-25株式会社東芝 Information processing apparatus, information processing method, and information processing program
KR102452644B1 (en)*2017-10-312022-10-11삼성전자주식회사Electronic apparatus, voice recognition method and storage medium
KR102468214B1 (en)*2018-02-192022-11-17삼성전자주식회사The system and an appratus for providig contents based on a user utterance
EP3841754A4 (en)*2018-09-132022-06-15iChannel.io Ltd.A system and computerized method for subtitles synchronization of audiovisual content using the human voice detection for synchronization
US10885903B1 (en)*2018-12-102021-01-05Amazon Technologies, Inc.Generating transcription information based on context keywords
KR102722495B1 (en)*2020-02-062024-10-29삼성전자주식회사Electronice device and control method thereof
US11032620B1 (en)*2020-02-142021-06-08Sling Media Pvt LtdMethods, systems, and apparatuses to respond to voice requests to play desired video clips in streamed media based on matched close caption and sub-title text
CA3223764A1 (en)*2021-05-102022-11-17Brandon WrightManaging content quality and related characteristics of a media playback system
US11683558B2 (en)*2021-06-292023-06-20The Nielsen Company (Us), LlcMethods and apparatus to determine the speed-up of media programs using speech recognition
US11785278B1 (en)*2022-03-182023-10-10Comcast Cable Communications, LlcMethods and systems for synchronization of closed captions with content output
JP2025051640A (en)*2023-09-222025-04-04中強光電股▲ふん▼有限公司 Electronic system and control method thereof
US12323647B2 (en)*2023-11-102025-06-03Avago Technologies International Sales Pte. LimitedVideo quality monitoring system

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US5729741A (en)*1995-04-101998-03-17Golden Enterprises, Inc.System for storage and retrieval of diverse types of information obtained from different media sources which includes video, audio, and text transcriptions
US5982448A (en)*1997-10-301999-11-09Reyes; Frances S.Multi-language closed captioning system
US6571210B2 (en)*1998-11-132003-05-27Microsoft CorporationConfidence measure system using a near-miss pattern
US6871107B1 (en)*1999-07-012005-03-22Ftr Pty, Ltd.Digital audio transcription system
US6505153B1 (en)*2000-05-222003-01-07Compaq Information Technologies Group, L.P.Efficient method for producing off-line closed captions
US6453294B1 (en)*2000-05-312002-09-17International Business Machines CorporationDynamic destination-determined multimedia avatars for interactive on-line communications
US7191117B2 (en)*2000-06-092007-03-13British Broadcasting CorporationGeneration of subtitles or captions for moving pictures
US6580437B1 (en)*2000-06-262003-06-17Siemens Corporate Research, Inc.System for organizing videos based on closed-caption information
WO2002009093A1 (en)*2000-07-202002-01-31Koninklijke Philips Electronics N.V.Feedback of recognized command confidence level
US7130790B1 (en)*2000-10-242006-10-31Global Translations, Inc.System and method for closed caption data translation
US7117231B2 (en)*2000-12-072006-10-03International Business Machines CorporationMethod and system for the automatic generation of multi-lingual synchronized sub-titles for audiovisual data
US7013273B2 (en)*2001-03-292006-03-14Matsushita Electric Industrial Co., Ltd.Speech recognition based captioning system
KR20030065332A (en)*2002-01-312003-08-06톰슨 라이센싱 소시에떼 아노님Method for processing auxiliary information in a video system
US7386454B2 (en)*2002-07-312008-06-10International Business Machines CorporationNatural error handling in speech recognition
JP2004302175A (en)2003-03-312004-10-28Fuji Television Network IncSystem, method, and program for speech recognition
JP3945778B2 (en)*2004-03-122007-07-18インターナショナル・ビジネス・マシーンズ・コーポレーション Setting device, program, recording medium, and setting method
JP4536481B2 (en)*2004-10-252010-09-01インターナショナル・ビジネス・マシーンズ・コーポレーション Computer system, method for supporting correction work, and program
JP4218758B2 (en)*2004-12-212009-02-04インターナショナル・ビジネス・マシーンズ・コーポレーション Subtitle generating apparatus, subtitle generating method, and program
US20070011012A1 (en)*2005-07-112007-01-11Steve YurickMethod, system, and apparatus for facilitating captioning of multi-media content
JP4158937B2 (en)*2006-03-242008-10-01インターナショナル・ビジネス・マシーンズ・コーポレーション Subtitle correction device
US20080270110A1 (en)*2007-04-302008-10-30Yurick Steven JAutomatic speech recognition with textual content input
US8229748B2 (en)*2008-04-142012-07-24At&T Intellectual Property I, L.P.Methods and apparatus to present a video program to a visually impaired person
US20100332214A1 (en)*2009-06-302010-12-30Shpalter ShaharSystem and method for network transmision of subtitles

Cited By (18)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20070048715A1 (en)*2004-12-212007-03-01International Business Machines CorporationSubtitle generation and retrieval combining document processing with voice processing
US7739116B2 (en)*2004-12-212010-06-15International Business Machines CorporationSubtitle generation and retrieval combining document with speech recognition
US20080151111A1 (en)*2006-12-212008-06-26Samsung Electronics Co., Ltd.Broadcast receiving apparatus and method for storing open caption information
US20130080163A1 (en)*2011-09-262013-03-28Kabushiki Kaisha ToshibaInformation processing apparatus, information processing method and computer program product
US9798804B2 (en)*2011-09-262017-10-24Kabushiki Kaisha ToshibaInformation processing apparatus, information processing method and computer program product
US20220014622A1 (en)*2014-02-282022-01-13Ultratec, Inc.Semiautomated relay method and apparatus
US20220014623A1 (en)*2014-02-282022-01-13Ultratec, Inc.Semiautomated relay method and apparatus
US11627221B2 (en)2014-02-282023-04-11Ultratec, Inc.Semiautomated relay method and apparatus
US11741963B2 (en)2014-02-282023-08-29Ultratec, Inc.Semiautomated relay method and apparatus
US20230352028A1 (en)*2014-02-282023-11-02Ultratec, Inc.Semiautomated relay method and apparatus
US12136426B2 (en)2014-02-282024-11-05Ultratec, Inc.Semiautomated relay method and apparatus
US12137183B2 (en)2014-02-282024-11-05Ultratec, Inc.Semiautomated relay method and apparatus
US12136425B2 (en)2014-02-282024-11-05Ultratec, Inc.Semiautomated relay method and apparatus
US12400660B2 (en)*2014-02-282025-08-26Ultratec, Inc.Semiautomated relay method and apparatus
US20180144747A1 (en)*2016-11-182018-05-24Microsoft Technology Licensing, LlcReal-time caption correction by moderator
US10192554B1 (en)*2018-02-262019-01-29Sorenson Ip Holdings, LlcTranscription of communications using multiple speech recognition systems
US11710488B2 (en)2018-02-262023-07-25Sorenson Ip Holdings, LlcTranscription of communications using multiple speech recognition systems
US12035070B2 (en)2020-02-212024-07-09Ultratec, Inc.Caption modification and augmentation systems and methods for use by hearing assisted user

Also Published As

Publication numberPublication date
US8311832B2 (en)2012-11-13
US20080270134A1 (en)2008-10-30

Similar Documents

PublicationPublication DateTitle
US8311832B2 (en)Hybrid-captioning system
US6442518B1 (en)Method for refining time alignments of closed captions
US6792409B2 (en)Synchronous reproduction in a speech recognition system
US9774747B2 (en)Transcription system
EP0899719B1 (en)Method for aligning text with audio signals
US8560327B2 (en)System and method for synchronizing sound and manually transcribed text
HazenAutomatic alignment and error correction of human generated transcripts for long speech recordings.
US7881930B2 (en)ASR-aided transcription with segmented feedback training
US8155958B2 (en)Speech-to-text system, speech-to-text method, and speech-to-text program
US20130035936A1 (en)Language transcription
US20070011012A1 (en)Method, system, and apparatus for facilitating captioning of multi-media content
US20020161582A1 (en)Method and apparatus for presenting images representative of an utterance with corresponding decoded speech
US20140180667A1 (en)System and method for real-time multimedia reporting
JP2009047920A (en) Apparatus and method for interacting with user by voice
US20050114131A1 (en)Apparatus and method for voice-tagging lexicon
EP3839953A1 (en)Automatic caption synchronization and positioning
Demuynck et al.A comparison of different approaches to automatic speech segmentation
JP6637332B2 (en) Spoken language corpus generation device and program thereof
US6377921B1 (en)Identifying mismatches between assumed and actual pronunciations of words
JP2017167318A (en) Minutes generating device and minutes generating program
JP2004343488A (en)Method, system, and program for inserting caption
CN110853627A (en)Method and system for voice annotation
KR101996551B1 (en)Apparatus and method for generating subtitles using speech recognition and script
Lanchantin et al.The development of the Cambridge University alignment systems for the Multi-Genre Broadcast challenge
JP4736478B2 (en) Voice transcription support device, method and program thereof

Legal Events

DateCodeTitleDescription
ASAssignment

Owner name:INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MIYAMOTO, KOHTAROH;NEGISHI, NORIKO;ARAKAWA, KENICHI;REEL/FRAME:017323/0682

Effective date:20051202

STCBInformation on status: application discontinuation

Free format text:ABANDONED -- FAILURE TO PAY ISSUE FEE


[8]ページ先頭

©2009-2025 Movatter.jp