Movatterモバイル変換


[0]ホーム

URL:


CN105989132A - Image file processing and speech controlling method - Google Patents

Image file processing and speech controlling method
Download PDF

Info

Publication number
CN105989132A
CN105989132ACN201510086642.2ACN201510086642ACN105989132ACN 105989132 ACN105989132 ACN 105989132ACN 201510086642 ACN201510086642 ACN 201510086642ACN 105989132 ACN105989132 ACN 105989132A
Authority
CN
China
Prior art keywords
data
information
account
personality
voice
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510086642.2A
Other languages
Chinese (zh)
Inventor
吴周亚
陈晓峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Chuqu Network Technology Co Ltd
Original Assignee
Shanghai Chuqu Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Chuqu Network Technology Co LtdfiledCriticalShanghai Chuqu Network Technology Co Ltd
Priority to CN201510086642.2ApriorityCriticalpatent/CN105989132A/en
Publication of CN105989132ApublicationCriticalpatent/CN105989132A/en
Pendinglegal-statusCriticalCurrent

Links

Landscapes

Abstract

The invention discloses an image file processing and speech controlling method. The speech controlling method comprises the steps of (S210) receiving speech information and extracting speech feature information and speech content information in the speech information; (S220) searching the combination of image data information, audio information and text information corresponding to the speech feature information; and (S230) sequentially displaying image data and text data and displaying audio data according to the speech content information. Through image file processing and speech controlling functions, an image can be achieved, augmented reality processing of the image can be achieved and interaction of virtuality and reality can also be achieved through speech control; 3D models learn from one another and influence one another; and higher interestingness and usability are strengthened for a user.

Description

Image file processes and sound control method
Technical field
The present invention relates to electronic information field, particularly relate to image file processing method and sound control method.
Background technology
Augmented reality (Augmented Reality is called for short AR), is the new technique grown up on the basis of virtual reality,Also mixed reality it is referred to as.It is that the information provided by computer system increases user's technology to real world perception, by voidThe Information application intended is to real world, and the dummy object, scene or the system prompt information superposition that are generated by computer are to true fieldJing Zhong, thus realize the enhancing to reality.
The function of augmented reality of the prior art, it is impossible to be controlled by voice, without from growing up and learning functionality.If passed throughImage file processes and voice control function, can realize image, not only can realize image enhaucament reality and process, also can realize passing throughVoice command realizes the interaction between virtual and reality, can mutually learn and affect, will increase to user higher between 3D modelInterest and ease for use.
Summary of the invention
It is an object of the invention to provide a kind of image file processing method and sound control method.
Image file processing method provided by the present invention, including: S110 acquisition records the first data of virtual image information;S120 obtains the second data recording real world images information;First data and the second data are combined process by S130, fromAnd obtain recording the 3rd data of virtual image and real world images.
Sound control method provided by the present invention, including: S210 receives voice messaging, and extracts the voice in this voice messagingCharacteristic information and voice content information;S220 find the image data information corresponding with this voice characteristics information, audio-frequency information,Text message combines;S230, according to voice content information, shows view data, text data and playing audio-fequency data successively.
The present invention is processed by image file and voice control function, can realize image, not only can realize image enhaucament reality and process,Also can realize realizing the interaction between virtual and reality by Voice command, can mutually learn and affect between 3D model, give and useFamily increases higher interest and ease for use.
Accompanying drawing explanation
Fig. 1 is the image file processing method schematic flow sheet described in the embodiment of the present invention one;
Fig. 2 is to obtain described in S120 described in the image file processing method described in the embodiment of the present invention one to record real world images letterThe schematic diagram of the step of the second data of breath;
Fig. 3 is the sound control method schematic flow sheet described in the embodiment of the present invention two;
Fig. 4 is to find the figure corresponding with this voice characteristics information described in the sound control method described in the embodiment of the present invention twoSchematic flow sheet as data message combination;
Fig. 5 is to show view data successively according to voice content information described in the sound control method described in the embodiment of the present invention twoSchematic flow sheet;
Fig. 6 is the schematic flow sheet of information feedback step in the sound control method described in the embodiment of the present invention two;
Fig. 7 is the schematic flow sheet of the sound control method learning step described in the embodiment of the present invention two;
Fig. 8 is to set up the first personality data and second matched with account in the sound control method described in the embodiment of the present invention twoThe schematic flow sheet of personality data.
Detailed description of the invention
For making the purpose of the embodiment of the present invention, technical scheme and advantage clearer, attached below in conjunction with in the embodiment of the present inventionFigure, is clearly and completely described the technical scheme in the embodiment of the present invention, it is clear that described embodiment is the present inventionA part of embodiment rather than whole embodiments.Based on the embodiment in the present invention, those of ordinary skill in the art are not havingMake the every other embodiment obtained under creative work premise, broadly fall into the scope of protection of the invention.
Embodiment one
As it is shown in figure 1, the present embodiment provides a kind of image file processing method, including:
S110 obtains the first data recording virtual image information;It will be understood by those skilled in the art that described virtual image isRefer to the image obtained by drafting, can be plane picture, it is also possible to be the stereo-picture presenting stereoeffect, such as mountain valley, outwardSpaces etc. have the scene image of three-dimensional sensorial effects;
S120 obtains the second data recording real world images information;It will be understood by those skilled in the art that described real world images isReferring to the image obtained by shooting, the validity of its image is more nearly reality, can be object, animal, face or human bodyImage;
First data and the second data are combined process by S130, thus obtain recording the 3rd of virtual image and real world imagesData.
Further, described combination processes and refers to that augmented reality (Augmented Reality is called for short AR) processes.Art technologyPersonnel are appreciated that described augmented reality, are the new techniques grown up on the basis of virtual reality, are also referred to as mixingReality.It is that the information provided by computer system increases user's technology to real world perception, virtual Information application is arrivedReal world, and dummy object, scene or the system prompt information superposition generated by computer is in real scene, thus realizeEnhancing to reality.
As in figure 2 it is shown, obtain the step of the second data recording real world images information described in S120, including:
First data of the virtual image information that records of storage in S121 reading equipment, and calculate mesh according to described first dataMark characteristic;
S122 identifies the first data;
S123 finds the part meeting target data feature as the second number recording real world images information in the data identifiedAccording to.
It will be understood by those skilled in the art that described target characteristic data include form parameter, range parameter, size parameter, divideResolution parameter and colorimetric parameter, those skilled in the art can be and institute according to which part that target characteristic data identify in photoState the first data to match, augmented reality process can be carried out, can be treated to be the part of the second data.Such as, if first countsAccording to being the neck drawing image with lower part, then with this first data match, augmented reality process can be carried out and canThe part being processed as the second data is the face range data that in photo, definition reaches this first data picture parameter request, i.e. thisTime the second data be shooting face information data.
Embodiment two
As it is shown on figure 3, the present embodiment provides a kind of sound control method, including:
S210 receives voice messaging, and extracts the voice characteristics information in this voice messaging and voice content information;Art technologyPersonnel are appreciated that described voice characteristics information includes pitch parameters and tamber parameter information, according to this voice characteristics information, and canIdentify the user of this this voice of input;
S220 finds the image data information corresponding with this voice characteristics information, audio-frequency information, text message combination;
S230, according to voice content information, shows view data, text data and playing audio-fequency data successively.
The combination of described image data information refers to the image data information combination with identical second data described in embodiment one.ThisSample can realize when inputting specific voice messaging, and the figure sector-meeting with identical real world images information is entered successively according to voice contentRow display.
As shown in Figure 4, described in find the combination of corresponding with this voice characteristics information image data information, including:
S221 sets up the matching relationship of view data combination and account;
S222 sets up the matching relationship of voice characteristics information and account;
Voice characteristics information is identified by S223, finds the account matched with this voice characteristics information, further according to the accountFind the view data combination matched with the account.
So, when user is by mike input voice, after system receives this voice messaging, by the tone in voice messagingParameter and tamber parameter extract, thus obtain pitch parameters and tamber parameter, can recognize that according to pitch parameters and tamber parameterGo out this user and find the account mated with this user, thus finding the view data combination mated with the account, as searchedTo all view data with this user's head portrait, including the process described in embodiment one, head image data and contextual data are passed throughAll 3rd data that augmented reality processes and obtains.
As it is shown in figure 5, described according to voice content information, show view data successively, including:
S231 sets up and the matching relationship of voice content information with the 3rd data;
Voice content information is identified by S232, finds the 3rd data with this voice content information match;
S233 shows the 3rd data.Such as when voice content is for dancing, inquire and dance the pictorial information matched, and rootAccording to being set for display, thus realize role's dancing that user utilizes Voice command to set.It addition, according to described in the present embodimentSound control method, it is also possible to realize virtual 3D house pet voice interface, or participate in virtual war, fight etc..
As shown in Figure 6, the sound control method described in the present embodiment, also include S240 information feedback step, specifically include:
S241 sets up the matching relationship of voice content information and account;
S242 statistics voice content information also goes out individual information according to voice content information analysis;
Set up the matching relationship of individual information and account;It will be understood by those skilled in the art that when the second data are a 3D pattern numberAccording to time, can be by repeatedly performing this information feedback step, it is achieved during user is mutual with equipment, accidentally, systemDrawing user personality by big data analysis, and individual character is given to 3D model, after permanent use, the feedback of 3D model willBecome closer to the individual character of user.
As it is shown in fig. 7, the sound control method described in the present embodiment, also include S250 learning procedure, specifically include:
S251 sets up the first personality data and the second personality data matched with account;
S252 set up with the first personality scale parameter of the first personality data match and with the second of the second personality data matchPersonality scale parameter, described first personality scale parameter and described second personality scale parameter and be 1;
S253 sets up the relation with increase of the first personality scale parameter and time;
S254 updates the first personality scale parameter over time.The most every 24 hours, described first personality scale parameter increased by 0.05.It will be understood by those skilled in the art that so when the second data are a 3D model data, its first personality being had will be along withThe growth of time and increase, it is had the second personality and growth is in time reduced, thus realizes the growth of 3D model.
As shown in Figure 8, described S251 sets up the first personality data and the second personality data matched with account, specifically includes:
S2511 sets up the matching relationship between the first account and the second account;
S2512 using the individual information of the second account as the first personality data of the first account;And the individual information of the first account is madeIt is the second personality data of the first account;It will be understood by those skilled in the art that the individual information of described first account is according toThe voice content Information Statistics that one account matches are analyzed and are obtained;The individual information of described second account according to the second account phaseThe voice content Information Statistics of coupling are analyzed and are obtained;So, growth over time, the first personality number of described first accountAccording to increasing, the second view data i.e. mated with the first account will have the individual character of increasing second account;Along with timeBetween growth, the second personality data of the first account can reduce, the second view data i.e. mated with the first account oneselfProperty can be fewer and feweri.The second view data as matched with the first account is maiden's photo, the voice matched with the first accountContent information is dancing content;The second view data matched with the second account is man's photo, the language mated with the second accountSound content information is battlefield command content;As time goes on, the maiden of the first account coupling will more and more demonstrate fingerWave the action of war, and less and less demonstrate dance movement, thus realize the second view data and second account of the first accountSecond view data at family learns mutually.By the sound control method described in the present embodiment, also can realize between the second view dataInfluence each other, such as user has " Sun Wukong " 3D model, carry out augmented reality experience time, " the pig of other usersEight guard against " model enter into this user augmented reality experience in, two 3D models with the personality data of respective user can be carried outAutomatic speech or word are interactive, influence each other in this process.
Last it is noted that above example is only in order to illustrate technical scheme, it is not intended to limit;Although referenceThe present invention has been described in detail by previous embodiment, it will be understood by those within the art that: it still can be to frontState the technical scheme described in each embodiment to modify, or wherein portion of techniques feature is carried out equivalent;And these are repaiiedChange or replace, not making the essence of appropriate technical solution depart from the spirit and scope of various embodiments of the present invention technical scheme.

Claims (10)

CN201510086642.2A2015-02-172015-02-17Image file processing and speech controlling methodPendingCN105989132A (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN201510086642.2ACN105989132A (en)2015-02-172015-02-17Image file processing and speech controlling method

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN201510086642.2ACN105989132A (en)2015-02-172015-02-17Image file processing and speech controlling method

Publications (1)

Publication NumberPublication Date
CN105989132Atrue CN105989132A (en)2016-10-05

Family

ID=57038416

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN201510086642.2APendingCN105989132A (en)2015-02-172015-02-17Image file processing and speech controlling method

Country Status (1)

CountryLink
CN (1)CN105989132A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN107185244A (en)*2017-05-242017-09-22上海悦游网络信息科技股份有限公司Pet combat system in one kind control game special scenes
CN108063936A (en)*2017-12-292018-05-22光锐恒宇(北京)科技有限公司Implementation method, device and the computer readable storage medium of augmented reality AR
CN114467140A (en)*2020-08-052022-05-10互动解决方案公司System for changing image based on voice

Citations (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN102495959A (en)*2011-12-052012-06-13无锡智感星际科技有限公司Augmented reality (AR) platform system based on position mapping and application method
CN103202010A (en)*2010-11-092013-07-10富士胶片株式会社 Augmented Reality Provider
CN103365854A (en)*2012-03-282013-10-23鸿富锦精密工业(深圳)有限公司System and method for retrieving video files

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN103202010A (en)*2010-11-092013-07-10富士胶片株式会社 Augmented Reality Provider
CN102495959A (en)*2011-12-052012-06-13无锡智感星际科技有限公司Augmented reality (AR) platform system based on position mapping and application method
CN103365854A (en)*2012-03-282013-10-23鸿富锦精密工业(深圳)有限公司System and method for retrieving video files

Cited By (5)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN107185244A (en)*2017-05-242017-09-22上海悦游网络信息科技股份有限公司Pet combat system in one kind control game special scenes
CN108063936A (en)*2017-12-292018-05-22光锐恒宇(北京)科技有限公司Implementation method, device and the computer readable storage medium of augmented reality AR
CN108063936B (en)*2017-12-292020-11-03光锐恒宇(北京)科技有限公司Method and device for realizing augmented reality AR and computer readable storage medium
CN114467140A (en)*2020-08-052022-05-10互动解决方案公司System for changing image based on voice
US11568877B2 (en)2020-08-052023-01-31Interactive Solutions Corp.System to change image based on voice

Similar Documents

PublicationPublication DateTitle
US12273600B2 (en)Intelligent commentary generation and playing methods, apparatuses, and devices, and computer storage medium
JP7253570B2 (en) Contextual in-game element recognition, annotation and interaction based on remote user input
CN107203953B (en)Teaching system based on internet, expression recognition and voice recognition and implementation method thereof
CN110652726B (en)Game auxiliary system based on image recognition and audio recognition
EP4184927A1 (en)Sound effect adjusting method and apparatus, device, storage medium, and computer program product
CN111556278A (en)Video processing method, video display device and storage medium
CN109637207A (en)A kind of preschool education interactive teaching device and teaching method
CN109739353A (en)A kind of virtual reality interactive system identified based on gesture, voice, Eye-controlling focus
CN109343695A (en)Exchange method and system based on visual human's behavioral standard
CN111027419A (en)Method, device, equipment and medium for detecting video irrelevant content
CN105989132A (en)Image file processing and speech controlling method
CN112764530A (en)Ammunition identification method based on touch handle and augmented reality glasses
WO2023216502A1 (en)Display control method and apparatus in game, storage medium and electronic device
CN104933278A (en)Multi-channel interactive method and system used for speech disorder rehabilitation training
CN112764531A (en)Augmented reality ammunition identification method
WO2018188088A1 (en)Clay toy system based on augmented reality and digital image processing and method therefor
CN112637692A (en)Interaction method, device and equipment
KR102482841B1 (en)Artificial intelligence mirroring play bag
Jung et al.Visual attention never fades: Selective progressive attention recalibration for detailed image captioning in multimodal large language models
Li et al.Multimodal causal reasoning benchmark: Challenging vision large language models to discern causal links across modalities
WO2023065963A1 (en)Interactive display method and apparatus, electronic device, and storage medium
CN114425162A (en)Video processing method and related device
CN114120338B (en) Hand-drawn trajectory recognition method and device, storage medium, and electronic device
ManuDevelopment Of Instagram Filter Using Spark AR In An Effort To Preserve Kupang Malay Language.
Yohannes et al.Virtual reality in puppet game using depth sensor of gesture recognition and tracking

Legal Events

DateCodeTitleDescription
C06Publication
PB01Publication
C10Entry into substantive examination
SE01Entry into force of request for substantive examination
WD01Invention patent application deemed withdrawn after publication

Application publication date:20161005

WD01Invention patent application deemed withdrawn after publication

[8]ページ先頭

©2009-2025 Movatter.jp