Movatterモバイル変換


[0]ホーム

URL:


CN109754781B - Voice translation terminal, mobile terminal, translation system, translation method and device thereof - Google Patents

Voice translation terminal, mobile terminal, translation system, translation method and device thereof
Download PDF

Info

Publication number
CN109754781B
CN109754781BCN201910173337.5ACN201910173337ACN109754781BCN 109754781 BCN109754781 BCN 109754781BCN 201910173337 ACN201910173337 ACN 201910173337ACN 109754781 BCN109754781 BCN 109754781B
Authority
CN
China
Prior art keywords
voice information
translation
voice
language
mobile terminal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910173337.5A
Other languages
Chinese (zh)
Other versions
CN109754781A (en
Inventor
周庐
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Kingsoft Internet Security Software Co Ltd
Original Assignee
Beijing Kingsoft Internet Security Software Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Kingsoft Internet Security Software Co LtdfiledCriticalBeijing Kingsoft Internet Security Software Co Ltd
Priority to CN201910173337.5ApriorityCriticalpatent/CN109754781B/en
Publication of CN109754781ApublicationCriticalpatent/CN109754781A/en
Priority to JP2020000107Uprioritypatent/JP3225682U/en
Priority to US16/801,342prioritypatent/US20200285815A1/en
Application grantedgrantedCritical
Publication of CN109754781BpublicationCriticalpatent/CN109754781B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Landscapes

Abstract

Translated fromChinese

本申请提出一种语音翻译终端、移动终端、翻译系统、翻译方法及其装置,其中,语音翻译终端包括:壳体、控制器、一个触发按键、传声器组、扬声器及通信组件;壳体形成空间用于容纳控制器、传声器组、扬声器及通信组件;一个触发按键设置在壳体上,传声器组用于在触发按键被触发后,采集第一语音信息;扬声器用于根据控制器的控制,播放第二语音信息;其中,第二语音信息为第一语音信息翻译后的语音信息;控制器用于控制通信组件向移动终端发送第一语音信息,控制通信组件接收由移动终端发送的第二语音信息。由此,用户无需在语音翻译终端上设置对应的按键来确定第一语音信息的语种,可以简化用户的操作步骤。

The present application proposes a voice translation terminal, a mobile terminal, a translation system, a translation method and a device thereof, wherein the voice translation terminal comprises: a housing, a controller, a trigger button, a microphone group, a loudspeaker and a communication component; the housing forms a space for accommodating the controller, the microphone group, the loudspeaker and the communication component; a trigger button is arranged on the housing, and the microphone group is used to collect the first voice information after the trigger button is triggered; the loudspeaker is used to play the second voice information according to the control of the controller; wherein the second voice information is the voice information translated from the first voice information; the controller is used to control the communication component to send the first voice information to the mobile terminal, and control the communication component to receive the second voice information sent by the mobile terminal. Thus, the user does not need to set a corresponding button on the voice translation terminal to determine the language of the first voice information, which can simplify the user's operation steps.

Description

Voice translation terminal, mobile terminal, translation system, translation method and device thereof
Technical Field
The present application relates to the field of speech recognition technologies, and in particular, to a speech translation terminal, a mobile terminal, a translation system, a translation method, and a device thereof.
Background
With the rapid development of economy and the increasing frequency of international communications, users may have a need for use in two or even more languages. At present, the voice translation terminal is popular with users with language translation requirements by virtue of a powerful language translation function. In the process of dialogue, the voice translation terminal is used for translation, so that users using different languages can communicate without barriers.
In the related art, the speech translation terminal can translate a specified language (e.g. english) into a specified target language (e.g. chinese), so that the speech translation terminal has at least two keys, one for determining the language spoken by the user and one for triggering collection of speech data. When the user uses the voice translation terminal to translate, the automatic translation from the language spoken by the user to another language can be realized.
However, when the user uses the voice translation terminal to translate, the user needs to press two keys successively, and the operation steps are complicated.
Disclosure of Invention
The application provides a voice translation terminal, a mobile terminal, a translation system, a translation method and a device thereof, which are used for realizing automatic translation of first voice information into second voice information through a server, and determining languages of the first voice information without setting corresponding keys on the voice translation terminal.
An embodiment of a first aspect of the present application provides a speech translation terminal, including:
the device comprises a shell, a controller, a trigger button, a microphone set, a loudspeaker and a communication assembly;
The housing forming space is used for accommodating the controller, the microphone set, the loudspeaker and the communication component;
the trigger key is arranged on the shell and is electrically connected with the controller;
the microphone set is electrically connected with the controller and is used for collecting first voice information after the trigger key is triggered;
The loudspeaker is electrically connected with the controller and used for playing second voice information according to the control of the controller, wherein the second voice information is the voice information translated by the first voice information;
the controller is electrically connected with the communication component and is used for controlling the communication component to send the first voice information to the mobile terminal and controlling the communication component to receive the second voice information sent by the mobile terminal.
As a first possible implementation manner of the embodiment of the first aspect of the present application, the speech translation terminal communicates with the mobile terminal through the communication component, the mobile terminal communicates with the server, the speech translation terminal sends the first speech information to the mobile terminal through the communication component, after the first speech information is preprocessed by the mobile terminal, the first speech information is sent to the server through the mobile terminal, so that the server translates the first speech information to obtain the second speech information according to a translation setting of the mobile terminal, the second speech information is sent to the mobile terminal through the server, and the speech translation terminal receives the second speech information sent by the mobile terminal through the communication component.
As a second possible implementation manner of the embodiment of the first aspect of the present application, the microphone set includes two microphones, one for collecting voice data input by a user, and the other for collecting noise data;
And the microphone set is used for subtracting and amplifying the voice data and the noise data to obtain the first voice information.
As a third possible implementation manner of the embodiment of the first aspect of the present application, the controller is configured to:
Receiving first voice information acquired by the microphone set, and compressing and subpackaging the first voice information;
The communication module is controlled to send compressed and sub-packaged first voice information to the mobile terminal, the communication module is controlled to receive the second voice information sent by the mobile terminal, the compressed and sub-packaged first voice information is preprocessed by the mobile terminal to obtain preprocessed first voice information, the preprocessed first voice information is sent to the server through the mobile terminal, so that the server translates the preprocessed first voice information to obtain second voice information according to translation setting of the mobile terminal, and the second voice information is sent to the mobile terminal through the server;
The second voice information is subjected to grouping and decompressing processing;
And controlling the loudspeaker to play the second voice information after the package and decompression processing.
As a fourth possible implementation manner of the embodiment of the first aspect of the present application, the communication component is a bluetooth component.
The voice translation terminal of the embodiment of the application collects the first voice information through the microphone set, then sends the first voice information to the mobile terminal through the communication component, and receives the second voice information sent by the mobile terminal, wherein the first voice information is transmitted to the server through the mobile terminal, the second voice information is obtained through translation of the server, the server sends the second voice information to the voice translation terminal through the mobile terminal, and finally the voice translation terminal plays the second voice information through the loudspeaker. According to the application, the user does not need to set the corresponding keys on the voice translation terminal to determine the languages of the first voice information, and can automatically translate the first voice information among different languages by triggering the unique trigger keys to acquire the first voice information, so that the operation steps of the user can be simplified, and the use experience of the user is improved.
An embodiment of a second aspect of the present application provides a mobile terminal, including:
the mobile communication device comprises a shell, a mobile communication assembly and a mobile controller, wherein the mobile communication assembly is electrically connected with the mobile controller, and is communicated with a voice translation terminal and a server;
The mobile communication component is used for receiving the first voice information sent by the voice translation terminal, sending the first voice information preprocessed by the mobile processor to the server, so that the server translates the preprocessed first voice information to obtain second voice information according to the translation setting of the mobile processor;
the mobile communication component is further used for receiving the second voice information sent by the server and sending the second voice information to the voice translation terminal;
The mobile processor is used for preprocessing the first voice information;
The mobile processor is further configured to generate translation setting information, send the translation setting information to the server through the mobile communication component, so that the server determines a corresponding translation setting according to the translation setting information, and translate the preprocessed first voice information based on the translation setting to obtain second voice information.
As a first possible implementation of the embodiment of the second aspect of the present application,
The mobile processor pre-processes the first voice information, wherein the mobile processor is specifically configured to:
integrity checking is carried out on the first voice information;
If the first voice information passes the integrity check, further decoding the first voice information;
and if the first voice information does not pass the integrity check, feeding back to the voice translation terminal so that the voice translation terminal resends the first voice information.
As a second possible implementation manner of the embodiment of the second aspect of the present application, the translation setting information includes information of a setting of mutual translation between the first language and the second language or information of a setting of translation from the first language to the second language.
According to the mobile terminal provided by the embodiment of the application, the first voice information is preprocessed by receiving the first voice information sent by the voice translation terminal, and then the preprocessed first voice information is sent to the server, so that the server translates the preprocessed first voice information to obtain the second voice information according to the translation setting of the mobile terminal, and after translating the second voice information by the server, the first voice information can be sent to the voice translation terminal through the mobile terminal. According to the application, the user does not need to set the corresponding keys on the voice translation terminal to determine the languages of the first voice information, and can automatically translate the first voice information among different languages by triggering the unique trigger keys to acquire the first voice information, so that the operation steps of the user can be simplified, and the use experience of the user is improved.
An embodiment of a third aspect of the present application provides a translation system, including a speech translation terminal, a mobile terminal, and a server, where the speech translation terminal communicates with the mobile terminal, and the mobile terminal communicates with the server;
The speech translation terminal includes:
The mobile terminal comprises a shell, a controller, a trigger button, a microphone set, a loudspeaker and a communication assembly, wherein the shell forms a space for accommodating the controller, the microphone set, the loudspeaker and the communication assembly, the trigger button is arranged on the shell and is electrically connected with the controller, the microphone set is electrically connected with the controller and is used for collecting first voice information after the trigger button is triggered, the loudspeaker is electrically connected with the controller and is used for playing second voice information according to the control of the controller, the second voice information is voice information translated by the first voice information, the controller is electrically connected with the communication assembly and is used for controlling the communication assembly to send the first voice information to a mobile terminal and controlling the communication assembly to receive the second voice information sent by the mobile terminal;
the mobile terminal includes:
The mobile communication device comprises a shell, a mobile communication component, a mobile processor, the mobile communication component, the mobile processor and the server, wherein the mobile communication component is electrically connected with the mobile controller, the mobile communication component is communicated with a voice translation terminal and the server, the mobile communication component is used for receiving the first voice information sent by the voice translation terminal, sending the first voice information preprocessed by the mobile processor to the server, so that the server can obtain second voice information according to translation setting of the mobile terminal through the server, receiving the second voice information sent by the server, sending the second voice information to the voice translation terminal, the mobile processor is used for preprocessing the first voice information, generating translation setting information, sending the translation setting information to the server through the mobile communication component, and accordingly determining corresponding translation setting by the server according to the translation setting information, and translating the preprocessed first voice information to obtain the second voice information based on the translation setting.
The server includes:
The mobile terminal comprises a receiver for acquiring preprocessed first voice information from the mobile terminal, a processor for translating the preprocessed first voice information into second voice information according to translation setting of the mobile terminal, and a transmitter for transmitting the second voice information to the mobile terminal.
As a first possible implementation manner of the embodiment of the third aspect of the present application, the microphone set includes two microphones, one for collecting voice data input by a user, and the other for collecting noise data;
And the microphone set is used for subtracting and amplifying the voice data and the noise data to obtain the first voice information.
As a second possible implementation manner of the embodiment of the third aspect of the present application, the controller is configured to:
Receiving first voice information acquired by the microphone set, and compressing and subpackaging the first voice information;
The communication module is controlled to send compressed and sub-packaged first voice information to the mobile terminal, the communication module is controlled to receive the second voice information sent by the mobile terminal, the compressed and sub-packaged first voice information is preprocessed by the mobile terminal to obtain preprocessed first voice information, the preprocessed first voice information is sent to the server through the mobile terminal, so that the server translates the preprocessed first voice information to obtain second voice information according to translation setting of the mobile terminal, and the second voice information is sent to the mobile terminal through the server;
The second voice information is subjected to grouping and decompressing processing;
And controlling the loudspeaker to play the second voice information after the package and decompression processing.
As a third possible implementation manner of the embodiment of the third aspect of the present application, the communication component is a bluetooth component.
As a fourth possible implementation manner of the embodiment of the third aspect of the present application, the mobile processor performs preprocessing on the first voice information, and the mobile processor is specifically configured to:
integrity checking is carried out on the first voice information;
If the first voice information passes the integrity check, further decoding the first voice information;
and if the first voice information does not pass the integrity check, feeding back to the voice translation terminal so that the voice translation terminal resends the first voice information.
As a fifth possible implementation manner of the embodiment of the third aspect of the present application, the translation setting information includes information of a setting of mutual translation between the first language and the second language or information of a setting of translation from the first language to the second language.
According to the translation system provided by the embodiment of the application, the first voice information is acquired through the voice translation terminal when the trigger button is triggered, the acquired first voice information is sent to the mobile terminal, then the mobile terminal preprocesses the first voice information, the preprocessed first voice information is sent to the server, the server translates the preprocessed first voice information into the second voice information according to the translation setting of the mobile terminal, then the server can send the second voice information to the voice translation terminal through the mobile terminal, and correspondingly, the voice translation terminal can play the second voice information through the loudspeaker after receiving the second voice information. According to the application, the user does not need to set the corresponding keys on the voice translation terminal to determine the languages of the first voice information, and can automatically translate the first voice information among different languages by triggering the unique trigger keys to acquire the first voice information, so that the operation steps of the user can be simplified, and the use experience of the user is improved.
An embodiment of a fourth aspect of the present application provides a translation method applied to a speech translation terminal, where the speech translation terminal communicates with a mobile terminal, the method including:
Collecting first voice information;
The method comprises the steps of sending the first voice information to the mobile terminal, and receiving second voice information sent by the mobile terminal, wherein the second voice information is voice information translated by the first voice information;
and playing the second voice information.
As a first possible implementation manner of the fourth aspect of the present application, the second voice information is obtained by preprocessing the first voice information by the mobile terminal, sending the preprocessed first voice information to a server, and translating the preprocessed first voice information by the server according to a translation setting of the mobile terminal.
As a second possible implementation manner of the fourth aspect of the present application, the collecting the first voice information includes:
collecting voice data and noise data input by a user;
And subtracting and amplifying the voice data and the noise data to obtain the first voice information.
As a third possible implementation manner of the fourth aspect of the present application, after the collecting the first voice information, the method further includes:
compressing and packetizing the first voice information;
the sending the first voice information to the mobile terminal comprises the following steps:
And sending the compressed and sub-packaged first voice information to the mobile terminal so that the mobile terminal preprocesses the compressed and sub-packaged first voice information, sending the preprocessed first voice information to the server, and translating the preprocessed first voice information by the server according to the translation setting of the mobile terminal to obtain the second voice information.
As a fourth possible implementation manner of the fourth aspect of the present application, the playing the second voice information includes:
The second voice information is subjected to grouping and decompressing processing;
and playing the second voice information after the group package and decompression processing.
The translation method of the embodiment of the application comprises the steps of collecting first voice information through the voice translation terminal, then sending the first voice information to the mobile terminal, and receiving second voice information sent by the mobile terminal, wherein the first voice information is transmitted to the server through the mobile terminal, the second voice information is obtained through translation of the server, the server sends the second voice information to the voice translation terminal through the mobile terminal, and finally the voice translation terminal plays the second voice information. Therefore, the automatic translation of the first voice information among different languages can be realized.
An embodiment of the fifth aspect of the present application proposes another translation method applied to a mobile terminal, where the mobile terminal communicates with a speech translation terminal and a server, the method including:
Receiving first voice information sent by the voice translation equipment;
preprocessing the first voice information;
The preprocessed first voice information is sent to the server, so that the server translates the preprocessed first voice information to obtain second voice information according to translation setting of the mobile processor;
receiving second voice information sent by the server;
and sending the second voice information to the voice translation equipment.
As a first possible implementation of the fifth aspect of the embodiments of the present application,
The preprocessing the first voice information comprises the following steps:
integrity checking is carried out on the first voice information;
If the first voice information passes the integrity check, further decoding the first voice information;
and if the first voice information does not pass the integrity check, feeding back to the voice translation terminal so that the voice translation terminal resends the first voice information.
As a second possible implementation manner of the fifth aspect of the embodiments of the present application, the translation setting information includes information of a setting of mutual translation between the first language and the second language or information of a setting of translation from the first language to the second language.
According to the translation method, the mobile terminal receives the first voice information, performs preprocessing on the first voice information, then sends the preprocessed first voice information to the server, so that the server translates the first voice information into the second voice information according to translation setting of the mobile terminal, finally sends the second voice information to the mobile terminal, and correspondingly, after receiving the second voice information, the mobile terminal can send the second voice information to the voice translation device, so that the voice translation device plays the second voice information. Therefore, the automatic translation of the first voice information among different languages can be realized.
An embodiment of a sixth aspect of the present application provides a translation apparatus, including:
the acquisition module is used for acquiring the first voice information;
The sending module is used for sending the first voice information to the mobile terminal and receiving the second voice information sent by the mobile terminal, wherein the second voice information is the voice information translated by the first voice information;
And the playing module is used for playing the second voice information.
As a first possible implementation manner of the sixth aspect of the present application, the second voice information is obtained by preprocessing the first voice information by the mobile terminal, sending the preprocessed first voice information to a server, and translating the preprocessed first voice information by the server according to a translation setting of the mobile terminal.
As a second possible implementation manner of the sixth aspect of the embodiments of the present application, the collecting module is specifically configured to:
collecting voice data and noise data input by a user;
And subtracting and amplifying the voice data and the noise data to obtain the first voice information.
As a third possible implementation manner of the sixth aspect of the embodiments of the present application, the apparatus further includes:
the processing module is used for compressing and packetizing the first voice information after the first voice information is acquired;
the sending module is specifically configured to:
And sending the compressed and packetized first voice information to the mobile terminal so that the mobile terminal preprocesses the compressed and packetized first voice information, sending the preprocessed first voice information to a server, and translating the preprocessed first voice information by the server according to the translation setting of the mobile terminal to obtain the second voice information.
As a fourth possible implementation manner of the sixth aspect of the embodiments of the present application, the playing module is specifically configured to:
The second voice information is subjected to grouping and decompressing processing;
and playing the second voice information after the group package and decompression processing.
The translation device of the embodiment of the application collects the first voice information through the voice translation terminal, then sends the first voice information to the mobile terminal and receives the second voice information sent by the mobile terminal, wherein the first voice information is transmitted to the server through the mobile terminal, the second voice information is obtained through the translation of the server, the server sends the second voice information to the voice translation terminal through the mobile terminal, and finally the voice translation terminal plays the second voice information. Therefore, the automatic translation of the first voice information among different languages can be realized.
An embodiment of the seventh aspect of the present application proposes another translation apparatus, including:
the receiving module is used for receiving the first voice information sent by the voice translation equipment;
The preprocessing module is used for preprocessing the first voice information;
The sending module is used for sending the preprocessed first voice information to the server so that the server translates the preprocessed first voice information to obtain second voice information according to the translation setting of the mobile processor;
The receiving module is further used for receiving second voice information sent by the server;
the sending module is further configured to send the second voice information to the voice translation device.
As a first possible implementation of the seventh aspect of the embodiments of the present application,
The preprocessing module is specifically configured to:
integrity checking is carried out on the first voice information;
If the first voice information passes the integrity check, further decoding the first voice information;
and if the first voice information does not pass the integrity check, feeding back to the voice translation terminal so that the voice translation terminal resends the first voice information.
As a second possible implementation manner of the seventh aspect of the embodiments of the present application, the translation setting information includes information of a setting of mutual translation between the first language and the second language or information of a setting of translation from the first language to the second language.
According to the translation device, the mobile terminal receives the first voice information, performs preprocessing on the first voice information, then sends the preprocessed first voice information to the server, so that the server translates the first voice information into the second voice information according to translation setting of the mobile terminal, finally sends the second voice information to the mobile terminal, and correspondingly, after receiving the second voice information, the mobile terminal sends the second voice information to the voice translation device, so that the voice translation device plays the second voice information. Therefore, the automatic translation of the first voice information among different languages can be realized.
An eighth aspect of the present application proposes another speech translation terminal, comprising a memory, a processor and a computer program stored on the memory and executable on the processor, said processor implementing the translation method according to the fourth aspect of the present application when executing said program.
An embodiment of the ninth aspect of the present application proposes another mobile terminal, which is characterized by comprising a memory, a processor and a computer program stored on the memory and executable on the processor, said processor implementing the translation method according to the embodiment of the fifth aspect of the present application when executing said program.
An embodiment of the tenth aspect of the present application proposes a computer readable storage medium, on which a computer program is stored, characterized in that the program, when being executed by a processor, implements a translation method as proposed by an embodiment of the fourth aspect of the present application, or implements a translation method as proposed by an embodiment of the fifth aspect of the present application.
Additional aspects and advantages of the application will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the application.
Drawings
The foregoing and/or additional aspects and advantages of the application will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings, in which:
Fig. 1 is a schematic structural diagram of a speech translation terminal according to an embodiment of the present application;
fig. 2 is a schematic product shape diagram of a speech translation terminal according to a second embodiment of the present application;
fig. 3 is a schematic structural diagram of a speech translation terminal according to a third embodiment of the present application;
FIG. 4 is a schematic diagram of a translation process according to a fourth embodiment of the present application;
Fig. 5 is a schematic structural diagram of a mobile terminal according to a fifth embodiment of the present application;
FIG. 6 is a schematic diagram of a translation system according to a sixth embodiment of the present application;
FIG. 7 is a flow chart of a translation method according to a seventh embodiment of the present application;
FIG. 8 is a flow chart of a translation method according to an embodiment of the present application;
fig. 9 is a schematic structural diagram of a translation device according to a ninth embodiment of the present application;
fig. 10 is a schematic structural diagram of a translation device according to a tenth embodiment of the present application;
Fig. 11 is a schematic structural diagram of a translation device according to an eleventh embodiment of the present application.
Detailed Description
Embodiments of the present application are described in detail below, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to like or similar elements or elements having like or similar functions throughout. The embodiments described below by referring to the drawings are illustrative and intended to explain the present application and should not be construed as limiting the application.
The following describes a speech translation terminal, a mobile terminal, a translation system, a translation method, and devices thereof according to embodiments of the present application with reference to the accompanying drawings.
Fig. 1 is a schematic structural diagram of a speech translation terminal according to an embodiment of the present application.
As shown in fig. 1, the speech translation terminal 100 may include a housing (not shown), a controller 110, a trigger button 120, a microphone array 130, a speaker 140, and a communication assembly 150.
Wherein the housing forms a space for accommodating the controller 110, the microphone set 130, the speaker 140 and the communication assembly 150.
A trigger button 120 is disposed on the housing and electrically connected to the controller 110.
In the embodiment of the present application, only one trigger button 120 is set on the speech translation terminal 100, and the user can input speech data by triggering the trigger button 120, and accordingly, when the speech translation terminal 100 detects that the user triggers the trigger button 120 in a listening manner, the first speech information can be collected through the speaker set 130.
As an example, referring to fig. 2, fig. 2 is a schematic product shape of a speech translation terminal according to a second embodiment of the present application. The front side of the speech translation terminal 100 may include a trigger button 120, and when the user triggers the trigger button 120, the controller may control the microphone set 130 to collect the first speech information.
It should be noted that, in fig. 2, the appearance of the speech translation terminal 100 is only shown as a square, and in practical application, in order to improve the aesthetic appearance of the speech translation terminal 100, a designer may set the appearance of the speech translation terminal 100 according to his own needs, for example, may be set to be circular, oval, or the like, which is not limited.
The microphone set 130 is electrically connected to the controller 110, and is configured to collect the first voice information after the trigger button 120 is triggered.
In the embodiment of the present application, the microphone set 130 is composed of at least two microphones, for example, when the microphone set 130 includes two microphones, one microphone may collect voice data input by a user, and the other microphone may collect noise data. For example, one microphone may be disposed on the front side of the speech translation terminal 100 and mainly used for collecting speech data input by the user, and it will be understood by those skilled in the art that the microphone may have a small portion of environmental noise in addition to collecting speech data of the user, and another microphone may be disposed on the back side of the speech translation terminal 100 and mainly used for collecting noise data, and those skilled in the art will understand that the noise data may also include a small portion of speech data input by the user. The microphone set 130 may subtract and amplify the voice data and the noise data to obtain the first voice information. Therefore, the acquired first voice information is the voice information obtained after noise reduction processing, and the signal quality of the first voice information can be improved, so that the accuracy of a translation result can be improved when the second voice information is obtained through subsequent translation.
The speaker 140 is electrically connected to the controller 110, and is configured to play second voice information according to control of the controller 110, where the second voice information is translated voice information of the first voice information.
The controller 110 is electrically connected to the communication assembly 150, and is configured to control the communication assembly 150 to send the first voice information to the mobile terminal, and control the communication assembly 150 to receive the second voice information sent by the mobile terminal.
In the embodiment of the application, the mobile terminal can be a mobile phone, a tablet personal computer, a personal digital assistant, a wearable device and other hardware devices with various operating systems, touch screens and/or display screens.
In the embodiment of the present application, the communication component 150 may be, for example, a bluetooth component, and the speech translation terminal 100 may communicate with the mobile terminal through the communication component 150.
Specifically, the speech translation terminal 100 may communicate with the mobile terminal through the communication component 150, for example, a bluetooth component, after the speech translation terminal 100 collects the first speech information, the speech translation terminal 100 may send the first speech information to the mobile terminal through a bluetooth low energy protocol (Bluetooth Low Energy, abbreviated as BLE), and then send the first speech information to the server through the mobile terminal. After the server receives the first voice information, the server may translate the first voice information to obtain second voice information, and then, the server may send the second voice information to the mobile terminal, and the mobile terminal may send the second voice information to the voice translation terminal 100 based on the BLE protocol, where correspondingly, after receiving the second voice information, the voice translation terminal 100 may play the second voice information through the speaker 140.
For example, the voice translation terminal 100 may be connected to the mobile terminal through bluetooth in advance, and start an Application (APP) in the mobile terminal, such as a leopard translation APP. After the voice translation terminal 100 collects the first voice information, the voice translation terminal 100 may send the first voice information to the mobile terminal through a BLE protocol, and the translation APP in the mobile terminal may send the first voice information to the server through the network. After the server receives the first voice information, the server can translate the first voice information to obtain second voice information, then the server can send the second voice information to the mobile terminal through the network, the translation APP in the mobile terminal sends the second voice information to the voice translation terminal 100 based on the BLE protocol, and accordingly, the voice translation terminal 100 can play the second voice information through the speaker 140 after receiving the second voice information.
In the embodiment of the application, the server can translate the first voice information into the second voice information according to the translation setting of the mobile terminal. Wherein the translation setting is determined based on the translation setting information input by the user.
Specifically, the user may set the translation setting information in the mobile terminal, for example, the user may set the translation setting information in the translation APP of the mobile terminal, and after the user sets the translation setting information, the mobile terminal may send the translation setting information to the server, so that the server may determine the corresponding translation setting according to the translation setting information set by the user.
As a possible implementation manner, the translation setting information includes information of a setting of mutual translation between the first language and the second language, at this time, the translation setting includes a setting of mutual translation between the first language and the second language, the server may identify the language of the first voice information, if the language of the first voice information is the first language, the server may translate the first voice information of the first language into the second voice information of the second language, and if the language of the first voice information is the second language, the server may translate the first voice information of the second language into the second voice information of the first language. Therefore, the server automatically recognizes the language spoken by the user, and automatically translates the language spoken by the user into the voice information corresponding to the other language, so that the corresponding key is not required to be arranged on the voice translation equipment 100 to determine the language spoken by the user, the operation of the user can be simplified, and the use experience of the user can be improved.
The translation setting may be Chinese-English inter-translation, chinese-law inter-translation, english-German inter-translation, etc., for example, the translation setting may be Chinese, the first language may be English, or the first language may be English, and the second language may be Chinese.
As an example, in the case of translation setting as a chinese-english inter-translation example, when the user is translating using the speech translation terminal 100, if the user presses the trigger 120 and speaks english, the speech translation terminal 100 plays chinese after lifting the trigger 120, and if the user presses the trigger 120 and speaks chinese, the speech translation terminal 100 plays english after lifting the trigger 110. That is, no matter what the user speaks in Chinese or English, the voice translation terminal translates into another corresponding voice, so that automatic translation between two languages can be realized, and the user can use the voice translation terminal conveniently.
As another possible implementation, the translation setting information includes information of a setting translated from the first language to the second language, and at this time, the translation setting includes a setting translated from the first language to the second language, and the server may translate the first voice information to the second voice information of the second language. The translation setting may be middle translation, english translation, middle translation, etc., for example, the translation setting may be middle translation, the first language may be chinese, and the second language may be english.
As an example, in the case where the translation is set as a middle translation example, when the user is translating using the speech translation terminal 100, if the user presses the trigger 120 and speaks english, the speech translation terminal 100 plays chinese after lifting the trigger 120. Therefore, the language can be automatically translated into another language according to the language spoken by the user, and the language translation device is convenient for the user to use.
The voice translation terminal 100 of the embodiment of the present application collects first voice information through the microphone set 130, then sends the first voice information to the mobile terminal through the communication component 150, and receives second voice information sent by the mobile terminal, where the first voice information is transmitted to the server through the mobile terminal, the second voice information is obtained through translation of the server, the server sends the second voice information to the voice translation terminal through the mobile terminal, and finally, the voice translation terminal 100 plays the second voice information through the speaker 140. According to the application, the user does not need to set the corresponding keys on the voice translation terminal to determine the languages of the first voice information, and can automatically translate the first voice information among different languages by triggering the unique trigger keys to acquire the first voice information, so that the operation steps of the user can be simplified, and the use experience of the user is improved.
When the server translates the first voice information into the second voice information according to the translation setting, the server recognizes the first voice information as text information, translates the first voice information into the corresponding second voice information according to the text information, and synthesizes the second voice information. Specifically, the server may identify the first voice information to obtain a language of the first voice information, convert the first voice information into a text file according to the language corresponding to the first voice information, then determine a target language according to the language corresponding to the first voice information and a translation mode, translate the text file into a translated file of the target language, and finally generate the translated file into second voice information, for example, convert a file in a text format into voice information in a voice format.
It will be appreciated that the phoneme arrangement is different between different languages, i.e. the phonons of the pronunciation, the strings of phonons, and the frequency and context of occurrence of the phonons are different, based on which the languages can be distinguished.
As a possible implementation manner, a trained language recognition model may be set in the server, and the language to which the first voice information belongs may be recognized through the trained language recognition model.
As another possible implementation manner, a plurality of voice samples may be obtained in advance, and languages corresponding to the voice samples may be labeled, after the server obtains the first voice information, the first voice information may be matched with the voice samples, the voice samples with a matching degree higher than a preset threshold value may be determined, and the determined languages of the voice samples may be used as the languages of the first voice information. The preset threshold is preset, for example, may be 95%.
In the embodiment of the application, after the server identifies the language of the first voice information, the first voice information can be converted into the text file based on the voice identification technology in the related technology, and then the target language is determined according to the language and the translation mode corresponding to the first voice information, and the text file is translated into the translation file of the target language. For example, when the language corresponding to the first voice information is the first language, the target language is the second language, and when the language corresponding to the first voice information is the second language, the target language is the first language, and the server may translate the text file into the translation file of the target language based on the translation rule between the language corresponding to the first voice information and the target language in the related art. The server may then synthesize the translation file into second speech information.
For example, when the target language is english, the server may convert the translated document from letters to phonemes (or syllables) and synthesize the translated document to obtain the second speech information. When the target language is Chinese, the server can convert the translation file from Chinese characters to phonemes (or pinyin) and synthesize the translation file to obtain second voice information.
In the embodiment of the application, the accuracy of the translation result can be improved by recognizing the first voice information as the text information and then translating and synthesizing the first voice information into the corresponding second voice information according to the text information.
As a possible implementation manner, after receiving the first voice information, the mobile terminal may perform preprocessing on the first voice information, then may send the preprocessed first voice information to the server, and correspondingly, when receiving the preprocessed first voice information, the server may translate the preprocessed first voice information according to a translation setting of the mobile terminal to obtain second voice information, then the server may send the second voice information to the mobile terminal, and correspondingly, the mobile terminal may send the second voice information to the voice translation device 100 after receiving the second voice information, and correspondingly, the voice translation terminal 100 may receive the second voice information sent by the mobile terminal through the communication component 150.
It can be understood that, in the process of collecting voice and generating a voice file, the voice translation terminal 100 may generate a header, a tail, and the like of the file, and the first voice information finally sent to the mobile terminal should be a complete voice file, so that after receiving the first voice information, the mobile terminal may perform integrity check on the file of the first voice information to determine whether the first voice information is a complete voice file, and if not, feedback the first voice information to the voice translation terminal 100 to enable the voice translation terminal 100 to resend the first voice information. If yes, further decoding the first voice information, and sending the decoded first voice information to the server.
Specifically, the mobile terminal may perform decoding processing on the first voice information according to the format of the voice file supported by the server. For example, for microsoft server, the supported voice file format is MP3 format, so that the first voice information can be decoded into a voice file in MP3 format, and then the decoded first voice information can be sent to microsoft server, so that the microsoft server can translate the first voice information according to the supported format to obtain the second voice information.
In the embodiment of the application, the first voice information is decoded according to the format of the voice file supported by the server, so that the server can be ensured to normally translate the first voice information.
As a possible implementation, on the basis of the embodiment shown in fig. 1, the housing may comprise a first housing and a second housing arranged to match each other, wherein,
The first housing is provided with a light emitting diode LED and a trigger key 120.
Wherein, the LED can be controlled to be turned on or off by the LED control module (or the controller 110) to prompt the user. For example, when the user presses the trigger button 120, the LED control module (or the controller 110) may control the LEDs to be turned on to prompt the user, and when the user does not press the trigger button 120, the LED control module (or the controller 110) may control the LEDs to be turned off.
The second housing is provided with a speaker 140, a charging slot, and a microphone set 130.
A communication assembly 150, a power module, and an access module are disposed between the first housing and the second housing.
In the embodiment of the present application, in order to improve accuracy of a subsequent translation result, the microphone set 130 may acquire the first voice information after the noise reduction processing.
The controller 110 is configured to receive the first voice information collected by the microphone set 130, and compress and packetize the first voice information.
In the embodiment of the present application, in order to ensure the signal quality of the first voice information and reduce the network resource occupation, thereby improving the translation efficiency, after the first voice information is collected, the controller 110 may further perform compression and packetization processing on the processed first voice information.
The communication component 150 is configured to send the compressed and packetized first voice information to the mobile terminal, and accordingly, when the mobile terminal receives the compressed and packetized first voice information, the mobile terminal may perform preprocessing on the compressed and packetized first voice information to obtain preprocessed first voice information, and then, the mobile terminal may send the compressed and packetized first voice information to the server through the network, so that the server identifies the preprocessed first voice information to obtain a language of the preprocessed first voice information, and translate the preprocessed first voice information according to a translation setting of the mobile terminal and the language of the preprocessed first voice information to obtain second voice information. After the server translates the second voice information, the second voice information may be sent to the mobile terminal through the network, and the mobile terminal may send the second voice information to the voice translation device 100, and accordingly, the voice translation device 100 may receive the second voice information sent by the mobile terminal through the communication component 150.
The controller 110 is further configured to obtain second voice information sent by the server through the mobile terminal, and perform a packet and decompression process on the second voice information.
And a speaker 140 for playing the second voice information after the package and decompression processing.
As an example, referring to fig. 3, when the user presses the trigger button, the LED control module may control the LED to be turned on, and at the same time, the first voice information after the denoising and enhancing process may be collected through the microphone set, and then the first voice information may be stored by the access module.
The controller can acquire the first voice information from the access module and compress and packetize the first voice information. And then, the communication component can be controlled to send the compressed and packetized first voice information to the mobile terminal, the mobile terminal preprocesses the compressed and packetized first voice information to obtain preprocessed first voice information, and then, the mobile terminal can send the preprocessed first voice information to the server.
After the server obtains the preprocessed first voice information, the server can translate the preprocessed first voice information according to the translation setting of the mobile terminal and the language of the preprocessed first voice information to obtain second voice information, and the second voice information is sent to the voice translation terminal through the mobile terminal. After the voice translation terminal receives the second voice information through the communication component, the controller can carry out the package and decompression processing on the second voice information, and the loudspeaker is controlled to play the second voice information after the package and decompression processing.
As an example, referring to fig. 4, fig. 4 is a schematic diagram of a translation process according to a fourth embodiment of the present application. The user can press a trigger button on the voice translation terminal to collect voice data input by the user. It should be noted that, the voice translation terminal has only one trigger button, no matter what language the user speaks (the language supported by the voice translation device only needs to be set in the translation APP in the mobile terminal, that is, the user can set the language spoken by the user and the finally translated language in the translation APP), the user can press the trigger button, and after the user finishes speaking, the user can release the trigger button.
After the voice data is collected, noise reduction and enhancement processing can be performed on the voice data to obtain first voice information in order to improve accuracy of subsequent translation results. In order to ensure the quality of the first voice information and reduce the occupation of network resources, thereby improving the translation efficiency, the voice translation terminal can compress and packetize the first voice information. Then, the voice translation terminal can send the compressed and packetized first voice information to the mobile terminal based on the customized BLE reliable transmission protocol, and the translation APP in the mobile terminal can perform front-end processing (preprocessing) on the compressed and packetized first voice information, and then send the preprocessed first voice information to the server through the network.
After receiving the first voice information, the server can identify the language of the preprocessed first voice information according to the translation setting of the mobile terminal, convert the preprocessed first voice information into a text file according to a voice identification technology, translate the text file into a translation file of a target language, and synthesize the translation file into second voice information. And then, the server can send second voice information to the mobile terminal through the network, and the translation APP in the mobile terminal sends the second voice information to the voice translation terminal based on the BLE protocol.
After receiving the second voice information, the voice translation terminal can package, decompress and distribute the second voice information, and then play the second voice information after package and decompression processing through a loudspeaker.
In order to achieve the above embodiment, the present application further provides a mobile terminal.
Fig. 5 is a schematic structural diagram of a mobile terminal according to a fifth embodiment of the present application.
As shown in fig. 5, the mobile terminal 300 includes a housing (not shown), a mobile communication component 310, and a mobile controller 320.
The mobile communication module 310 is electrically connected to the mobile controller 320, and the mobile communication module 310 communicates with the speech translation terminal 100 and the server.
The mobile communication component 310 is configured to receive the first voice information sent by the voice translation terminal 100, send the first voice information preprocessed by the mobile processor 320 to the server, so that the server translates the preprocessed first voice information to obtain the second voice information according to the translation setting of the mobile processor 320.
The mobile communication module 310 is further configured to receive the second voice information sent by the server, send the second voice information to the voice translation terminal 100, the mobile processor 320 is configured to pre-process the first voice information, and the mobile processor 320 is further configured to generate translation setting information, send the translation setting information to the server through the mobile communication module 310, so that the server determines a corresponding translation setting according to the translation setting information, and translate the pre-processed first voice information based on the translation setting to obtain the second voice information.
In the embodiment of the application, the translation setting is determined according to the translation setting information input by the user. Specifically, the user may set the translation setting information in the mobile terminal, for example, the user may set the translation setting information in the translation APP of the mobile terminal, and after the user sets the translation setting information, the mobile terminal 300 may send the translation setting information to the server, so that the server may determine the corresponding translation setting according to the translation setting information set by the user.
As one possible implementation, the translation setting information includes information of a setting of mutual translation between the first language and the second language, and at this time, the translation setting includes a setting of mutual translation between the first language and the second language. The translation setting may be Chinese-English inter-translation, chinese-law inter-translation, english-German inter-translation, etc., for example, the translation setting may be Chinese, the first language may be English, or the first language may be English, and the second language may be Chinese.
In the embodiment of the present application, after the voice translation terminal 100 collects the first voice information, the voice translation terminal may send the first voice information to the mobile terminal 300 through the communication component, and correspondingly, the mobile terminal 300 may receive the first voice information through the mobile communication component 310, after the mobile terminal 300 receives the first voice information, the mobile processor 320 may perform preprocessing on the first voice information, and send the preprocessed first voice information to the server through the mobile communication component 310, so that the server may identify the language of the preprocessed first voice information, and if the language of the preprocessed first voice information is the first language, the server may translate the preprocessed first voice information of the first language into the second voice information of the second language, and if the language of the preprocessed first voice information is the second language, the server may translate the preprocessed first voice information of the second language into the second voice information of the first language. Therefore, the server automatically recognizes the language spoken by the user, and automatically translates the language spoken by the user into the voice information corresponding to the other language, so that the corresponding key is not required to be arranged on the voice translation equipment 100 to determine the language spoken by the user, the operation of the user can be simplified, and the use experience of the user can be improved.
As another possible implementation, the translation setting information includes information of a setting translated from the first language to the second language, and at this time, the translation setting includes a setting translated from the first language to the second language. The translation setting may be middle translation, english translation, middle translation, etc., for example, the translation setting may be middle translation, the first language may be chinese, and the second language may be english.
In the embodiment of the present application, after receiving the first voice information, the mobile terminal 300 may perform preprocessing on the first voice information and send the preprocessed first voice information to the server, so that the server may translate the first voice information into the second voice information in the second language.
The mobile terminal 300 of the embodiment of the present application performs preprocessing on the first voice information by receiving the first voice information sent by the voice translation terminal, and then sends the preprocessed first voice information to the server, so that the server translates the preprocessed first voice information to obtain second voice information according to the translation setting of the mobile terminal, and after translating the second voice information by the server, the server can send the first voice information to the voice translation terminal. According to the application, the user does not need to set the corresponding keys on the voice translation terminal to determine the languages of the first voice information, and can automatically translate the first voice information among different languages by triggering the unique trigger keys to acquire the first voice information, so that the operation steps of the user can be simplified, and the use experience of the user is improved.
As a possible implementation manner, when the mobile processor 320 preprocesses the first voice information, the mobile processor performs integrity check on the first voice information, determines whether the first voice information passes the integrity check, if the first voice information does not pass the integrity check, feeds back to the voice translation terminal, so that the voice translation terminal retransmits the first voice information, and if the first voice information passes the integrity check, further decodes the first voice information.
It can be understood that, in the process of collecting voice and generating a voice file, the voice translation terminal 100 may generate a header, a tail, and the like of the file, and the first voice information finally sent to the mobile terminal 300 should be a complete voice file, so that after receiving the first voice information, the mobile terminal 300 may perform integrity check on the first voice information by the mobile processor 320 to determine whether the first voice information is a complete voice file, and if not, feedback the first voice information to the voice translation terminal 100 to enable the voice translation terminal 100 to resend the first voice information. If yes, further decoding the first voice information, and sending the decoded first voice information to the server.
Specifically, the mobile processor 320 may perform decoding processing on the first voice information according to the format of the voice file supported by the server. For example, for microsoft server, the supported voice file format is MP3 format, so that the first voice information can be decoded into a voice file in MP3 format, and then the decoded first voice information can be sent to microsoft server, so that the microsoft server can translate the first voice information according to the supported format to obtain the second voice information.
In the embodiment of the application, the first voice information is decoded according to the format of the voice file supported by the server, so that the server can be ensured to normally translate the first voice information.
In order to implement the above embodiment, the present application further proposes a translation system.
Fig. 6 is a schematic structural diagram of a translation system according to a sixth embodiment of the present application.
As shown in fig. 6, the translation system may include a speech translation terminal 100, a server 200, and a mobile terminal 300. Wherein,
The speech translation terminal 100 includes a housing (not shown), a controller 110, a trigger button 120, a microphone array 130, a speaker 140, and a communication assembly 150. Wherein,
The shell forms a space for accommodating the controller 110, the microphone set 130, the loudspeaker 140 and the communication assembly 150, the trigger key 120 is arranged on the shell and is electrically connected with the controller 110, the microphone set 130 is electrically connected with the controller 110 and is used for collecting first voice information after the trigger key 120 is triggered, the loudspeaker 140 is electrically connected with the controller 110 and is used for playing second voice information according to the control of the controller 110, wherein the second voice information is voice information after the first voice information is translated, the controller 110 is electrically connected with the communication assembly 150 and is used for controlling the communication assembly 150 to send the first voice information to the mobile terminal 300 and controlling the communication assembly 150 to receive the second voice information sent by the mobile terminal 300.
In the embodiment of the present application, only one trigger button 120 is set on the speech translation terminal 100, and the user can input speech data by triggering the trigger button 120. As an example, referring to fig. 2, the front side of the speech translation terminal 100 may include a trigger key 120, and a user may input speech data by triggering the trigger key 120. Accordingly, when the voice translation terminal 100 detects that the user triggers the trigger key 120 through the interception mode, the speaker set 130 may collect the first voice information.
In the embodiment of the present application, the microphone set 130 is composed of at least two microphones, for example, when the microphone set 130 includes two microphones, one microphone may collect voice data input by a user, and the other microphone may collect noise data. For example, one microphone may be disposed on the front side of the speech translation terminal 100 and mainly used for collecting speech data input by the user, and it will be understood by those skilled in the art that the microphone may have a small portion of environmental noise in addition to collecting speech data of the user, and another microphone may be disposed on the back side of the speech translation terminal 100 and mainly used for collecting noise data, and those skilled in the art will understand that the noise data may also include a small portion of speech data input by the user. The microphone set 130 may subtract and amplify the voice data and the noise data to obtain the first voice information. Therefore, the acquired first voice information is the voice information obtained after noise reduction processing, and the signal quality of the first voice information can be improved, so that the accuracy of a translation result can be improved when the second voice information is obtained through subsequent translation. In the embodiment of the present application, the communication component 150 may be, for example, a bluetooth component, and the speech translation terminal 100 may communicate with the mobile terminal through the communication component 150.
The mobile terminal 300 includes a housing (not shown), a mobile communication component 310, and a mobile processor 320. Wherein,
The mobile communication module 310 is electrically connected to the mobile controller 320, the mobile communication module 310 is in communication with the voice translation terminal 100 and the server 200, the mobile communication module 310 is configured to receive the first voice information sent by the voice translation terminal 100, send the first voice information preprocessed by the mobile processor 320 to the server 200, so that the server 200 can translate the preprocessed first voice information according to the translation setting of the mobile terminal 300 to obtain the second voice information, the mobile communication module 310 is further configured to receive the second voice information sent by the server 200, send the second voice information to the voice translation terminal 100, the mobile processor 320 is configured to preprocess the first voice information, the mobile processor 320 is further configured to generate translation setting information, and send the translation setting information to the server 200 through the mobile communication module 310, so that the server 200 can determine a corresponding translation setting according to the translation setting information, and translate the preprocessed first voice information based on the translation setting to obtain the second voice information.
The server 200 includes a receiver 210, a processor 220, and a transmitter 230. Wherein,
The mobile terminal 300 includes a receiver 210 for acquiring the preprocessed first voice information from the mobile terminal 300, a processor 220 for translating the preprocessed first voice information into second voice information according to a translation setting of the mobile terminal 300, and a transmitter 230 for transmitting the second voice information to the mobile terminal 300.
In the embodiment of the present application, the speech translation terminal 100 may communicate with the mobile terminal 300 through the communication component 150, for example, a bluetooth component, after the speech translation terminal 100 collects the first speech information, the speech translation terminal 100 may send the first speech information to the mobile communication component 310 of the mobile terminal 300 through the BLE protocol, and then, the mobile processor 320 in the mobile terminal 300 performs preprocessing on the first speech information. After the mobile processor 320 pre-processes the first voice information, the pre-processed first voice information may be transmitted to the server 200 through the mobile communication component 310. When the server 200 receives the preprocessed first voice information, the preprocessed first voice information may be translated by the processor 220 to obtain second voice information, and then the server 200 may send the second voice information to the mobile terminal 300 through the transmitter 230, and the mobile communication component 310 in the mobile terminal 300 may send the second voice information to the voice translation terminal 100 based on the BLE protocol, and accordingly, after receiving the second voice information, the voice translation terminal 100 may play the first voice information through the speaker 140.
Specifically, the processor 220 in the server 200 may translate the preprocessed first voice information into the second voice information according to the translation setting of the mobile terminal 300. Wherein the translation setting is determined based on the translation setting information input by the user.
In the embodiment of the application, the user can set the translation setting information in the mobile terminal, for example, the user can set the translation setting information in the translation APP of the mobile terminal, and after the user sets the translation setting information, the mobile terminal can send the translation setting information to the server, so that the server can determine the corresponding translation setting according to the translation setting information set by the user.
As a possible implementation manner, the translation setting information includes information of a setting of mutual translation between the first language and the second language, where the translation setting includes a setting of mutual translation between the first language and the second language, the processor 220 may identify the language of the preprocessed first voice information, if the language of the preprocessed first voice information is the first language, the processor 220 may translate the preprocessed first voice information of the first language into the second voice information of the second language, and if the language of the preprocessed first voice information is the second language, the processor 220 may translate the preprocessed first voice information of the second language into the second voice information of the first language. Therefore, the server automatically recognizes the language spoken by the user, and automatically translates the language spoken by the user into the voice information corresponding to the other language, so that the corresponding key is not required to be arranged on the voice translation equipment 100 to determine the language spoken by the user, the operation of the user can be simplified, and the use experience of the user can be improved.
The translation setting may be Chinese-English inter-translation, chinese-law inter-translation, english-German inter-translation, etc., for example, the translation setting may be Chinese, the first language may be English, or the first language may be English, and the second language may be Chinese.
As another possible implementation, the translation setting information includes information of a setting translated from the first language to the second language, and at this time, the translation setting includes a setting translated from the first language to the second language, and the processor 220 may translate the preprocessed first voice information to the second voice information of the second language.
The translation setting may be middle translation, english translation, middle translation, etc., for example, the translation setting may be middle translation, the first language may be chinese, and the second language may be english.
According to the translation system provided by the embodiment of the application, the first voice information is acquired through the voice translation terminal when the trigger button is triggered, the acquired first voice information is sent to the mobile terminal, then the mobile terminal preprocesses the first voice information, the preprocessed first voice information is sent to the server, the server translates the preprocessed first voice information into the second voice information according to the translation setting of the mobile terminal, then the server sends the second voice information to the voice translation terminal through the mobile terminal, and correspondingly, the voice translation terminal plays the second voice information after receiving the second voice information. According to the application, the user does not need to set the corresponding keys on the voice translation terminal to determine the languages of the first voice information, and can automatically translate the first voice information among different languages by triggering the unique trigger keys to acquire the first voice information, so that the operation steps of the user can be simplified, and the use experience of the user is improved.
As a possible implementation, the mobile processor 320 pre-processes the first voice information, specifically by performing an integrity check on the first voice information, further decodes the first voice information if the first voice information passes the integrity check, and feeds back to the voice translation terminal 100 if the first voice information does not pass the integrity check, so that the voice translation terminal 100 resends the first voice information.
It can be understood that, in the process of collecting voice and generating a voice file, the voice translation terminal 100 may generate a header, a tail, and the like of the file, and finally, the first voice information sent to the mobile terminal 300 is a complete voice file, so that after receiving the first voice information, the mobile terminal 300 may perform integrity check on the file of the first voice information to determine whether the first voice information is a complete voice file, and if not, feedback the first voice information to the voice translation terminal 100 to enable the voice translation terminal 100 to resend the first voice information. If so, further decoding the first voice information, and transmitting the decoded first voice information to the server 200.
Specifically, the mobile terminal 300 may perform decoding processing on the first voice information according to the format of the voice file supported by the server 200. For example, for microsoft server, the supported voice file format is MP3 format, so that the first voice information can be decoded into a voice file in MP3 format, and then the decoded first voice information can be sent to microsoft server, so that the microsoft server can translate the first voice information according to the supported format to obtain the second voice information.
In the embodiment of the application, the first voice information is decoded according to the format of the voice file supported by the server, so that the server can be ensured to normally translate the first voice information.
As a possible implementation manner, in order to ensure the signal quality of the first voice information and reduce the network resource occupation, so as to improve the translation efficiency, after the first voice information is collected, the controller 110 may further perform compression and packetization processing on the first voice information.
The communication component 150 is configured to send the compressed and packetized first voice information to the mobile terminal 300, where the mobile terminal 300 may pre-process the compressed and packetized first voice information, and then send the pre-processed first voice information to the server 200, so that the server 200 identifies the pre-processed first voice information to obtain a language of the pre-processed first voice information, and translate the pre-processed first voice information according to a translation setting of the mobile terminal 300 and the language of the pre-processed first voice information to generate the second voice information.
The controller 110 is further configured to obtain the second voice information sent by the server 200, and perform a packet and decompression process on the second voice information.
And a speaker 140 for playing the second voice information after the package and decompression processing.
In order to implement the above embodiment, the present application further provides a translation method.
Fig. 7 is a flow chart of a translation method according to a seventh embodiment of the present application.
As shown in fig. 7, the translation method is applied to the voice translation terminal provided in the above embodiment, where the voice translation terminal communicates with the mobile terminal, and the translation method may include the following steps:
Step 101, first voice information is collected.
In the embodiment of the application, the voice translation terminal can be provided with the trigger button, the user can input voice data by triggering the trigger button, and correspondingly, when the voice translation terminal detects that the user triggers the trigger button in a interception mode, the first voice information can be acquired through the loudspeaker group.
Wherein the microphone set is composed of at least two microphones, for example, when the microphone set includes two microphones, one microphone may collect voice data input by a user and the other microphone may collect noise data. For example, one microphone may be disposed on the front side of the speech translation terminal and used for collecting speech data inputted by the user, and it will be understood by those skilled in the art that the microphone may have a small portion of environmental noise in addition to collecting speech data of the user, and another microphone may be disposed on the back side of the speech translation terminal and used for collecting noise data, and those skilled in the art will understand that the noise data may also include a small portion of speech data inputted by the user. The microphone set may subtract and amplify the voice data and the noise data to obtain the first voice information. Therefore, the acquired first voice information is the voice information obtained after noise reduction processing, and the signal quality of the first voice information can be improved, so that the accuracy of a translation result can be improved when the second voice information is obtained through subsequent translation.
Step 102, sending first voice information to the mobile terminal, and receiving second voice information sent by the mobile terminal, wherein the second voice information is the voice information after the translation of the first voice information.
In the embodiment of the application, the voice translation terminal can communicate with the mobile terminal through the communication component, such as the Bluetooth component, after the voice translation terminal collects the first voice information, the voice translation terminal can send the first voice information to the mobile terminal through the BLE protocol, and then the mobile terminal can send the first voice information to the server. After the server receives the first voice information, the second voice information can be obtained by translating the first voice information. After the server translates the second voice information, the second voice information can be sent to the voice translation terminal through the mobile terminal.
Specifically, the server may translate the first voice information into the second voice information according to the translation setting of the mobile terminal. Wherein the translation setting is determined based on the translation setting information input by the user.
In the embodiment of the application, the user can set the translation setting information in the mobile terminal, for example, the user can set the translation setting information in the translation APP of the mobile terminal, and after the user sets the translation setting information, the mobile terminal can send the translation setting information to the server, so that the server can determine the corresponding translation setting according to the translation setting information set by the user.
As a possible implementation manner, the translation setting information includes information of a setting of mutual translation between the first language and the second language, at this time, the translation setting includes a setting of mutual translation between the first language and the second language, the server may identify the language of the first voice information, if the language of the first voice information is the first language, the server may translate the first voice information of the first language into the second voice information of the second language, and if the language of the first voice information is the second language, the server may translate the first voice information of the second language into the second voice information of the first language. Therefore, the server automatically recognizes the language spoken by the user, and automatically translates the language spoken by the user into the voice information corresponding to the other language, so that the corresponding keys are not required to be arranged on the voice translation equipment to determine the language spoken by the user, the operation of the user can be simplified, and the use experience of the user is improved.
The translation setting may be Chinese-English inter-translation, chinese-law inter-translation, english-German inter-translation, etc., for example, the translation setting may be Chinese, the first language may be English, or the first language may be English, and the second language may be Chinese.
As another possible implementation, the translation setting information includes information of a setting translated from the first language to the second language, and at this time, the translation setting includes a setting translated from the first language to the second language, and the server may translate the first voice information to the second voice information of the second language.
The translation setting may be middle translation, english translation, middle translation, etc., for example, the translation setting may be middle translation, the first language may be chinese, and the second language may be english.
Step 103, playing the second voice information.
In the embodiment of the application, after the voice translation terminal receives the second voice information, the second voice information can be played through the loudspeaker.
The translation method of the embodiment of the application comprises the steps of collecting first voice information through the voice translation terminal, then sending the first voice information to the mobile terminal, and receiving second voice information sent by the mobile terminal, wherein the first voice information is transmitted to the server through the mobile terminal, the second voice information is obtained through the translation of the server, the server sends the second voice information to the voice translation terminal through the mobile terminal, and finally the voice translation terminal plays the second voice information. Therefore, the automatic translation of the first voice information among different languages can be realized.
As a possible implementation manner, the first voice information received by the server is preprocessed by the mobile terminal, that is, the second voice information is obtained by preprocessing the first voice information by the mobile terminal, sending the preprocessed first voice information to the server, and translating the preprocessed first voice information by the server according to a translation setting of the mobile terminal.
It can be understood that in the process of collecting voice and generating a voice file, the voice translation terminal can generate a header, a tail and the like of the file, and finally, the first voice information sent to the mobile terminal is a complete voice file, so that after receiving the first voice information, the mobile terminal can perform file integrity check on the first voice information to determine whether the first voice information is the complete voice file, and if not, the first voice information is fed back to the voice translation terminal so as to enable the voice translation terminal to resend the first voice information. If yes, further decoding the first voice information, and sending the decoded first voice information to the server.
Specifically, the mobile terminal may perform decoding processing on the first voice information according to the format of the voice file supported by the server. For example, for microsoft server, the supported voice file format is MP3 format, so that the first voice information can be decoded into a voice file in MP3 format, and then the decoded first voice information can be sent to microsoft server, so that the microsoft server can translate the first voice information according to the supported format to obtain the second voice information.
In the embodiment of the application, the first voice information is decoded according to the format of the voice file supported by the server, so that the server can be ensured to normally translate the first voice information.
As a possible implementation manner, in order to ensure the quality of the first voice information and reduce the network resource occupation, so as to improve the translation efficiency, after the microphone group collects the first voice information, the first voice information may be compressed and packetized.
Accordingly, step 102 may include transmitting the compressed and packetized first voice information to the mobile terminal. After receiving the compressed and sub-packaged first voice information, the mobile terminal can preprocess the compressed and sub-packaged first voice information and send the preprocessed first voice information to the server, so that the server can identify the preprocessed first voice information to acquire the language of the preprocessed first voice information, and translate the preprocessed first voice information according to the translation setting of the mobile terminal and the language of the preprocessed first voice information to generate second voice information.
After the server translates to obtain the second voice information, the second voice information can be sent to the voice translation terminal through the mobile terminal, and correspondingly, the voice translation terminal can carry out package and decompression processing on the second voice information after receiving the second voice information, and then can play the second voice information after the package and decompression processing.
In order to implement the above embodiment, the present application further provides a translation method.
Fig. 8 is a flow chart of a translation method according to an embodiment of the present application.
As shown in fig. 8, the translation method is applied to the mobile terminal provided in the above embodiment, and the mobile terminal communicates with the speech translation terminal and the server, and the translation method may include the following steps:
step 201, receiving first voice information sent by a voice translation device.
In the embodiment of the application, after the voice translation terminal acquires the first voice information, the first voice information can be sent to the mobile terminal through the communication component, and correspondingly, the mobile terminal can receive the first voice information sent by the voice translation device.
Step 202, preprocessing the first voice information.
In the embodiment of the application, after receiving the first voice information, the mobile terminal can preprocess the first voice information to obtain the preprocessed first voice information.
It can be understood that, in the process of collecting voice and generating a voice file, the voice translation terminal can generate a header, a tail, and the like of the file, and finally, the first voice information sent to the mobile terminal is a complete voice file, so as to be a possible implementation manner of the embodiment of the application, after receiving the first voice information, the mobile terminal can perform integrity check on the first voice information to determine whether the first voice information is a complete voice file, and if not, the first voice information is fed back to the voice translation terminal, so that the voice translation terminal resends the first voice information. If yes, further decoding the first voice information, and sending the decoded first voice information to the server.
Specifically, the mobile terminal may perform decoding processing on the first voice information according to the format of the voice file supported by the server. For example, for microsoft server, the supported voice file format is MP3 format, so that the first voice information can be decoded into a voice file in MP3 format, and then the decoded first voice information can be sent to microsoft server, so that the microsoft server can translate the first voice information according to the supported format to obtain the second voice information.
In the embodiment of the application, the first voice information is decoded according to the format of the voice file supported by the server, so that the server can be ensured to normally translate the first voice information.
Step 203, the preprocessed first voice information is sent to the server, so that the server translates the preprocessed first voice information to obtain second voice information according to the translation setting of the mobile processor.
In the embodiment of the application, the translation setting is determined according to the translation setting information input by the user. Specifically, the user may set the translation setting information in the mobile terminal, for example, the user may set the translation setting information in the translation APP of the mobile terminal, and after the user sets the translation setting information, the mobile terminal may send the translation setting information to the server, so that the server may determine the corresponding translation setting according to the translation setting information set by the user.
As one possible implementation, the translation setting information includes information of a setting of mutual translation between the first language and the second language, and at this time, the translation setting includes a setting of mutual translation between the first language and the second language. The translation setting may be Chinese-English inter-translation, chinese-law inter-translation, english-German inter-translation, etc., for example, the translation setting may be Chinese, the first language may be English, or the first language may be English, and the second language may be Chinese.
In the embodiment of the application, after the mobile terminal preprocesses the first voice information, the mobile terminal can send the preprocessed first voice information to the server, so that the server can recognize the language of the preprocessed first voice information, if the language of the preprocessed first voice information is the first language, the server can translate the preprocessed first voice information of the first language into the second voice information of the second language, and if the language of the preprocessed first voice information is the second language, the server can translate the preprocessed first voice information of the second language into the second voice information of the first language. Therefore, the server automatically recognizes the language spoken by the user, and automatically translates the language spoken by the user into the voice information corresponding to the other language, so that the corresponding keys are not required to be arranged on the voice translation equipment to determine the language spoken by the user, the operation of the user can be simplified, and the use experience of the user is improved.
As another possible implementation, the translation setting information includes information of a setting translated from the first language to the second language, and at this time, the translation setting includes a setting translated from the first language to the second language. The translation setting may be middle translation, english translation, middle translation, etc., for example, the translation setting may be middle translation, the first language may be chinese, and the second language may be english.
In the embodiment of the application, after the mobile terminal preprocesses the first voice information, the preprocessed first voice information can be sent to the server, so that the server can translate the preprocessed first voice information into the second voice information of the second language.
Step 204, receiving the second voice information sent by the server, and sending the second voice information to the voice translation device.
In the embodiment of the application, after the server translates to obtain the second voice information, the second voice information can be sent to the mobile terminal, and then the mobile terminal sends the second voice information to the voice translation terminal.
According to the translation method, the mobile terminal receives the first voice information, performs preprocessing on the first voice information, then sends the preprocessed first voice information to the server, so that the server translates the first voice information into the second voice information according to translation setting of the mobile terminal, finally sends the second voice information to the mobile terminal, and correspondingly, after receiving the second voice information, the mobile terminal can send the second voice information to the voice translation device, so that the voice translation device plays the second voice information. Therefore, the automatic translation of the first voice information among different languages can be realized.
In order to realize the embodiment, the application also provides a translation device.
Fig. 9 is a schematic structural diagram of a translation device according to a ninth embodiment of the present application.
As shown in fig. 9, the translating apparatus includes an acquisition module 410, a sending module 420, and a playing module 430. Wherein,
The acquisition module 410 is configured to acquire the first voice information.
As a possible implementation manner, the collection module 410 is specifically configured to collect voice data and noise data input by a user, and subtract and amplify the voice data and the noise data to obtain first voice information.
The sending module 420 is configured to send the first voice information to the mobile terminal and receive the second voice information sent by the mobile terminal, where the second voice information is the voice information translated by the first voice information.
As a possible implementation manner, the second voice information is obtained by preprocessing the first voice information by the mobile terminal, sending the preprocessed first voice information to the server, and translating the preprocessed first voice information by the server according to a translation setting of the mobile terminal.
And a playing module 430, configured to play the second voice information.
Further, as a possible implementation manner of the embodiment of the present application, referring to fig. 10, the translation device may further include a processing module 440 based on the embodiment shown in fig. 9.
The processing module 440 is configured to compress and packetize the first voice information after the first voice information is collected.
The sending module 420 is specifically configured to send the compressed and packetized first voice information to the mobile terminal, so that the mobile terminal pre-processes the compressed and packetized first voice information, and send the pre-processed first voice information to the server, where the server translates the pre-processed first voice information according to the translation setting of the mobile terminal to obtain the second voice information.
The playing module 430 is specifically configured to perform a process of packaging and decompressing the second voice information and play the packaged and decompressed second voice information.
It should be noted that the explanation of the embodiment of the translation method in fig. 7 is also applicable to the translation device of this embodiment, and will not be repeated here.
The translation device of the embodiment of the application collects the first voice information through the voice translation terminal, then sends the first voice information to the mobile terminal and receives the second voice information sent by the mobile terminal, wherein the first voice information is transmitted to the server through the mobile terminal, the second voice information is obtained through the translation of the server, the server sends the second voice information to the voice translation terminal through the mobile terminal, and finally the voice translation terminal plays the second voice information. Therefore, the automatic translation of the first voice information among different languages can be realized.
In order to realize the embodiment, the application also provides a translation device.
Fig. 11 is a schematic structural diagram of a translation device according to an eleventh embodiment of the present application.
As shown in fig. 11, the translation apparatus includes a receiving module 510, a preprocessing module 520, and a transmitting module 530.
The receiving module 510 is configured to receive first voice information sent by the voice translating apparatus.
The preprocessing module 520 is configured to preprocess the first voice information.
As a possible implementation manner, the preprocessing module 520 is specifically configured to perform integrity check on the first voice information, further decode the first voice information if the first voice information passes the integrity check, and feed back to the voice translation terminal if the first voice information does not pass the integrity check, so that the voice translation terminal resends the first voice information.
And the sending module 530 is configured to send the preprocessed first voice information to the server, so that the server translates the preprocessed first voice information to obtain the second voice information according to the translation setting of the mobile processor.
As one possible implementation, the translation setting information includes information of a setting of mutual translation between the first language and the second language or information of a setting of translation from the first language to the second language.
The receiving module 510 is further configured to receive the second voice information sent by the server.
The sending module 530 is further configured to send the second speech information to the speech translation device.
It should be noted that the explanation of the embodiment of the translation method in fig. 8 is also applicable to the translation device in this embodiment, and will not be repeated here.
According to the translation device, the mobile terminal receives the first voice information, performs preprocessing on the first voice information, then sends the preprocessed first voice information to the server, so that the server translates the first voice information into the second voice information according to the translation setting of the mobile terminal, and finally sends the second voice information to the mobile terminal. Therefore, the automatic translation of the first voice information among different languages can be realized.
In order to implement the above embodiment, the present application further provides a speech translation terminal, which includes a memory, a processor, and a computer program stored in the memory and capable of running on the processor, where the processor implements the translation method according to the embodiment of fig. 7 of the present application when executing the program.
In order to implement the above embodiment, the present application further provides a server, which is characterized by comprising a memory, a processor, and a computer program stored in the memory and capable of running on the processor, where the processor implements the translation method according to the embodiment of fig. 8 of the present application when executing the program.
In order to implement the above embodiment, the present application also proposes a computer readable storage medium, on which a computer program is stored, characterized in that the program, when executed by a processor, implements a translation method as proposed in the foregoing embodiment of fig. 7 of the present application, or implements a translation method as proposed in the foregoing embodiment of fig. 8 of the present application.
In the description of the present specification, a description referring to terms "one embodiment," "some embodiments," "examples," "specific examples," or "some examples," etc., means that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present application. In this specification, schematic representations of the above terms are not necessarily directed to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Furthermore, the different embodiments or examples described in this specification and the features of the different embodiments or examples may be combined and combined by those skilled in the art without contradiction.
Furthermore, the terms "first," "second," and the like, are used for descriptive purposes only and are not to be construed as indicating or implying a relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defining "a first" or "a second" may explicitly or implicitly include at least one such feature. In the description of the present application, the meaning of "plurality" means at least two, for example, two, three, etc., unless specifically defined otherwise.
Any process or method descriptions in flow charts or otherwise described herein may be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps of the process, and additional implementations are included within the scope of the preferred embodiment of the present application in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order from that shown or discussed, depending on the functionality involved, as would be understood by those reasonably skilled in the art of the embodiments of the present application.
Logic and/or steps represented in the flowcharts or otherwise described herein, e.g., a ordered listing of executable instructions for implementing logical functions, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. For the purposes of this description, a "computer-readable medium" can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. More specific examples (a non-exhaustive list) of the computer-readable medium include an electrical connection (an electronic device) having one or more wires, a portable computer diskette (a magnetic device), a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber device, and a portable compact disc read-only memory. Additionally, the computer-readable medium may even be paper or other suitable medium upon which the program is printed, as the program may be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
Portions of the present application may be implemented in hardware, software, firmware, or a combination thereof. In the above-described embodiments, the various steps or methods may be implemented in software or firmware stored in a memory and executed by a suitable instruction execution system. If implemented in hardware as in another embodiment, it may be implemented in any one or combination of techniques known in the art, discrete logic circuits with logic gates for performing logic functions on data signals, application specific integrated circuits with appropriate combinational logic gates, programmable gate arrays, field programmable gate arrays, and the like.
Those of ordinary skill in the art will appreciate that all or part of the steps carried out in the method of the above-described embodiments may be implemented by a program to instruct related hardware, and the program may be stored in a computer readable storage medium, where the program when executed includes one or a combination of the steps of the method embodiments.
In addition, each functional unit in the embodiments of the present application may be integrated in one processing module, or each unit may exist alone physically, or two or more units may be integrated in one module. The integrated modules may be implemented in hardware or in software functional modules. The integrated modules may also be stored in a computer readable storage medium if implemented as software functional modules and sold or used as a stand-alone product.
The above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, or the like. While embodiments of the present application have been shown and described above, it will be understood that the above embodiments are illustrative and not to be construed as limiting the application, and that variations, modifications, alternatives and variations may be made to the above embodiments by one of ordinary skill in the art within the scope of the application.

Claims (25)

The mobile terminal comprises a shell, a controller, a trigger button, a microphone set, a loudspeaker and a communication component, wherein the shell forms a space for accommodating the controller, the microphone set, the loudspeaker and the communication component, the trigger button is arranged on the shell and is electrically connected with the controller, the microphone set is electrically connected with the controller and is used for collecting first voice information after the trigger button is triggered, the loudspeaker is electrically connected with the controller and is used for playing second voice information according to the control of the controller, the second voice information is voice information translated by the first voice information, the controller is electrically connected with the communication component and is used for controlling the communication component to send the first voice information to a mobile terminal and controlling the communication component to receive the second voice information sent by the mobile terminal, one of the microphone set comprises two microphones and is used for collecting voice data input by a user, the other microphone set is used for collecting noise data, and the communication component is a Bluetooth component;
The mobile communication device comprises a shell, a mobile communication component, a mobile processor and a mobile processor, wherein the mobile communication component is electrically connected with the mobile processor, the mobile communication component is communicated with a voice translation terminal and a server, the mobile communication component is used for receiving first voice information sent by the voice translation terminal, sending the first voice information preprocessed by the mobile processor to the server so that the server can obtain second voice information according to translation setting of the mobile terminal through translating the preprocessed first voice information, the mobile communication component is also used for receiving the second voice information sent by the server and sending the second voice information to the voice translation terminal, the mobile processor is used for preprocessing the first voice information, the mobile processor is also used for generating translation setting information, sending the translation setting information to the server through the mobile communication component so that the server can determine corresponding translation setting according to the translation setting information and translate the preprocessed first voice information based on the translation setting to obtain second voice information.
CN201910173337.5A2019-03-072019-03-07 Voice translation terminal, mobile terminal, translation system, translation method and device thereofActiveCN109754781B (en)

Priority Applications (3)

Application NumberPriority DateFiling DateTitle
CN201910173337.5ACN109754781B (en)2019-03-072019-03-07 Voice translation terminal, mobile terminal, translation system, translation method and device thereof
JP2020000107UJP3225682U (en)2019-03-072020-01-15 Voice translation terminal, mobile terminal and translation system
US16/801,342US20200285815A1 (en)2019-03-072020-02-26Speech translation terminal, mobile terminal, translation system, translation method, and translation device

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN201910173337.5ACN109754781B (en)2019-03-072019-03-07 Voice translation terminal, mobile terminal, translation system, translation method and device thereof

Publications (2)

Publication NumberPublication Date
CN109754781A CN109754781A (en)2019-05-14
CN109754781Btrue CN109754781B (en)2025-04-15

Family

ID=66406972

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN201910173337.5AActiveCN109754781B (en)2019-03-072019-03-07 Voice translation terminal, mobile terminal, translation system, translation method and device thereof

Country Status (1)

CountryLink
CN (1)CN109754781B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN110245360A (en)*2019-06-102019-09-17北京金山安全软件有限公司Handheld voice translator and use method thereof
CN111046678A (en)*2019-12-102020-04-21深圳市润屋科技有限公司Device and method for realizing simultaneous interpretation by extensible connection based on network
CN111985252B (en)*2020-08-212024-05-07Oppo广东移动通信有限公司 Dialogue translation method and device, storage medium and electronic device
CN114765024A (en)*2021-01-112022-07-19博泰车联网(南京)有限公司Voice translation method, device and storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN209947430U (en)*2019-03-072020-01-14北京金山安全软件有限公司Voice translation terminal, mobile terminal and translation system

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
EP2237270B1 (en)*2009-03-302012-07-04Nuance Communications, Inc.A method for determining a noise reference signal for noise compensation and/or noise reduction
KR102108500B1 (en)*2013-02-222020-05-08삼성전자 주식회사Supporting Method And System For communication Service, and Electronic Device supporting the same
JP2015060332A (en)*2013-09-182015-03-30株式会社東芝Voice translation system, method of voice translation and program
CN106454605A (en)*2016-11-302017-02-22南京小脚印网络科技有限公司Intelligent translation earphone system
CN106935240A (en)*2017-03-242017-07-07百度在线网络技术(北京)有限公司Voice translation method, device, terminal device and cloud server based on artificial intelligence
CN107341148A (en)*2017-06-272017-11-10深圳市沃特沃德股份有限公司Interpretation method, interpreting equipment and translation system
CN208044602U (en)*2017-12-292018-11-02科大讯飞股份有限公司Supplementary translation equipment and translation system
CN207764800U (en)*2017-12-292018-08-24科大讯飞股份有限公司Interpreting equipment and translation system
CN108415904B (en)*2018-01-122022-04-19广东思派康电子科技有限公司 A dual-channel real-time translation method
CN108899018A (en)*2018-05-082018-11-27深圳市沃特沃德股份有限公司automatic translation device and method
CN108920470A (en)*2018-06-122018-11-30深圳市合言信息科技有限公司A kind of language of automatic detection audio and the method translated
CN108959273B (en)*2018-06-152022-07-08Oppo广东移动通信有限公司Translation method, electronic device and storage medium
CN109359307B (en)*2018-10-172023-07-07深圳市沃特沃德股份有限公司Translation method, device and equipment for automatically identifying languages

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN209947430U (en)*2019-03-072020-01-14北京金山安全软件有限公司Voice translation terminal, mobile terminal and translation system

Also Published As

Publication numberPublication date
CN109754781A (en)2019-05-14

Similar Documents

PublicationPublication DateTitle
CN109754781B (en) Voice translation terminal, mobile terminal, translation system, translation method and device thereof
US20200285815A1 (en)Speech translation terminal, mobile terminal, translation system, translation method, and translation device
KR100819928B1 (en) Voice recognition device of mobile terminal and its method
CN109147784B (en)Voice interaction method, device and storage medium
US11120813B2 (en)Image processing device, operation method of image processing device, and computer-readable recording medium
US10599785B2 (en)Smart sound devices and language translation system
CN108197572B (en)Lip language identification method and mobile terminal
WO2011048826A1 (en)Speech translation system, control apparatus and control method
US9009033B2 (en)Real-time voice recognition on a handheld device
CN108260051B (en)Voice remote control system, portable transmission equipment and intelligent equipment
CN113724718A (en)Target audio output method, device and system
US9779731B1 (en)Echo cancellation based on shared reference signals
CN108733656A (en)Speech translation apparatus, system and method
WO2014173325A1 (en)Gutturophony recognition method and device
US20170364516A1 (en)Linguistic model selection for adaptive automatic speech recognition
CN109545216A (en)A kind of audio recognition method and speech recognition system
JP6457706B1 (en) Translation system, translation method, and translation apparatus
CN202721771U (en) TV system with audio recognition function
CN209947430U (en)Voice translation terminal, mobile terminal and translation system
CN111713119B (en)Headset, headset system and method in headset system
US7478046B2 (en)Server-client type speech recognition apparatus and method
CN113992468A (en)Smart home voice control method
US20040034528A1 (en)Server and receiving terminal
KR102495028B1 (en)Sound Device with Function of Whistle Sound Recognition
CN208337877U (en)A kind of loudspeaker of voice control

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp