Movatterモバイル変換


[0]ホーム

URL:


CN113810725B - Video processing method, device, storage medium and video communication terminal - Google Patents

Video processing method, device, storage medium and video communication terminal
Download PDF

Info

Publication number
CN113810725B
CN113810725BCN202111187261.5ACN202111187261ACN113810725BCN 113810725 BCN113810725 BCN 113810725BCN 202111187261 ACN202111187261 ACN 202111187261ACN 113810725 BCN113810725 BCN 113810725B
Authority
CN
China
Prior art keywords
video
target object
frame rate
scene
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111187261.5A
Other languages
Chinese (zh)
Other versions
CN113810725A (en
Inventor
周建军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Huasheng Software Technology Co ltd
TCL Commercial Information Technology Huizhou Co Ltd
Original Assignee
Shenzhen Huasheng Software Technology Co ltd
TCL Commercial Information Technology Huizhou Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Huasheng Software Technology Co ltd, TCL Commercial Information Technology Huizhou Co LtdfiledCriticalShenzhen Huasheng Software Technology Co ltd
Priority to CN202111187261.5ApriorityCriticalpatent/CN113810725B/en
Publication of CN113810725ApublicationCriticalpatent/CN113810725A/en
Application grantedgrantedCritical
Publication of CN113810725BpublicationCriticalpatent/CN113810725B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Classifications

Landscapes

Abstract

Translated fromChinese

本申请公开了一种视频处理方法、装置、存储介质及视频通讯终端,包括:获取拍摄场景中目标对象区域的第一视频;获取拍摄场景中非目标对象区域的第二视频,第一视频的视频帧率大于第二视频的视频帧率;将第一视频和第二视频发送至接收终端。通过得到视频帧率低于第一视频的第二视频,从而降低了视频整体的数据量,提高了传输效率,保证了接收终端接收第一视频和第二视频的进度,并不会造成第一视频和第二视频在接收终端上播放卡顿的情况,从而提高了视频显示效果。

The present application discloses a video processing method, device, storage medium and video communication terminal, including: obtaining a first video of a target object area in a shooting scene; obtaining a second video of a non-target object area in the shooting scene, wherein the video frame rate of the first video is greater than the video frame rate of the second video; and sending the first video and the second video to a receiving terminal. By obtaining a second video with a lower video frame rate than the first video, the overall data volume of the video is reduced, the transmission efficiency is improved, the progress of the first video and the second video received by the receiving terminal is guaranteed, and the first video and the second video will not be played on the receiving terminal. The situation of freeze, thereby improving the video display effect.

Description

Video processing method, device, storage medium and video communication terminal
Technical Field
The present application relates to the field of video communication technologies, and in particular, to a video processing method, a device, a storage medium, and a video communication terminal.
Background
Aiming at the video conference requirement, terminal equipment such as a conference machine and the like is required to be used, video acquisition and transmission are carried out at a sending terminal, and the transmitted video is received and displayed at a receiving terminal, so that video communication between the two terminals is realized.
The too high or too low resolution of the video collected by the sending terminal can affect the display effect of the video at the receiving terminal, for example, the video image displayed by the receiving terminal is blurred, noise or mosaic occurs, or display jamming phenomenon exists.
Disclosure of Invention
The embodiment of the application provides a video processing method, a video processing device, a storage medium and a video communication terminal, which can improve video display effect.
In a first aspect, an embodiment of the present application provides a video processing method, including:
Acquiring a first video of a target object area in a shooting scene;
Acquiring a second video of a non-target object area in a shooting scene, wherein the video frame rate of the first video is larger than that of the second video;
and transmitting the first video and the second video to the receiving terminal.
In a second aspect, an embodiment of the present application further provides a video processing apparatus, including:
the first video processing module is used for acquiring a first video of a target object area in a shooting scene;
The second video processing module is used for acquiring a second video of a non-target object area in a shooting scene, and the video frame rate of the first video is larger than that of the second video;
And the video sending module is used for sending the first video and the second video to the receiving terminal.
In a third aspect, embodiments of the present application also provide a computer-readable storage medium having stored thereon a computer program which, when run on a computer, causes the computer to perform a video processing method as provided by any of the embodiments of the present application.
In a fourth aspect, an embodiment of the present application further provides a video communication terminal, including a processor and a memory, where the memory has a computer program, and the processor is configured to execute the video processing method according to any one of the embodiments of the present application by calling the computer program.
According to the technical scheme provided by the embodiment of the application, the first video of the target object area in the shooting scene is obtained in the sending terminal, and the second video of the non-target object area in the shooting scene is obtained, so that the target object area and the non-target object area in the same shooting scene are separated, the first video of the target object area and the second video of the non-target object area are obtained, wherein the video frame rate of the first video is larger than the video frame rate of the second video. Compared with the prior art that the target object area and the non-target object area are not distinguished, and the processing mode in the prior art is caused by overlarge video data volume by transmitting a video to the receiving terminal, the video is blocked, and the display effect of the video on the receiving terminal is further affected. In the embodiment of the application, the video frame rate of the first video is higher, and the video frame rate of the second video is lower, so that the overall data size of the video is reduced, and the transmission efficiency of the first video and the second video is further improved, thereby ensuring that the first video and the second video on the receiving terminal can be smoothly displayed after the first video is sent to the receiving terminal, and improving the display effect of the video. Furthermore, the first video and the second video can be transmitted respectively, the transmission pressure can be reduced, the video transmission efficiency is improved, the fluency of the transmission process is further guaranteed, after the receiving terminal receives the first video and the second video, the first video and the second video can be displayed fluently, and the display effect of the video is guaranteed. Based on the above, the video playing is not blocked during the video display, the display effect of the target object area is clear, and the mosaic or noise phenomenon is not generated.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are needed in the description of the embodiments will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present application, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a flow chart of a video processing method executed by a transmitting terminal according to an embodiment of the present application.
Fig. 2 is a flowchart of a video processing method performed by a receiving terminal according to an embodiment of the present application.
Fig. 3 is a schematic structural diagram of a video processing apparatus according to an embodiment of the present application.
Fig. 4 is a schematic structural diagram of a video communication terminal according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application. It will be apparent that the described embodiments are only some, but not all, embodiments of the application. All other embodiments, which can be made by a person skilled in the art without any inventive effort, are intended to be within the scope of the present application based on the embodiments of the present application.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the application. The appearances of such phrases in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those of skill in the art will explicitly and implicitly appreciate that the embodiments described herein may be combined with other embodiments.
The embodiment of the application provides a video processing method, and an execution subject of the video processing method can be the video processing device provided by the embodiment of the application or a video communication terminal integrated with the video processing device. The video processing device can be realized in a hardware or software mode, and the video communication terminal can be terminal equipment supporting video call, such as a smart phone, a tablet personal computer, a palm computer, a conference machine and the like.
Referring to fig. 1, fig. 1 is a flowchart illustrating a video processing method performed by a transmitting terminal according to an embodiment of the present application. The specific flow of the video processing method provided by the embodiment of the application can be as follows:
101. And acquiring a first video of a target object area in the shooting scene.
Illustratively, the transmitting terminal acquires a first video of a target object area in a shooting scene. The first video of the target object area can be shot through a camera arranged on the sending terminal, and the first video can be sent to the sending terminal after the first video of the target object area is shot through an external camera. It is to be understood that the manner of acquiring the first video is not limited herein, as long as the first video can be acquired.
The target object area refers to an area where the target object is located, and when the target object moves, the target object area changes correspondingly. The target object may be a person or a moving object, and as long as a certain attribute of the target object is set, the target object may be distinguished according to the attribute feature to obtain the first video of the target object, and based on this, any manner capable of obtaining the first video of the target object region may be used in this embodiment.
102. And acquiring a second video of the non-target object area in the shooting scene, wherein the video frame rate of the first video is larger than that of the second video.
In the same shooting scene, after the target object area is positioned, the areas except the target object area are non-target object areas, and the non-target object areas change along with the change of the target object area.
The second video may be obtained by capturing a non-target object area by a camera configured by the transmitting terminal, or may be obtained by capturing a non-target object area by an external camera.
The first video and the second video may be obtained through the same camera, or may be obtained through different cameras, which is not limited herein, as long as two videos can be obtained, one of which is the first video and the other is the second video.
The video frame rate of the first video is greater than that of the second video, and the target object is an object to be focused, so that the required display effect can be achieved only by ensuring that the first video has a higher video frame rate, and the video frame rate of the second video is only required to be lower than that of the first video, wherein the video frame rate of the second video can also be set to be the lowest value, and the video frame rate of the first video is necessarily greater than that of the second video when the first video is acquired subsequently, so that the overall data volume of the first video and the second video is reduced, and the transmission efficiency is further ensured. It should be understood that the specific setting manner of the video frame rates of the first video and the second video is not limited herein, so long as the manner of obtaining the video frame rate of the first video greater than the video frame rate of the second video can be achieved.
Through obtaining the first video with higher video frame rate and obtaining the second video with lower video frame rate, because the video frame rate of the first video is higher, the video frame rate of the second video is lower, the overall data size of the video can be reduced, and then the transmission efficiency of the first video and the second video is improved, so that after the first video is sent to a receiving terminal, the first video and the second video on the receiving terminal can be smoothly displayed, the display effect of the video is ensured, video playing is not blocked, the display effect of a target object area is clear, and mosaic or noise phenomenon can not occur.
103. And transmitting the first video and the second video to the receiving terminal.
The first video and the second video may be transmitted to the receiving terminal as two video streams, respectively, or the first video and the second video may be synthesized into one video stream and transmitted to the receiving terminal, and the data terminal analyzes the video stream, and the specific transmission mode is not limited herein, so long as the first video and the second video can be transmitted to the receiving terminal.
By way of example, the first video and the second video can be compressed and then sent to the receiving terminal, so that the transmission quantity of the video is reduced, the transmission efficiency is improved, and correspondingly, the display effect of the video at the receiving terminal is also ensured.
In particular, the application is not limited by the order of execution of the steps described, as some of the steps may be performed in other orders or concurrently without conflict.
The transmitting terminal in the embodiment of the application acquires the first video with high frame rate and the second video with low frame rate, and transmits the first video and the second video to the receiving terminal, and in the video transmission process, compared with one video frame rate in the prior art, the scheme of the embodiment of the application can reduce the data volume of the second video in the transmission process, thereby reducing the overall data volume of the first video and the second video, further improving the transmission efficiency of the first video and the second video, and smoothly playing the first video and the second video after the receiving terminal receives the first video and the second video. Meanwhile, the video frame rate of the first video is higher, the playing of the first video is not influenced, and the first video is correspondingly displayed as the target object, so that the display effect of the target object on the receiving terminal can be ensured.
It can be understood that when the first video and the second video in the embodiment of the present application are both high definition videos, by adopting the scheme of the embodiment of the present application, the data amount in the transmission process of the first video and the second video can be reduced, so that the receiving terminal is guaranteed not to have a play clamping phenomenon when receiving the first video and the second video, and further, the play fluency of the first video and the second video is guaranteed, furthermore, the first video records the video of the target object, and by smoothly playing the first video at the receiving terminal, the definition of the first video display can be guaranteed, the play fluency is improved, and further, the overall video display effect is improved.
Based on the above, the scheme provided by the embodiment of the application not only can ensure the play fluency of the video at the receiving terminal, but also can ensure the display definition of the video at the receiving terminal, thereby solving the problems that in the prior art, the video is blocked when the high-definition video is played and the video is blurred when the non-high-definition video is played, ensuring the high-definition display and the play fluency simultaneously, greatly improving the video display effect and being more beneficial to application and popularization in the video communication field.
The method described in the previous examples is described in further detail below by way of example.
The above-mentioned ways of obtaining the first video of the target object area in the shooting scene and obtaining the second video of the non-target object area in the shooting scene are various:
For example, the shooting frame rate of the camera can be set directly, so that a first video with a high video frame rate can be obtained through shooting, and a second video with a low video frame rate can be obtained through shooting.
For example, the shooting frame rate of the camera may not be set in advance, and the video frame rate of the first video may be larger than the video frame rate of the second video by performing high-frame-rate encoding processing on the video frame rate of the first video and performing low-frame-rate encoding processing on the video frame rate of the second video after shooting the first video and the second video.
It should be understood that, setting the video frame rate of the first video and setting the video frame rate of the second video may be achieved by setting the shooting frame rate of the camera, or may be achieved by performing frame rate encoding after shooting, or may be achieved by combining the mode of setting the camera and frame rate encoding, and the specific implementation is not limited herein, so long as the first video can have a higher video frame rate relative to the second video.
It should be noted that, in the embodiment of the present application, the first video with the higher video frame rate, the second video with the lower video frame rate, the video frame rate of the first video being higher/greater than that of the second video, and the high frame rate encoding and the low frame rate encoding, the descriptions about the high and the low are only expressions in relative sense, and only refer to the relative magnitudes in numerical values of the video frame rates.
In the following embodiments, two ways of capturing the first video and the second video will be described by way of example, and it should be understood that these two ways of capturing the first video and the second video should not be construed as limiting the present application.
In an embodiment, the camera may be set to obtain the first video and the second video, where step 101 includes:
shooting a target object area in a shooting scene based on a first preset frame rate to obtain a first video;
step 102 comprises:
Shooting a non-target object area in a shooting scene based on a second preset frame rate to obtain a second video, wherein the first preset frame rate is larger than the second preset frame rate.
For example, the target object region and the non-target object region may be photographed by one camera in succession, or the target object region and the non-target object region may be photographed by two cameras, respectively.
When shooting by one camera, the shooting frame rate of the camera can be set to be a first preset frame rate, and after shooting a first video of a target object, the shooting frame rate of the camera is set to be a second preset frame rate, so that a second video of a non-target object is obtained through shooting. Based on this way, shooting can be performed, the setting of the camera can be reduced, and the specific positions of the target object and the non-target object can be distinguished.
Of course, two cameras may be used to photograph respectively, where the photographing frame rate of one camera may be set to a first preset frame rate, so as to obtain a first video with the first preset frame rate, and the photographing frame rate of the other camera may be set to a second preset frame rate, so as to obtain a second video with the second preset frame rate.
In some embodiments, the AI camera may also be used to automatically capture and track face images, thereby distinguishing between target object regions and non-target object regions in the captured scene.
It is understood that specific values of the first preset frame rate and the second preset frame rate are not limited herein, and may be set by those skilled in the art according to actual needs.
In this embodiment, by setting the shooting frame rate of the camera, it is possible to directly obtain the first video with the first preset frame rate and the second video with the second preset frame rate, where the first preset frame rate is greater than the second preset frame rate, so that the subsequent need to process the video frame rate is avoided, and the video processing efficiency is improved.
In another embodiment, the video encoding process may be further performed on the obtained first video and the second video, so that the video frame rate of the first video is a first preset frame rate, and the video frame rate of the second video is a second preset frame rate.
Wherein, step 101 includes:
shooting the shooting scene to obtain a scene image of the shooting scene;
And carrying out video coding on the image content in the scene image of the target object area according to a first preset frame rate to obtain a first video.
Step 102 comprises:
And carrying out video coding on the content in the scene image of the non-target object area according to a second preset frame rate to obtain a second video, wherein the first preset frame rate is larger than the second preset frame rate.
When the scene image is shot at the same frame rate, the target object area and the non-target object area are segmented from the scene image, so that the video of the target object area and the video of the non-target object area are obtained, the video of the target object area is subjected to video coding processing according to a first preset frame rate to obtain a first video, the video of the non-target object area is subjected to video coding processing according to a second preset frame rate to obtain a second video, and the first preset frame rate is larger than the second preset frame rate, so that the video frame rate of the first video is larger than the second preset video frame rate.
For example, the segmentation of the target object region and the non-target object region may be obtained by using an image segmentation model, or by performing feature recognition on an image to extract the image content of the region where the target object is located, and since there are various ways of performing segmentation, it will be understood that any way of obtaining the first video of the target object region and the way of obtaining the second video of the non-target object region may be used in this embodiment.
According to the embodiment, the first video is obtained by carrying out video coding on the image content of the target object area at the first preset frame rate, the second video is obtained by carrying out video coding on the image content of the non-target object area at the second preset frame rate, the video frame rate of the first video and/or the second video can be freely adjusted, the real-time adjustment of the video frame rate is realized, the data transmission requirement is met, and the limitation that the camera can only set a limited number of shooting frame rates is solved.
Based on this, the first preset frame rate and the second preset frame rate in the present embodiment may not be fixed values, may be automatically adjusted according to actual needs, or may be set and adjusted by a user, and when the adjustment is automatically performed, the adjustment basis may be a network transmission amount or a memory occupation amount, and specific implementation manners are not described herein, so long as the adaptive adjustment of the first preset frame rate and/or the second preset frame rate can be achieved.
In some embodiments, after shooting the shooting scene and obtaining the scene image of the shooting scene, the method further includes:
Blurring the image content in the scene image of the non-target object area to obtain a blurred scene image;
And carrying out video coding on the virtual scene image according to a first preset frame rate to obtain a first video.
For example, after the scene image is captured, blurring processing may be performed on the image content in the scene image where the non-target object area is located, so that the image display of the non-target object is blurred, and further contrast is formed between the non-target object and the target object, so as to highlight the display effect of the target object, and make the visual display of the target object clearer.
In this embodiment, the image of the non-target object area does not need to be encoded according to the second preset frame rate, where the bit depth of each pixel point of the image can be changed by blurring the image content of the non-target object area, so that the data size of the image of the non-target object area is reduced, the transmission efficiency can be improved, and further the display effect of the target object on the receiving terminal is improved.
And transmitting the first video and the second video to the receiving terminal for playing, wherein the first video and the second video can be firstly processed by the transmitting terminal and then transmitted to the receiving terminal for direct playing and displaying, or the first video and the second video can be firstly transmitted to the receiving terminal and then processed and played on the receiving terminal. Therefore, the processing of the first video and the second video is not limited to whether it is based on the transmitting terminal or the receiving terminal.
While there are also a number of ways for the processing of the first video and the second video:
for example, the second video may be subjected to frame insertion processing, so as to obtain a video frame rate identical to that of the first video, and then the first video and the second video of each frame are spliced to obtain a spliced image and play the spliced image frame by frame.
For another example, the first video and the second video may be played simultaneously, and the first video may be played superimposed on the second video, so long as the display of the first video is not affected.
For another example, the third video may be formed by performing frame extraction processing on the second video, and further extracting an image frame in the second video, where the image frame is consistent with the time stamp of the image frame of the first video.
In the following embodiments, the two modes of frame insertion and frame extraction are merely exemplified, but it should be noted that these two modes should not be construed as limiting the present application.
In one embodiment, the second video is processed in a frame-extracting manner, and step 103 includes:
Extracting image frames matched with the first video from the second video according to the time stamps of the image frames in the first video and the second video to obtain a third video;
And transmitting the first video and the third video to the receiving terminal.
For example, the image frames in the second video may be decimated as the time stamps of the image frames in the first video, and further the decimated image frames may be used as a third video, where each image frame in the third video has a matching image frame in the first video.
The consistency with the time stamp refers to that the image frames which are synchronous in time with the image frames of the first video are screened out from the image frames of the second video, the image frames which are not coincident in time can be directly discarded, and therefore frame extraction processing of the second video is achieved.
After the third video is obtained, each image frame in the third video is provided with the image frame matched with the first video on the time stamp, and the image frame in the second video can be smaller in this way, so that the data size of the obtained third video is smaller, and the transmission efficiency of the first video and the third video is greatly improved.
In another embodiment, the second video is processed in an interleaved manner, and step 103 includes:
copying the image frames of the second video according to the first preset frame rate, and inserting the image frames into the second video frame by frame to obtain a third video;
And transmitting the first video and the third video to the receiving terminal.
By way of example, the replication rate of the image frames of the second video may be obtained by calculating the difference between the first preset frame rate and the second preset frame rate, and by inserting between two temporally adjacent image frames after the image frames of the second video are replicated, the frame inserting process of the second video is further implemented, the video frame rate of the second video is improved, and a third video is obtained, where the video frame rate of the third video is the same as the video frame rate of the first video.
The frame inserting mode can realize synchronous playing of the first video and the third video, and reduces the processing process of the first video and the third video. The method can realize that the simple superposition of the first video and the third video can restore the video of the shooting scene and restore the video display effect of the shooting scene.
In some embodiments, the shooting scene is a conference scene, the target object region is a region in which a speaker in the conference scene is located, and the non-target object region is a region other than the speaker, wherein the non-target object region may be a conference background;
acquiring a first video of a target object area in a shooting scene, including:
Shooting a conference scene to obtain a conference image of the conference scene;
Acquiring a first video of a speaker;
acquiring a second video of a non-target object area in a shooting scene, wherein the second video comprises:
and matching the conference images from a preset conference database to obtain a second video.
For example, only a speaker in a conference scene may be photographed, and video encoding processing may be performed, thereby obtaining a first video. Whereas for a second video of a non-target object region, it can be derived by matching the history data.
The second video obtained after the processing mode of the embodiment can be stored in the preset conference database, when the second video is needed to be used as a background, the video corresponding to the conference image can be directly matched from the preset conference database, and when the second video is not matched, the second video can be obtained through shooting.
It can be understood that the target object may be a conference speaker or an object moving in a conference, and since the images except for the target object may be referred to as a conference background and the conference background is static, only the first video of the conference speaker may be captured, and the video may not be captured for the conference background, so long as capturing the first video is ensured, and the first video is sent to the receiving terminal.
Furthermore, since the conference background is stationary, the frame rate of the second video of the conference background is reduced, and the second video does not change greatly when being played, and the display effect is not affected. Of course, a solid background picture can be set instead of the second video, thereby greatly reducing the video data amount and improving the video display effect.
According to the embodiment of the application, the conference scene is shot, so that the conference background can be matched according to the shot conference image, wherein the conference background can be video or picture, the conference background is prevented from being shot, or the conference image is processed to obtain the video of the conference background, the process of obtaining the second video is omitted, the video processing efficiency is greatly improved, and the video display effect is greatly improved for the receiving terminal.
As can be seen from the above, the receiving terminal according to the embodiment of the present application may be used as a transmitting terminal to implement the video processing method described above, and when the receiving terminal receives the video transmitted from the transmitting terminal, the receiving terminal may process the two received videos and play the two processed videos, or may directly play the two processed videos. In the following embodiments, a manner of processing two received videos and then playing the processed videos will be described in detail.
As shown in fig. 2, fig. 2 is a flowchart of a video processing method performed by a receiving terminal according to an embodiment of the present application, and in this embodiment, after the step 103, the method further includes:
201. Receiving a fourth video and a fifth video, wherein the video frame rate of the fourth video is greater than that of the fifth video;
202. Splicing the image frames in the fifth video into the image frames of the fourth video according to the time stamp to obtain a spliced sixth video;
203. and playing the sixth video.
If the fifth video is a video after the frame extraction processing performed at the transmitting terminal, each image frame of the fifth video has a matched image frame in the fourth video, and the image frames of the fifth video and the fourth video may be directly spliced according to the timestamp, where the splicing refers to the alignment processing of the target object area in the fourth video and the non-target object area in the fifth video, so that the spliced sixth video can achieve the effect of restoring the image of the shooting scene in visual effect.
If the fifth video is not subjected to frame extraction processing at the sending terminal, different mixed playing can be directly performed on the fourth video and the fifth video, and only the target object area and the non-target object area are aligned in the image frame.
Of course, the frame inserting process may be performed on the fifth video at the receiving terminal, and the frame inserting process may refer to the content mentioned in the above embodiment, which is not described herein again.
The processing of the fifth video and the fourth video can also be performed through a neural network repair model, and the repaired sixth video can be obtained by inputting the fifth video and the fourth video into the neural network repair model. The neural network repair model can be obtained through sample set training.
As can be seen from the above, when the video processing method provided by the embodiment of the application is used as a transmitting terminal, the first video with higher video frame rate and the second video with lower video frame rate can be obtained, and because the video frame rate of the first video is higher and the video frame rate of the second video is lower, the overall data size of the video can be reduced, and the transmission efficiency of the first video and the second video can be further improved, so that after the first video is transmitted to the receiving terminal, the first video and the second video on the receiving terminal can be smoothly displayed, the display effect of the video is ensured, video playing is not blocked, the display effect of the target object area is clear, and mosaic or noise phenomenon can not occur. Moreover, the frame extraction processing is performed on the second video, so that the data volume can be greatly reduced, the transmission efficiency is improved, the transmission pressure is relieved and the transmission efficiency is improved by dividing the video into the first video and the second video for transmission respectively.
A video processing apparatus is also provided in an embodiment. Referring to fig. 3, fig. 3 is a schematic diagram of a video processing apparatus 300 according to an embodiment of the application. The video processing device 300 is applied to a video communication terminal, and the video processing device 300 includes a first video processing module 301, a second video processing module 302, and a video transmitting module 303, as follows:
The first video processing module 301 is configured to obtain a first video of a target object area in a shooting scene;
the second video processing module 302 is configured to obtain a second video of a non-target object area in the shooting scene, where a video frame rate of the first video is greater than a video frame rate of the second video;
The video sending module 303 is configured to send the first video and the second video to a receiving terminal.
In some embodiments, the first video processing module 301 is further configured to:
Extracting image frames matched with the first video from the second video according to the time stamps of the image frames in the first video and the second video to obtain a third video;
And transmitting the first video and the third video to the receiving terminal. In some embodiments, the first video processing module 301 is further configured to:
shooting the shooting scene to obtain a scene image of the shooting scene;
And carrying out video coding on the image content in the scene image of the target object area according to a first preset frame rate to obtain a first video.
In some embodiments, the first video processing module 301 is further configured to:
Blurring the image content in the scene image of the non-target object area to obtain a blurred scene image;
And carrying out video coding on the virtual scene image according to a first preset frame rate to obtain a first video.
In some embodiments, the second video processing module 302 is further configured to:
And carrying out video coding on the content in the scene image of the non-target object area according to a second preset frame rate to obtain a second video, wherein the first preset frame rate is larger than the second preset frame rate.
In some embodiments, the video transmission module 303 is further configured to:
Extracting image frames matched with the first video from the second video according to the time stamps of the image frames in the first video and the second video to obtain a third video;
And transmitting the first video and the third video to the receiving terminal.
In some embodiments, the shooting scene is a conference scene, and the target object area is an area where a speaker in the conference scene is located:
the first video processing module 301 is further configured to:
Shooting a conference scene to obtain a conference image of the conference scene;
Acquiring a first video of a speaker;
the second video processing module 302 is further configured to:
and matching the conference images from a preset conference database to obtain a second video.
In some embodiments, the video processing apparatus further comprises:
the video receiving module 304 is configured to receive a fourth video and a fifth video, where a video frame rate of the fourth video is greater than a video frame rate of the fifth video;
splicing the image frames in the fifth video into the image frames of the fourth video according to the time stamp to obtain a spliced sixth video;
And playing the sixth video.
It should be noted that, the video processing apparatus 300 provided in the embodiment of the present application belongs to the same concept as the video processing method in the above embodiment, and any method provided in the embodiment of the video processing method may be implemented by the video processing apparatus 300, and detailed implementation processes of the method are shown in the embodiment of the video processing method, which is not described herein.
As can be seen from the above, when the video processing device provided by the embodiment of the application is used as a transmitting terminal, the first video with a higher video frame rate and the second video with a lower video frame rate can be obtained, and because the video frame rate of the first video is higher and the video frame rate of the second video is lower, the overall data size of the video can be reduced, and the transmission efficiency of the first video and the second video can be further improved, so that after the first video is transmitted to the receiving terminal, the first video and the second video on the receiving terminal can be smoothly displayed, the display effect of the video is ensured, video playing is not blocked, the display effect of the target object area is clear, and mosaic or noise phenomenon can not occur. Moreover, the frame extraction processing is performed on the second video, so that the data volume can be greatly reduced, the transmission efficiency is improved, the transmission pressure is relieved and the transmission efficiency is improved by dividing the video into the first video and the second video for transmission respectively.
The embodiment of the application also provides a video communication terminal which can be terminal equipment supporting video call, such as a smart phone, a tablet personal computer, a palm computer, a conference machine and the like. Fig. 4 is a schematic structural diagram of a video communication terminal according to an embodiment of the present application, as shown in fig. 4. The video communication terminal 400 includes a processor 401 having one or more processing cores, a memory 402 having one or more computer readable storage media, and a computer program stored on the memory 402 and executable on the processor. The processor 401 is electrically connected to the memory 402. It will be appreciated by those skilled in the art that the video communication terminal structure shown in the drawings does not constitute a limitation on the video communication terminal and may include more or less components than those shown, or may combine certain components, or may have a different arrangement of components.
The processor 401 is a control center of the video communication terminal 400, connects various parts of the entire video communication terminal 400 using various interfaces and lines, and performs various functions of the video communication terminal 400 and processes data by running or loading software programs and/or modules stored in the memory 402 and calling data stored in the memory 402, thereby performing overall monitoring of the video communication terminal 400.
In the embodiment of the present application, the processor 401 in the video communication terminal 400 loads the instructions corresponding to the processes of one or more application programs into the memory 402 according to the following steps, and the processor 401 executes the application programs stored in the memory 402, so as to implement various functions:
Acquiring a first video of a target object area in a shooting scene;
Acquiring a second video of a non-target object area in a shooting scene, wherein the video frame rate of the first video is larger than that of the second video;
and transmitting the first video and the second video to the receiving terminal.
The specific implementation of each operation above may be referred to the previous embodiments, and will not be described herein.
In the foregoing embodiments, the descriptions of the embodiments are emphasized, and for parts of one embodiment that are not described in detail, reference may be made to related descriptions of other embodiments.
As can be seen from the above, when the video communication terminal provided in this embodiment is used as a transmitting terminal, the first video with a higher video frame rate and the second video with a lower video frame rate can be obtained, and because the video frame rate of the first video is higher and the video frame rate of the second video is lower, the overall data size of the video can be reduced, and the transmission efficiency of the first video and the second video can be further improved, so that after the first video is transmitted to the receiving terminal, the first video and the second video on the receiving terminal can be smoothly displayed, the display effect of the video is ensured, the video playing is not blocked, the display effect of the target object area is clear, and the mosaic or noise phenomenon can not occur. Moreover, the frame extraction processing is performed on the second video, so that the data volume can be greatly reduced, the transmission efficiency is improved, the transmission pressure is relieved and the transmission efficiency is improved by dividing the video into the first video and the second video for transmission respectively.
Those of ordinary skill in the art will appreciate that all or a portion of the steps of the various methods of the above embodiments may be performed by instructions, or by instructions controlling associated hardware, which may be stored in a computer-readable storage medium and loaded and executed by a processor.
To this end, an embodiment of the present application provides a computer readable storage medium, and those skilled in the art will understand that all or part of the steps in implementing the method of the above embodiment may be implemented by a program to instruct related hardware, and the program may be stored in a computer readable storage medium, where the program when executed includes the following steps:
Acquiring a first video of a target object area in a shooting scene;
Acquiring a second video of a non-target object area in a shooting scene, wherein the video frame rate of the first video is larger than that of the second video;
and transmitting the first video and the second video to the receiving terminal.
The specific implementation of each operation above may be referred to the previous embodiments, and will not be described herein.
The storage medium may be ROM/RAM, magnetic disk, optical disk, etc. The steps in any video processing method provided by the embodiment of the present application can be executed by the computer program stored in the storage medium, so that the beneficial effects that can be achieved by any video processing method provided by the embodiment of the present application can be achieved, and detailed descriptions of the foregoing embodiments are omitted.
The foregoing describes in detail a video processing method, apparatus, medium and video communication terminal provided by the embodiments of the present application, and specific examples are used herein to describe the principles and embodiments of the present application, and the above description of the embodiments is only for aiding in understanding the method and core concept of the present application, and meanwhile, for those skilled in the art, according to the concept of the present application, there are variations in the specific embodiments and application ranges, and in summary, the present disclosure should not be construed as limiting the present application.

Claims (7)

CN202111187261.5A2021-10-122021-10-12Video processing method, device, storage medium and video communication terminalActiveCN113810725B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202111187261.5ACN113810725B (en)2021-10-122021-10-12Video processing method, device, storage medium and video communication terminal

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202111187261.5ACN113810725B (en)2021-10-122021-10-12Video processing method, device, storage medium and video communication terminal

Publications (2)

Publication NumberPublication Date
CN113810725A CN113810725A (en)2021-12-17
CN113810725Btrue CN113810725B (en)2025-04-08

Family

ID=78897545

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202111187261.5AActiveCN113810725B (en)2021-10-122021-10-12Video processing method, device, storage medium and video communication terminal

Country Status (1)

CountryLink
CN (1)CN113810725B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN114520873B (en)*2021-12-312024-10-01广州文远知行科技有限公司Video acceleration method and system for sweeper

Citations (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN105635624A (en)*2014-10-272016-06-01华为技术有限公司Video image processing method, device and system
JP2019003325A (en)*2017-06-132019-01-10キヤノン株式会社 Image processing apparatus, image processing method, and program
CN111405199A (en)*2020-03-272020-07-10维沃移动通信(杭州)有限公司Image shooting method and electronic equipment

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
JP4680087B2 (en)*2006-02-152011-05-11三菱電機株式会社 Image transmission device
JP7042571B2 (en)*2017-08-102022-03-28キヤノン株式会社 Image processing device and its control method, program
CN109068052B (en)*2018-07-242020-11-10努比亚技术有限公司Video shooting method, mobile terminal and computer readable storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN105635624A (en)*2014-10-272016-06-01华为技术有限公司Video image processing method, device and system
JP2019003325A (en)*2017-06-132019-01-10キヤノン株式会社 Image processing apparatus, image processing method, and program
CN111405199A (en)*2020-03-272020-07-10维沃移动通信(杭州)有限公司Image shooting method and electronic equipment

Also Published As

Publication numberPublication date
CN113810725A (en)2021-12-17

Similar Documents

PublicationPublication DateTitle
CN108650542B (en)Method for generating vertical screen video stream and processing image, electronic equipment and video system
US7852368B2 (en)Method and apparatus for composing images during video communications
US20030128273A1 (en)Video processing apparatus, control method therefor, and storage medium
US9305331B2 (en)Image processor and image combination method thereof
US20170134654A1 (en)Video recording method and device
CN105827935B (en)A kind of method and terminal of terminal sectional drawing
JP6946566B2 (en) Static video recognition
CN107040794A (en)Video broadcasting method, server, virtual reality device and panoramic virtual reality play system
CN112423110A (en)Live video data generation method and device and live video playing method and device
CN113973190A (en)Video virtual background image processing method and device and computer equipment
KR20130140904A (en)System and method for video caption re-overlaying for video adaptation and retargeting
DE102020108357A1 (en) RE-ENCODING PREDICTED IMAGES IN LIVE VIDEOSTREAM APPLICATIONS
CN106713942B (en)Video processing method and device
CN107786803A (en)A kind of image generating method, device and terminal device
CN111757087A (en)VR video processing method and device and electronic equipment
CN113438550A (en)Video playing method, video conference method, live broadcasting method and related devices
CN113810725B (en)Video processing method, device, storage medium and video communication terminal
CN113014817B (en) Method, device and electronic equipment for obtaining high-definition high-frame video
CN107491934B (en)3D interview system based on virtual reality
CN112601099A (en)Live image processing method and device, storage medium and electronic equipment
CN112887653A (en)Information processing method and information processing device
CN116208812A (en) A video frame interpolation method and system based on stereo event and intensity camera
CN116489265A (en)Method for real-time link of Maya to UE5 screen capturing
CN106851105A (en)It is a kind of to record the method and apparatus that high-resolution moves facial image
KR102637147B1 (en)Vertical mode streaming method, and portable vertical mode streaming system

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp