Movatterモバイル変換


[0]ホーム

URL:


CN119006686A - Combined scene experience generation system and device based on XR technology - Google Patents

Combined scene experience generation system and device based on XR technology
Download PDF

Info

Publication number
CN119006686A
CN119006686ACN202410974424.1ACN202410974424ACN119006686ACN 119006686 ACN119006686 ACN 119006686ACN 202410974424 ACN202410974424 ACN 202410974424ACN 119006686 ACN119006686 ACN 119006686A
Authority
CN
China
Prior art keywords
data
module
user
scene
light
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202410974424.1A
Other languages
Chinese (zh)
Inventor
刘方勇
叶康健
江水瑶
廖鹏
程建辉
姚俊光
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Tianqing Digital Co ltd
Original Assignee
Shenzhen Tianqing Digital Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Tianqing Digital Co ltdfiledCriticalShenzhen Tianqing Digital Co ltd
Priority to CN202410974424.1ApriorityCriticalpatent/CN119006686A/en
Publication of CN119006686ApublicationCriticalpatent/CN119006686A/en
Pendinglegal-statusCriticalCurrent

Links

Classifications

Landscapes

Abstract

Translated fromChinese

本发明公开了一种基于XR技术组合场景体验生成系统及装置,涉及拓展现实技术领域,包括用户交互模块、建模模块、场景数据管理模块、设备适配支持模块、音效模拟模块、网络通信模块、内容更新模块;通过设置有字幕显示模块,可以对用户的语音信号的频率进行识别,并将用户的语音信号转化为文字信息,使得可以通过字幕对他人说话的内容进行显示,使得在网络不加或者语言不通时,也可以进行多人交互,使得多人交互系统的适用性更广,适应性更强,将光线接收数据、模拟边缘数据以及光线衰减数据输入蒙特卡洛光线追踪算法进行模拟光线信息生成,得到模拟XR场景中模拟光线信息,从而使虚拟场景中的光感更真实。

The present invention discloses a system and device for generating scene experience based on XR technology combination, which relates to the field of extended reality technology, including a user interaction module, a modeling module, a scene data management module, a device adaptation support module, a sound effect simulation module, a network communication module, and a content update module; by providing a subtitle display module, the frequency of the user's voice signal can be identified, and the user's voice signal can be converted into text information, so that the content of other people's speech can be displayed through subtitles, so that multi-person interaction can be carried out when the network is not connected or the language is not communicated, so that the multi-person interaction system has wider applicability and stronger adaptability, and the light reception data, simulated edge data and light attenuation data are input into the Monte Carlo ray tracing algorithm to generate simulated light information, and the simulated light information in the simulated XR scene is obtained, so that the light sensation in the virtual scene is more realistic.

Description

Combined scene experience generation system and device based on XR technology
Technical Field
The invention relates to the technical field of augmented reality, in particular to a scene experience generating system and device based on an XR technology.
Background
The extended display (XR) comprises various forms of XR such as Augmented Reality (AR), virtual Reality (VR), mixed Reality (MR) and the like, and the appearance and the use of the XR also enable interaction in a virtual space; XR scene technology has made significant progress in a number of areas including entertainment, medical, education, training and simulation. These techniques can provide an immersive virtual environment that enables a user to interact with digitized characters and objects. However, the effective virtual environment needs high-quality digital images and visual effects and user-friendly interactivity, but when the language of the crowd with multi-person interaction is not feasible, the difficulty of the multi-person interaction is caused, and basically only the crowd with the same language can interact in a multi-person interaction scene, so that the adaptability is not strong.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a combined scene experience generating system and device based on XR technology, and solves the problems in the background technology.
In order to achieve the above purpose, the invention is realized by the following technical scheme: the system and the device for generating the combined scene experience based on the XR technology comprise a user interaction module, a modeling module, a scene data management module, a device adaptation support module, an audio simulation module, a network communication module and a content updating module;
The user interaction module is used for acquiring interactive voice data and limb action images of a target user and identifying user intention, obtaining the intention of the target user and converting the intention into an instruction which can be understood by a system so as to realize interactive operation on a virtual scene, and the user interaction module also comprises a feedback module for providing feedback in visual and auditory modes and enhancing the immersion of the user;
The modeling module is used for carrying out environment scanning on the target scene area to obtain environment depth data and space layout data, and carrying out data cleaning to obtain target three-dimensional data; extracting spatial features of the target three-dimensional data and the environmental depth data, and registering depth information through an iterative nearest point algorithm to obtain an initial spatial model;
The scene data management module is used for storing, loading and editing geometric data in an XR scene, supports FBX, OBJ, GLTF data formats, is provided with a data compression and buffering mechanism so as to improve the loading speed and the system performance, and also comprises an editing tool for a developer to quickly create and modify the scene;
The device adaptation support module is used for identifying the type and the performance of the device and ensuring that an XR scene experience system can normally run on various devices, and the device adaptation support module can automatically adjust rendering parameters and interaction modes so as to ensure the consistency and the fluency of user experience;
the sound effect simulation module is used for simulating sound effects in an XR scene, including environmental sound, background music and sound source localization, the immersion of a user is further enhanced through high-quality sound effect processing, and the sound effect simulation system supports various audio formats and coding modes and has real-time processing and sound mixing capabilities;
The network communication module is used for realizing real-time communication and data synchronization between users, and the users can interact and cooperate with other users in a virtual scene through the network communication module; the network communication module supports various network protocols and transmission modes so as to ensure the stability and instantaneity of data transmission;
The content updating module is used for managing and maintaining content resources in the XR scene experience system, and ensuring that the system is always kept in an optimal state and providing the latest functions and experience for users by regularly releasing updating packages and repairing patches; the content updating module can provide perfect user support and service to solve the problems encountered by users in the using process;
Preferably, the modeling module further includes a 3D rendering engine module, where the 3D rendering engine module is configured to calculate, in real time, a position, a ray, and a material property of an object in the virtual space according to a position and a viewing angle of a user, calculate, in real time, a color and a brightness of each pixel based on simulated ray information, and generate a high-quality image, and the 3D rendering engine module further has an efficient performance optimization mechanism, so as to ensure that a smooth frame rate can be maintained in a complex scene.
Preferably, the simulated light information in the 3D rendering engine module is used for identifying a shadow area of the simulated light virtual object to obtain the shadow area; and constructing the fuzzy edge data of the shadow area to obtain simulated edge data, and simultaneously, calculating the light attenuation data of the shadow area through a light attenuation data calculation formula to obtain the light attenuation data, wherein the light attenuation data calculation formula is as follows:
Wherein A (P) is light attenuation data, Isource is light intensity data, and P represents a P point of a shadow area in the simulated light virtual object; d is the distance from the P point to the light source; sigma (x) is the medium absorption coefficient of the simulated ray virtual object; μ is a central position parameter of the shadow region; σshadow is the shadow area; k is the shape parameter of the shadow region; alpha is the scattering coefficient;
And inputting the light receiving data, the simulated edge data and the light attenuation data into the Monte Carlo ray tracing algorithm to generate simulated light information, so as to obtain the simulated light information in the simulated XR scene.
Preferably, the gesture recognition of the user interaction module captures gesture data of a user through a depth perception or image capturing device (such as an RGB-D camera, a depth camera and the like), records video or image data when the user executes various gestures, removes noise in the image or video, improves data quality, and segments gesture areas from the background for subsequent processing.
Preferably, the user gesture data is converted into a uniform size and format so as to facilitate comparison and analysis, shape features such as outlines, boundaries, corner points and the like of gestures are extracted, dynamic features such as motion tracks, speeds, accelerations and the like of the gestures are analyzed, and the gestures are classified by using a trained machine learning model (such as a support vector machine, a neural network and the like); training the model by using the marked gesture data set so that different gestures can be recognized; defining one or more instructions or operations associated therewith for each gesture; when the model identifies the gesture of the user, converting the gesture into a corresponding instruction according to the gesture mapping relation; and sending the generated instruction to the XR system or the application program so as to execute corresponding operation.
Preferably, the modeling module performs data cleaning on the environmental depth data and the spatial layout data, and performs definition correction on the spatial layout data through a depth image super-resolution algorithm to obtain corrected layout data; performing edge smoothing processing on the corrected layout data to obtain smoothed layout data; filling the missing data into the removed depth data to obtain filling depth data; and inputting the smooth layout data and the filling depth data into a point cloud conversion algorithm to perform space data conversion to obtain the target three-dimensional data.
Preferably, the user interaction module further comprises a binocular stereo correction module, the binocular stereo correction module is used for improving efficiency and accuracy of stereo matching and eliminating image distortion, and the binocular stereo correction module corrects a calculation formula as follows:
dst(x,y)=src(mapx(x,y),mapy(x,y))
Wherein src is a binocular camera image, dst (x, y) is a corrected binocular camera image, map_x and map_y are mapping matrices, in dst (x, y), x is the abscissa of the binocular camera corrected image, and y is the ordinate of the binocular camera corrected image; in map_x (x, y) or map_y (x, y), x is the abscissa of the binocular camera image, and y is the ordinate of the binocular camera image.
Preferably, the adapter support module further comprises a gyroscope unit, an eyeball tracking unit and a display unit, wherein the gyroscope unit is used for collecting the rotation angle of the head of the user, the eyeball tracking unit is used for tracking the eyeballs of the user and determining the position where the eyes of the user stay, and the display unit comprises a voice recognition unit, a voice conversion unit, a frequency analysis unit, a voice distinguishing unit and a text display unit;
Preferably, the voice recognition unit is used for recognizing the voice signal of the user, the frequency analysis unit is used for analyzing the frequency of the voice signal recognized by the voice recognition unit, the voice distinguishing unit judges whether the voice signal belongs to the user according to the analysis result of the frequency analysis unit, the voice conversion unit is used for converting the voice signal belonging to the user into text information, and the text display unit is used for displaying the text information converted by the voice conversion unit on the display unit of other people in the multi-person interaction scene in the form of subtitles.
Preferably, the device comprises an XR headset, wherein the XR headset is used for realizing user man-machine interaction, two sides of the XR headset are connected with a frame, the frame is used for wearing the XR headset on the head of a user, the inside of the XR headset is electrically connected with headphones, headphones on two sides of the XR headset are used for realizing feedback virtual scene audio, a display is embedded on the front surface of the XR headset, the display is used for feeding back real-time pictures of the virtual scene, a scanning device is installed at the top of the XR headset, and the scanning device is used for scanning gesture instructions of the user.
The invention provides a combined scene experience generating system and device based on an XR technology. The device comprises the following
The beneficial effects are that:
According to the system and the device for generating the combined scene experience based on the XR technology, the subtitle display module is arranged, so that the frequency of a voice signal of a user can be identified, the voice signal of the user is converted into character information, the character information of the user is displayed on the display unit of the other person during multi-person interaction through the character display unit, the content of speaking of the other person can be displayed through the subtitle, the multi-person interaction can be performed when a network is poor or a language is not feasible, the applicability of the multi-person interaction system is wider, the applicability is stronger, the light receiving data, the simulation edge data and the light attenuation data are input into a Monte Carlo ray tracing algorithm to generate simulated light information, and the simulated light information in a simulated XR scene is obtained, so that the light sensation in a virtual scene is more real.
Drawings
FIG. 1 is a schematic diagram of the overall structure of the present invention;
Fig. 2 is a schematic diagram of the module structure of the present invention.
In the figure, 1-XR headset, 2-gantry, 3-scanning device, 4-earphone, 5-display.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The system comprises a user interaction module, a modeling module, a scene data management module, a device adaptation support module, an audio simulation module, a network communication module and a content updating module;
the user interaction module is used for acquiring the interaction voice data and the limb action image of the target user and identifying the user intention to obtain the target user intention, converting the target user intention into an instruction which can be understood by the system so as to realize the interaction operation of the virtual scene, and the user interaction module also comprises a feedback module for providing feedback in visual and auditory forms and enhancing the immersion of the user;
The modeling module is used for carrying out environment scanning on the target scene area to obtain environment depth data and space layout data, and carrying out data cleaning to obtain target three-dimensional data; extracting spatial features of the target three-dimensional data and the environmental depth data, and registering depth information through an iterative nearest point algorithm to obtain an initial spatial model;
The scene data management module is used for storing, loading and editing geometric data in the XR scene, supports FBX, OBJ, GLTF data formats, is provided with a data compression and buffering mechanism so as to improve the loading speed and the system performance, and simultaneously comprises an editing tool for a developer to quickly create and modify the scene;
The device adaptation support module is used for identifying the type and the performance of the device, ensuring that the XR scene experience system can normally operate on various devices, and automatically adjusting rendering parameters and interaction modes so as to ensure the consistency and the fluency of user experience;
The sound effect simulation module is used for simulating sound effects in an XR scene, including environmental sounds, background music and sound source localization, the immersion of a user is further enhanced through high-quality sound effect processing, and the sound effect simulation system supports various audio formats and coding modes and has real-time processing and sound mixing capabilities;
the network communication module is used for realizing real-time communication and data synchronization between users, and the users can interact and cooperate with other users in a virtual scene through the network communication module; the network communication module supports various network protocols and transmission modes so as to ensure the stability and real-time performance of data transmission;
The content updating module is used for managing and maintaining content resources in the XR scene experience system, and ensuring that the system is always kept in an optimal state and provides the latest functions and experience for users by regularly releasing updating packages and repairing patches; the content updating module can provide perfect user support and service to solve the problems encountered by users in the using process;
the modeling module further comprises a 3D rendering engine module, the 3D rendering engine module is used for calculating the position, light and material properties of an object in the virtual space in real time according to the position and the visual angle of a user, calculating the color and the brightness of each pixel in real time based on the simulated light information, generating a high-quality image, and the 3D rendering engine module is further provided with an efficient performance optimization mechanism so as to ensure that the smooth frame rate can be maintained in a complex scene.
The simulated light information in the 3D rendering engine module is used for identifying a shadow area of the simulated light virtual object to obtain the shadow area; and constructing fuzzy edge data of the shadow area to obtain simulated edge data, and simultaneously, calculating light attenuation data of the shadow area through a light attenuation data calculation formula to obtain the light attenuation data, wherein the light attenuation data calculation formula is as follows:
Wherein A (P) is light attenuation data, Isource is light intensity data, and P represents a P point of a shadow area in the simulated light virtual object; d is the distance from the P point to the light source; sigma (x) is the medium absorption coefficient of the simulated ray virtual object; μ is a central position parameter of the shadow region; σshadow is the shadow area; k is the shape parameter of the shadow region; alpha is the scattering coefficient;
And inputting the light receiving data, the simulated edge data and the light attenuation data into a Monte Carlo ray tracing algorithm to generate simulated light information, so as to obtain the simulated light information in the simulated XR scene.
The user interaction module is used for capturing gesture data of a user through depth perception or image capturing equipment (such as an RGB-D camera, a depth camera and the like), recording video or image data when the user executes various gestures, removing noise in the image or video, improving data quality, and dividing gesture areas from the background for subsequent processing.
The user gesture data is converted into a uniform size and format so as to be convenient for comparison and analysis, shape characteristics such as outlines, boundaries, corner points and the like of gestures are extracted, dynamic characteristics such as motion tracks, speeds, accelerations and the like of the gestures are analyzed, and the gestures are classified by using a trained machine learning model (such as a support vector machine, a neural network and the like); training the model by using the marked gesture data set so that different gestures can be recognized; defining one or more instructions or operations associated therewith for each gesture; when the model identifies the gesture of the user, converting the gesture into a corresponding instruction according to the gesture mapping relation; and sending the generated instruction to the XR system or the application program so as to execute corresponding operation.
The modeling module performs data cleaning on the environment depth data and the space layout data, and performs definition correction on the space layout data through a depth image super-resolution algorithm to obtain corrected layout data; performing edge smoothing processing on the corrected layout data to obtain smoothed layout data; filling the missing data of the removed depth data to obtain filling depth data; and inputting the smooth layout data and the filling depth data into a point cloud conversion algorithm to perform space data conversion, so as to obtain target three-dimensional data.
The user interaction module further comprises a binocular stereo correction module, the binocular stereo correction module is used for improving the efficiency and accuracy of stereo matching and eliminating image distortion, and the binocular stereo correction module corrects the calculation formula as follows:
dst(x,y)=src(mapx(x,y),mapy(x,y))
Wherein src is a binocular camera image, dst (x, y) is a corrected binocular camera image, map_x and map_y are mapping matrices, in dst (x, y), x is the abscissa of the binocular camera corrected image, and y is the ordinate of the binocular camera corrected image; in map_x (x, y) or map_y (x, y), x is the abscissa of the binocular camera image, and y is the ordinate of the binocular camera image.
The adapter support module further comprises a gyroscope unit, an eyeball tracking unit and a display unit, wherein the gyroscope unit is used for collecting the rotation angle of the head of the user, the eyeball tracking unit is used for tracking the eyeball of the user and determining the stay position of the eye of the user, and the display unit comprises a voice recognition unit, a voice conversion unit, a frequency analysis unit, a voice distinguishing unit and a text display unit;
The voice recognition unit is used for recognizing the voice signal of the user, the frequency analysis unit is used for analyzing the frequency of the voice signal recognized by the voice recognition unit, the voice distinguishing unit judges whether the voice signal belongs to the user according to the analysis result of the frequency analysis unit, the voice conversion unit is used for converting the voice signal belonging to the user into text information, and the text display unit is used for displaying the text information converted by the voice conversion unit on the display unit of other people in the multi-person interaction scene in the form of subtitles.
According to the above technical solution, when the user's gaze moves from the center point to a certain point, the gyro unit detects that the XR headset is turned horizontally from the initial stateThe eyeball tracking unit detects that the iris of the user horizontally rotates from the central pointThe gyroscope and the eyeball tracking unit detect that the angle of the XR headset and the eyeball rotating leftwards is negative, the angle of the eyeball rotating rightwards is positive, and the movement analysis unit calculates the distance L of the user's gaze moving horizontally in the multi-person interaction scene according to the following formula;
The current position where the gaze of the user stays is the distance of the horizontal movement L of the center point of the multi-person interaction scene and then the distance of the vertical movement L'; through the calculation mode, the position where the user stays in the multi-person interaction scene at present can be accurately calculated and determined, at the moment, the multi-person interaction scene can be controlled by utilizing the iris of the user, and the position change of the user in the multi-person interaction scene is irregular, so that the position where the user stays in the gaze can be accurately judged by utilizing the calculation formula, and the determination of the position where the user stays in the gaze is facilitated;
Through being provided with caption display module, can discern user's speech signal's frequency to turn into text information with user's speech signal, the text information of user is shown on the display element of other people when many people are interacted through text display element, make can show the content that other people speak through the caption, make when network is bad or language is not expert, also can carry out many people and exchange, make many people interaction system's suitability wider, the adaptability is stronger, with light receiving data, simulation edge data and light decay data input Monte Carlo ray tracing algorithm carry out the formation of simulation ray information, obtain the simulation ray information in the simulation XR scene, thereby make the light sense in the virtual scene more true.
While the fundamental and principal features of the invention and advantages of the invention have been shown and described, it will be apparent to those skilled in the art that the invention is not limited to the details of the foregoing exemplary embodiments, but may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. The present embodiments are, therefore, to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein. Any reference sign in a claim should not be construed as limiting the claim concerned.
Furthermore, it should be understood that although the present disclosure describes embodiments, not every embodiment is provided with a separate embodiment, and that this description is provided for clarity only, and that the disclosure is not limited to the embodiments described in detail below, and that the embodiments described in the examples may be combined as appropriate to form other embodiments that will be apparent to those skilled in the art.

Claims (10)

Translated fromChinese
1.一种基于XR技术组合场景体验生成系统,其特征在于:包括用户交互模块、建模模块、场景数据管理模块、设备适配支持模块、音效模拟模块、网络通信模块、内容更新模块;1. A scene experience generation system based on XR technology combination, characterized by: including a user interaction module, a modeling module, a scene data management module, a device adaptation support module, a sound effect simulation module, a network communication module, and a content update module;所述用户交互模块用于获取目标用户的交互语音数据和肢体动作图像并进行用户意图识别,并将其转化为系统可以理解的指令,所述用户交互模块还包括反馈模块;The user interaction module is used to obtain the interactive voice data and body movement images of the target user and perform user intention recognition, and convert them into instructions that can be understood by the system. The user interaction module also includes a feedback module;所述建模模块用于对目标场景区域进行环境扫描,得到环境深度数据以及空间布局数据并进行数据清洗,得到目标三维数据;The modeling module is used to perform environmental scanning on the target scene area, obtain environmental depth data and spatial layout data, and perform data cleaning to obtain target three-dimensional data;所述场景数据管理模块用于存储、加载和编辑XR场景中的几何数据,所述场景数据管理模块支持FBX、OBJ、GLTF数据格式;The scene data management module is used to store, load and edit geometric data in the XR scene. The scene data management module supports FBX, OBJ, and GLTF data formats;所述设备适配支持模块用于识别设备的类型和性能,确保XR场景体验系统能够在各种设备上正常运行;The device adaptation support module is used to identify the type and performance of the device to ensure that the XR scene experience system can run normally on various devices;所述音效模拟模块用于模拟XR场景中的声音效果,包括环境音、背景音乐、声源定位,通过高质量的音效处理;The sound effect simulation module is used to simulate the sound effects in the XR scene, including ambient sound, background music, and sound source positioning, through high-quality sound effect processing;所述网络通信模块用于实现用户之间的实时通信和数据同步,通过网络通信模块,用户可以在虚拟场景中与其他用户进行互动、协作;The network communication module is used to realize real-time communication and data synchronization between users. Through the network communication module, users can interact and collaborate with other users in a virtual scene;所述内容更新模块用于管理和维护XR场景体验系统中的内容资源,通过定期发布更新包和修复补丁,确保系统始终保持在最佳状态,并为用户提供最新的功能和体验。The content update module is used to manage and maintain the content resources in the XR scene experience system. By regularly releasing update packages and repair patches, it ensures that the system is always kept in the best condition and provides users with the latest functions and experiences.2.根据权利要求1所述一种基于XR技术组合场景体验生成系统,其特征在于:所述建模模块还包括3D渲染引擎模块,所述3D渲染引擎模块用于根据用户的位置和视角,实时计算虚拟空间中物体的位置、光线、材质属性,基于模拟光线信息实时计算出每个像素的颜色和亮度,生成高质量的图像。2. According to claim 1, a scene experience generation system based on XR technology combination is characterized in that: the modeling module also includes a 3D rendering engine module, and the 3D rendering engine module is used to calculate the position, light, and material properties of objects in the virtual space in real time according to the user's position and viewing angle, and calculate the color and brightness of each pixel in real time based on the simulated light information to generate high-quality images.3.根据权利要求2所述一种基于XR技术组合场景体验生成系统,其特征在于:所述3D渲染引擎模块中模拟光线信息通过对所述模拟光线虚拟物体进行阴影区域识别,得到阴影区域;对所述阴影区域进行模糊边缘数据构建,得到模拟边缘数据,同时,通过光线衰减数据计算公式对所述阴影区域进行光线衰减数据计算,得到光线衰减数据,其中,所述光线衰减数据计算公式如:3. According to claim 2, a scene experience generation system based on XR technology combination is characterized in that: the simulated light information in the 3D rendering engine module obtains the shadow area by identifying the shadow area of the simulated light virtual object; fuzzy edge data is constructed for the shadow area to obtain simulated edge data, and at the same time, the light attenuation data of the shadow area is calculated by the light attenuation data calculation formula to obtain the light attenuation data, wherein the light attenuation data calculation formula is as follows:其中,A(P)为光线衰减数据,Isource为光线强度数据,P表示模拟光线虚拟物体中阴影区域的P点;d为P点到光源的距离;σ(x)为所述模拟光线虚拟物体的介质吸收系数;μ为阴影区域的中心位置参数;σshadow为阴影区域面积;k为阴影区域的形状参数;α为散射系数;Wherein, A(P) is the light attenuation data, Isource is the light intensity data, P represents the point P of the shadow area in the virtual object simulating light; d is the distance from the point P to the light source; σ(x) is the medium absorption coefficient of the virtual object simulating light; μ is the center position parameter of the shadow area; σshadow is the area of the shadow area; k is the shape parameter of the shadow area; α is the scattering coefficient;将所述光线接收数据、所述模拟边缘数据以及所述光线衰减数据输入所述蒙特卡洛光线追踪算法进行模拟光线信息生成,得到所述模拟XR场景中模拟光线信息。The light reception data, the simulated edge data, and the light attenuation data are input into the Monte Carlo ray tracing algorithm to generate simulated light information to obtain simulated light information in the simulated XR scene.4.根据权利要求1所述一种基于XR技术组合场景体验生成系统,其特征在于:所述用户交互模块手势识别通过深度感知或图像捕获的设备来捕获用户的手势数据,记录用户执行各种手势时的视频或图像数据,去除图像或视频中的噪声,提高数据质量,从背景中分割出手势区域,以便后续处理。4. According to claim 1, a scene experience generation system based on XR technology combination is characterized in that: the user interaction module gesture recognition captures the user's gesture data through a depth perception or image capture device, records the video or image data when the user performs various gestures, removes noise in the image or video, improves data quality, and segments the gesture area from the background for subsequent processing.5.根据权利要求3所述一种基于XR技术组合场景体验生成系统,其特征在于:所述用户手势数据转换为统一的尺寸和格式,提取手势的轮廓、边界、角点等形状特征,分析手势的运动轨迹、速度、加速度等动态特征,使用学习模型对手势进行分类;使用标注好的手势数据集对模型进行训练,为每个手势定义一个或多个与之关联的指令或操作;当模型识别出用户的手势时,根据手势映射关系将手势转化为相应的指令;将生成的指令发送给XR系统或应用程序,以执行相应的操作。5. According to claim 3, a scene experience generation system based on XR technology combination is characterized in that: the user gesture data is converted into a unified size and format, the shape features of the gesture such as contour, boundary, corner point, etc. are extracted, the dynamic features of the gesture such as motion trajectory, speed, acceleration, etc. are analyzed, and the gesture is classified using a learning model; the model is trained using a labeled gesture data set, and one or more associated instructions or operations are defined for each gesture; when the model recognizes the user's gesture, the gesture is converted into a corresponding instruction according to the gesture mapping relationship; the generated instruction is sent to the XR system or application to perform the corresponding operation.6.根据权利要求1所述一种基于XR技术组合场景体验生成系统,其特征在于:所述建模模块对所述环境深度数据以及空间布局数据并进行数据清洗,通过深度图像超分辨率算法对所述空间布局数据进行清晰度校正,得到校正布局数据;对所述校正布局数据进行边缘平滑处理,得到平滑布局数据;对所述去除深度数据进行缺失数据填充,得到填充深度数据;将所述平滑布局数据以及所述填充深度数据输入点云转换算法进行空间数据转换,得到所述目标三维数据。6. According to claim 1, a scene experience generation system based on XR technology combination is characterized in that: the modeling module performs data cleaning on the environmental depth data and spatial layout data, performs clarity correction on the spatial layout data through a deep image super-resolution algorithm to obtain corrected layout data; performs edge smoothing on the corrected layout data to obtain smoothed layout data; fills missing data on the removed depth data to obtain filled depth data; inputs the smoothed layout data and the filled depth data into a point cloud conversion algorithm for spatial data conversion to obtain the target three-dimensional data.7.根据权利要求1所述一种基于XR技术组合场景体验生成系统,其特征在于:所述用户交互模块还包括双目立体校正模块,所述双目立体校正模块用于提高立体匹配的效率和准确性,消除图像畸变,所述双目立体校正模块矫正计算公式为:7. According to claim 1, a scene experience generation system based on XR technology combination is characterized in that: the user interaction module also includes a binocular stereo correction module, the binocular stereo correction module is used to improve the efficiency and accuracy of stereo matching and eliminate image distortion, and the correction calculation formula of the binocular stereo correction module is:dst(x,y)=src(mapx(x,y),mapy(x,y))dst(x,y)=src(mapx (x,y),mapy (x,y))其中,src为双目相机图像,dst(x,y)为被校正处理后的双目相机校正图像,map_x、map_y为映射矩阵,在dst(x,y)中,x为双目相机校正图像的横坐标,y为双目相机校正图像的纵坐标;在map_x(x,y)或map_y(x,y)中,x为双目相机图像的横坐标,y为双目相机图像的纵坐标。Wherein, src is the binocular camera image, dst(x,y) is the binocular camera correction image after correction processing, map_x and map_y are mapping matrices, in dst(x,y), x is the horizontal coordinate of the binocular camera correction image, y is the vertical coordinate of the binocular camera correction image; in map_x(x,y) or map_y(x,y), x is the horizontal coordinate of the binocular camera image, y is the vertical coordinate of the binocular camera image.8.根据权利要求1所述一种基于XR技术组合场景体验生成系统,其特征在于:所述适配器支持模块还包括陀螺仪单元、眼球追踪单元和显示单元,其中陀螺仪单元用于对用户头部的转动角度进行采集,所述眼球追踪单元用于对用户的眼球进行追踪,确定用户的目光所停留的位置,所述显示单元包括语音识别单元、语音转换单元、频率分析单元、语音区分单元和文字显示单元。8. According to claim 1, a scene experience generation system based on XR technology combination is characterized in that: the adapter support module also includes a gyroscope unit, an eye tracking unit and a display unit, wherein the gyroscope unit is used to collect the rotation angle of the user's head, the eye tracking unit is used to track the user's eyes and determine the position where the user's gaze rests, and the display unit includes a speech recognition unit, a speech conversion unit, a frequency analysis unit, a speech distinction unit and a text display unit.9.根据权利要求1所述一种基于XR技术组合场景体验生成系统,其特征在于:所述语音识别单元用于对用户的语音信号进行识别,所述频率分析单元用于对语音识别单元所识别的语音信号的频率进行分析,所述语音区分单元根据频率分析单元的分析结果判定该语音信号是否属于用户自身,所述语音转换单元用于将属于用户自身的语音信号转换为文字信息,所述文字显示单元用于将语音转换单元所转换的文字信息以字幕的形式显示在多人交互场景下他人的显示单元上。9. According to claim 1, a scene experience generation system based on XR technology combination is characterized in that: the voice recognition unit is used to recognize the user's voice signal, the frequency analysis unit is used to analyze the frequency of the voice signal recognized by the voice recognition unit, the voice distinction unit determines whether the voice signal belongs to the user himself according to the analysis result of the frequency analysis unit, the voice conversion unit is used to convert the voice signal belonging to the user himself into text information, and the text display unit is used to display the text information converted by the voice conversion unit in the form of subtitles on the display unit of others in a multi-person interaction scenario.10.根据权利要求1所述一种基于XR技术组合场景体验生成装置,其特征在于:包括XR头戴设备(1),所述XR头戴设备(1)用于实现用户人机交互,所述XR头戴设备(1)两侧与机架(2)连接,所述XR头戴设备(1)内部与耳机(4)电性连接,所述XR头戴设备(1)正面嵌装有显示器(5),所述XR头戴设备(1)顶部安装有扫描装置(3)。10. According to claim 1, a scene experience generation device based on XR technology combination is characterized in that: it includes an XR head-mounted device (1), the XR head-mounted device (1) is used to realize user human-computer interaction, the two sides of the XR head-mounted device (1) are connected to the frame (2), the inside of the XR head-mounted device (1) is electrically connected to the earphone (4), the front of the XR head-mounted device (1) is embedded with a display (5), and the top of the XR head-mounted device (1) is installed with a scanning device (3).
CN202410974424.1A2024-07-192024-07-19Combined scene experience generation system and device based on XR technologyPendingCN119006686A (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202410974424.1ACN119006686A (en)2024-07-192024-07-19Combined scene experience generation system and device based on XR technology

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202410974424.1ACN119006686A (en)2024-07-192024-07-19Combined scene experience generation system and device based on XR technology

Publications (1)

Publication NumberPublication Date
CN119006686Atrue CN119006686A (en)2024-11-22

Family

ID=93475508

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202410974424.1APendingCN119006686A (en)2024-07-192024-07-19Combined scene experience generation system and device based on XR technology

Country Status (1)

CountryLink
CN (1)CN119006686A (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN104604256A (en)*2012-08-312015-05-06杜比实验室特许公司Reflected sound rendering of object-based audio
CN113109943A (en)*2021-04-062021-07-13深圳市思麦云科技有限公司XR-based simulation multi-person interaction system
CN116912403A (en)*2023-07-032023-10-20上海鱼微阿科技有限公司XR equipment and obstacle information sensing method thereof
CN116977600A (en)*2023-07-032023-10-31上海鱼微阿科技有限公司XR equipment and XR equipment height acquisition method
CN117994477A (en)*2024-04-022024-05-07虚拟现实(深圳)智能科技有限公司Method, device, equipment and storage medium for realizing XR (X-ray) augmented reality scene

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN104604256A (en)*2012-08-312015-05-06杜比实验室特许公司Reflected sound rendering of object-based audio
CN113109943A (en)*2021-04-062021-07-13深圳市思麦云科技有限公司XR-based simulation multi-person interaction system
CN116912403A (en)*2023-07-032023-10-20上海鱼微阿科技有限公司XR equipment and obstacle information sensing method thereof
CN116977600A (en)*2023-07-032023-10-31上海鱼微阿科技有限公司XR equipment and XR equipment height acquisition method
CN117994477A (en)*2024-04-022024-05-07虚拟现实(深圳)智能科技有限公司Method, device, equipment and storage medium for realizing XR (X-ray) augmented reality scene

Similar Documents

PublicationPublication DateTitle
CN106993249B (en) A method and device for processing audio data of a sound field
US20230041730A1 (en)Sound effect adjustment
WO2021043053A1 (en)Animation image driving method based on artificial intelligence, and related device
WO2022148083A1 (en)Simulation 3d digital human interaction method and apparatus, and electronic device and storage medium
CN117391122A (en)3D digital human-assisted chat method established in meta universe
CN105959718A (en)Real-time interaction method and device in video live broadcasting
CN107038745A (en)A kind of 3D tourist sights roaming interaction method and device
CN112562056A (en)Control method, device, medium and equipment for virtual light in virtual studio
CN112749611A (en)Face point cloud model generation method and device, storage medium and electronic equipment
CN204406327U (en)Based on the limb rehabilitating analog simulation training system of said three-dimensional body sense video camera
CN115049016A (en)Model driving method and device based on emotion recognition
CN116524087A (en) Audio-driven speaker video synthesis method and system fusing neural radiation fields
CN104933278B (en)A kind of multi-modal interaction method and system for disfluency rehabilitation training
CN113221840B (en) A kind of portrait video processing method
CN119359870A (en) Multi-pose portrait animation generation method, device, equipment and storage medium
CN119006686A (en)Combined scene experience generation system and device based on XR technology
CN102880288A (en)Three-dimensional (3D) display human-machine interaction method, device and equipment
CN116863043A (en)Face dynamic capture driving method and device, electronic equipment and readable storage medium
CN111435268A (en)Human-computer interaction method based on image recognition and reconstruction and system and device using same
JP7161200B2 (en) karaoke production system
JP2024506299A (en) Scene understanding using occupancy grids
CN113823129A (en)Method and device for guiding disassembly and assembly of turning wheel equipment based on mixed reality
CN116958353B (en)Holographic projection method based on dynamic capture and related device
CN119379873B (en)AIGC-based digital person and holographic display method, system and storage medium
CN117292094B (en)Digitalized application method and system for performance theatre in karst cave

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination

[8]ページ先頭

©2009-2025 Movatter.jp