This view provides a standardized installation procedure for your users. You can configure this view to automatically display during session initialization and limited tracking, while giving users specific instructions to best facilitate ARKit's world tracking.
Environmental detection is a technology that captures real-world images from a camera and organizes this information into environmental textures, such as a cube map, which contains views in all directions from a specific point in the scene. Using this environment texture to render 3D objects can reflect real-world images in the rendered objects. The result is usually real reflections and lighting of virtual objects, influenced by perceptions of the real world.
Light Estimation is discrete visual cues for a given image and provide detailed information about the lighting in a given scene. You can then use this information when rendering virtual objects to illuminate them under the same conditions as the placed scene, so that these objects feel more realistic and enhance the user's immersive experience.
Post-processing is the process of applying full-screen filters and effects before the camera's image buffer is displayed on the screen. It can greatly improve the visual effect of your product with very little setup time. You can use post-processing effects to simulate physical camera and film propertie
AR Interaction is an advanced, component-based interaction system. It also contains auxiliary components, you can use these components to extend the function of drawing visual effects and hooking your own interactive events.
- AR gesture system maps screen touches to gesture events
- AR gesture interactor and interactive conversion gestures, such as position, selection, conversion, rotation and zoom to object operations
- AR annotations inform users about AR objects placed in the real world
Create AR experiences that merge digital and physical realities for consumers, industrial use, ad campaigns and more.
Use Server Localizer
The environment where the current user is located is located through the data stored on the cloud server to load the grid map of the current environment.
Turn Off Server Localized
When the map is located through the cloud, stop using the cloud location to prevent excessive API calls from increasing costs.
Cloud positioning/device positioning interval frequency
Mode of rendering point cloud data.
|Do Not Render||Do not render scene point cloud data|
|Editor Only||Only render in Unity Editor (not in Play mode)|
|Editor And Runtime||Always render, whether it is Editor-Play or when it is running on the device|
The face mesh is a 3D model of a face. It works in combination with the face tracker in XRMOD Engine to create a surface that reconstructs someone's expressions.
Maximum Face Count
Maximum number of faces that can be tracked
Occlusion: Allows the virtual content to be occluded by the detected environment depth (environmental occlusion) or the detected human body depth (human occlusion). Some devices provide in-depth information about the real world. For example, there is a feature called person occlusion. iOS devices with A12 bionic chip (newer) can detect the depth information of the human body in the frame of the augmented reality camera. Newer Android phones and iOS devices are equipped with lidar scanners that can provide depth images of the environment, and each pixel contains a depth estimate between the device and the physical environment.
Enable AR Occlusion
Enable AR Occlusion, it will turn on the Occlusion algorithm.
This feature is quite costly in performance and memory, please use it with caution
Environment Depth Mode
Environment segmentation depth algorithm mode
|Disabled||Environment depth is disabled and will not be generated|
|Fastest||Environment depth is enabled and will be generated at the fastest resolution|
|Medium||Ambient depth is enabled and will be generated at medium resolution|
|Best||Ambient depth is enabled and will be generated at the best resolution|
Human Segmentation Depth Mode
Human body segmentation depth algorithm mode
|Disabled||Segmentation depth is disabled and will not be generated|
|Fastest||Segmentation depth is enabled and will be generated without additional image filtering|
|Best||Segmentation depth is enabled and will produce additional image filtering|
Human Segmentation Stencil Mode
Human body segmentation template algorithm mode
|Disabled||The split template is disabled and will not be generated|
|Fastest||The split template is enabled and will be generated in the fastest resolution|
|Medium||The split template is enabled and will be generated in moderate resolution|
|Best||The split template is enabled and will be generated with the best resolution|
Occlusion Preference Mode
AR blocking algorithm preference mode |Mode|Description| |---|---| |Prefer Environment Occlusion|The preferred is the use of the environment depth occlusion| |Prefer Human Occlusion|Priority use human split templates and depth to block|
ARWorld Scale Block
AR World Scale scales the size of the current AR experience scene according to the size value set by the developer to get a better visual effect and experience.
|AR World Scale||Float||The size value of the AR experience scene world at runtime. Unity 1 unit is equal to 1 meter in the real world. Default scale size is 1.|
Multiplayer Block is a system for building multiplayer capabilities for AR experiences.