Coaching Overlay Feature

This view provides a standardized installation procedure for your users. You can configure this view to automatically display during session initialization and limited tracking, while giving users specific instructions to best facilitate ARKit's world tracking.
Coaching Overlay

Environment Probe

Environmental detection is a technology that captures real-world images from a camera and organizes this information into environmental textures, such as a cube map, which contains views in all directions from a specific point in the scene. Using this environment texture to render 3D objects can reflect real-world images in the rendered objects. The result is usually real reflections and lighting of virtual objects, influenced by perceptions of the real world.
Environment Probe

Light Estimation

Light Estimation is discrete visual cues for a given image and provide detailed information about the lighting in a given scene. You can then use this information when rendering virtual objects to illuminate them under the same conditions as the placed scene, so that these objects feel more realistic and enhance the user's immersive experience.
Light Estimation

Post Processing Feature

Post-processing is the process of applying full-screen filters and effects before the camera's image buffer is displayed on the screen. It can greatly improve the visual effect of your product with very little setup time.
You can use post-processing effects to simulate physical camera and film properties.
Post Processing

AR Interaction Feature

AR Interaction is an advanced, component-based interaction system. It also contains auxiliary components, you can use these components to extend the function of drawing visual effects and hooking your own interactive events.
  • AR gesture system maps screen touches to gesture events
  • AR gesture interactor and interactive conversion gestures, such as position, selection, conversion, rotation and zoom to object operations
  • AR annotations inform users about AR objects placed in the real world

Immersal Setting

Use Server Localizer

The environment where the current user is located is located through the data stored on the cloud server to load the grid map of the current environment.

Turn Off Server Localized

When the map is located through the cloud, stop using the cloud location to prevent excessive API calls from increasing costs.

Localization Interval

Cloud positioning/device positioning interval frequency

Render Mode

Mode of rendering point cloud data。
Render Mode
Do Not Render
Do not render scene point cloud data
Editor Only
Only render in Unity Editor (not in Play mode)
Editor And Runtime
Always render, whether it is Editor-Play or when it is running on the device

Face Mesh

Maximum Face Count

Maximum number of faces that can be tracked
Face Mesh


Occlusion: Allows the virtual content to be occluded by the detected environment depth (environmental occlusion) or the detected human body depth (human occlusion). Some devices provide in-depth information about the real world. For example, there is a feature called person occlusion. iOS devices with A12 bionic chip (newer) can detect the depth information of the human body in the frame of the augmented reality camera. Newer Android phones and iOS devices are equipped with lidar scanners that can provide depth images of the environment, and each pixel contains a depth estimate between the device and the physical environment.

Enable AR Occlusion

Enable AR Occlusion, it will turn on the Occlusion algorithm.
This feature is quite costly in performance and memory, please use it with caution

Environment Depth Mode

Environment depth mode
Environment depth is disabled and will not be generated
Environment depth is enabled and will be generated at the fastest resolution
Ambient depth is enabled and will be generated at medium resolution
Ambient depth is enabled and will be generated at the best resolution

Human Segmentation Depth Mode

Human body segmentation depth algorithm mode
Segmentation depth is disabled and will not be generated
Segmentation depth is enabled and will be generated without additional image filtering
Segmentation depth is enabled and will produce additional image filtering

Human Segmentation Stencil Mode

Human body segmentation template algorithm mode
The split template is disabled and will not be generated
The split template is enabled and will be generated in the fastest resolution
The split template is enabled and will be generated in moderate resolution
The split template is enabled and will be generated with the best resolution

Occlusion Preference Mode

AR blocking algorithm preference mode
Prefer Environment Occlusion
The preferred is the use of the environment depth occlusion
Prefer Human Occlusion
Priority use human split templates and depth to block

ARWorld Scale Block

AR World Scale scales the size of the current AR experience scene according to the size value set by the developer to get a better visual effect and experience.
AR World Scale
The size value of the AR experience scene world at runtime. Unity 1 unit is equal to 1 meter in the real world. Default scale size is 1.

Multiplayer Block

Multiplayer Block is a system for building multiplayer capabilities for AR experiences.
Last modified 1mo ago