Skip to content

Release v0.7.5

Hair Segmentation

The most important new feature added in this release is Hair Segmentation. We have added the new type of tracking that allows to implement virtual try-on effects like hair recoloring, patching of hairs in headwear virtual try-on, and many more. Hair Segmentation is GPGPU powered and real-time, it supports smooth visual effects on mobile devices.

The standard set of core tools to enable hair segmentation in an app:

Already available plugins implementing processings of segmentation masks are compatible with hair segmentation mask by default. Developers can use familiar building blocks to assemble pipelines of visually rich AR effects. ShaderPlugin can be used as a starting point of a custom visual effect on top of hair segmentation mask.

Mask Post-Processing

MaskBinaryPlugin applying binarization operation to a segmentation mask. This plugin can be used within mask processing pipeline to filter pixels based on their probability and separate foreground and background pixels by a threshold.

MaskStepPlugin is a similar but more advanced version of the plugin applying smoothstep operation between two thresholds. Compared to more agressive Heaviside step function used in simple binarization, smoothstep provides smooth transition of mask value between foreground and background regions still separating them. We recommend using MaskStepPlugin for smoother visuals.

MaskFilterPlugin performs temporal filtering of a segmentation mask that is based on minimization of entropy in probability space. It can be integrated into a mask post-processing pipeline to decrease temporal noise.

One notable new plugin designed specifically for hair segmentation mask is MaskSharpPlugin. When used in post-processing pipeline of a segmentation mask it significantly improves its accuracy. For example, in case of hair segmentation it can even highlight separate locks of hairs that were missed by a neural network. We recommend to apply a slight spacial smoothing (about 0.5) of a mask before MaskSharpPlugin. If not provided by the Processor temporal smoothing of a segmentation mask can also improve final results. While improvement of mask quality is significant MaskSharpPlugin is really fast and can be used in mobile browsers without sacrificing frames per second.

Body Patching

In this release we reimplemented body patching processing and corresponding plugins. Partial body patching is now not limited in radius of search for the closes body part defined by “patch” and “keep” meshes. This means that no matter how long is the distance between a pixel of a segmentation mask and the closest mesh it’ll be correctly classified as “patch” or “keep” pixel. Moreover this stage of processing is now blazingly fast and processing time doesn’t depend on a search radius. This makes usage of body patching more convenient across all virtual try-on types. Additionally, the same algorithm is applied within patching procedure itself improving it’s performance, quality, and accuracy. The same improvements and changes are applied within full body patching.

On top of that, to improve quality of image processing and overcome floating point precision limits in shaders on mobile devices we implement multi-stage refinement of results of intermediate processing stages. Additionally, memory access pattern is optimized for all GPGPU kernels across Engeenee SDK.

Wrist Tracking

Hand pose detection was optimized to provide more precise and smooth tracking. We’ve implemented totally new approach for wrist detection that provides much more robust and stable results and overcomes major disadvantages of the previous algorithm. New wrist estimation borrows many principles from phalanx refinement procedure introduced in the previous release and bringing promising results. Wrist is approximated by an elliptical cylindroid with subpixel accuracy. New detector has much better performance, it’s at least twice faster, thus virtual try-on of watches and bracelets has gained significant boost in FPS. On top of that, GPGPU kernels of wrist detector were significantly optimized for real-time usage on mobile devices. Wrist detection can be enabled setting corresponding parameter. WristTrackPlugin may be utilized to align scene node with a wrist estimation.

HandAlignPlugin has been considerably reworked. Now we utilize minimization of back-projection error when fitting hand armature in detected keypoints and iteratively converge skeleton in approximated pose to keypoints at the same time preserving relative lengths and scales. We utilized similar approach when aligning a body armature. HandAlignPlugin still requires some work to be 100% ready for virtual try-on of gloves, we are planning to complete implementation within the next release cycle.

Pose Alignment

This release introduces number of improvements in alignment of armature and detected 3D pose including more natural kinematics, both forward and inverse. We improved fitting of a skeleton into detected points relaxing some hard requirements that when enforced lead to less natural deformations of a mesh. Kinematics of shoulder bones has been reworked to provide less deformations within shoulders and underarms areas. We also fixed rescaling issue near shoulders using the same approach we implemented for limbs in the previous version of the SDK.

Spine curve parameter of ClothAlignPlugin has been fixed and when not specified or set to null | undefined default 1.0 value is used. At the same time 0.0 value is now treated properly and disables curvature recovery. Behaviour of spine bones is also improved to reduce unnatural rescaling.

Image Texture

ImageTexture supports generation of mipmap levels via new convenient API. Mipmaps can considerably improve performance of certain image processing algorithms or visual effect shaders. Number of mipmap levels required to reach 1x1 resolution is evaluated internally. Mipmaps are generated automatically on texture update or upload. ShaderProgram provides an option to opt-in for output texture with auto-generated mipmaps that can be useful when chaining image processing stages.

New read() method of a ShaderProgram copies GPU texture data to the host memory providing access to raw pixel values. Helper TextureReader class provides functionality to get texture data on the client side and use it within image processing pipelines. The important utility provided by TextureReader is non-blocking asynchronous read of texture data using readAsync(). In scenarios where data is not needed immediately or can be accessed with a delay, usage of readAsync() can considerably improve performance as blocking reads stall the whole GPU pipeline and wait until all operations having the texture as a dependency are finished, this in turn significantly decreases performance and utilization of a GPU device. Implementation of readAsync() is thread-safe and correctly handles WebGL state recovery.

Generation of mipmaps and asynchronous read of textures allowed to reimplement BrightnessPlugin and derived LightsPlugin and make them very fast. Now they can be used in rendering pipelines without sacrificing frames per second. Additionally BrightnessPlugin was fine-tuned to react on illumination changes mush faster.

API Improvements

We’ve improved API of video capture setup and made it more flexible. Previously, we had two sets of capture options available. One option is to provide standard MediaStreamConstraints defining set of device capabilities and allowing complete control over selection of a video device. The second option is our simplified VideoParams that provide basic selection of a video device. The main downside was that advanced features like rotation or cropping of the input video streams were available only through VideoParams. To effectively resolve this limitation we’ve merged VideoParams and MediaStreamConstraints options. Now fine-grained specification of the device is available via opts field of VideoParams. This allows for example to request specific deviceId or even focus or white balance modes. Provided MediaStreamConstraints have higher priority than the rest or parameters of VideoParams.

SDK users can provide custom options to underlying babylon.js or three.js renderer when initializing BabylonUniRenderer or ThreeRenderer. This might be useful when more fine-grained control over rendering pipeline or canvas settings are required.

Other Changes

  • Documentation page of Engeenee SDK is now generated by Starlight/Astro.
  • Several fixes where made in saving state of rendering pipeline when WebGL context is shared between several clients.
  • User can provide custom options to underlying babylon.js or three.js renderer when constructing BabylonUniRenderer or ThreeRenderer.
  • Fine-tuning of temporal filter for 2D and 3D pose keypoints.
  • Optimization of memory access in GPGPU kernels with performance improvement.