Skip to content

post-processing

2 posts with the tag “post-processing”

Release v0.7.5

Hair Segmentation

The most important new feature added in this release is Hair Segmentation. We have added the new type of tracking that allows to implement virtual try-on effects like hair recoloring, patching of hairs in headwear virtual try-on, and many more. Hair Segmentation is GPGPU powered and real-time, it supports smooth visual effects on mobile devices.

The standard set of core tools to enable hair segmentation in an app:

Already available plugins implementing processings of segmentation masks are compatible with hair segmentation mask by default. Developers can use familiar building blocks to assemble pipelines of visually rich AR effects. ShaderPlugin can be used as a starting point of a custom visual effect on top of hair segmentation mask.

Mask Post-Processing

MaskBinaryPlugin applying binarization operation to a segmentation mask. This plugin can be used within mask processing pipeline to filter pixels based on their probability and separate foreground and background pixels by a threshold.

MaskStepPlugin is a similar but more advanced version of the plugin applying smoothstep operation between two thresholds. Compared to more agressive Heaviside step function used in simple binarization, smoothstep provides smooth transition of mask value between foreground and background regions still separating them. We recommend using MaskStepPlugin for smoother visuals.

MaskFilterPlugin performs temporal filtering of a segmentation mask that is based on minimization of entropy in probability space. It can be integrated into a mask post-processing pipeline to decrease temporal noise.

One notable new plugin designed specifically for hair segmentation mask is MaskSharpPlugin. When used in post-processing pipeline of a segmentation mask it significantly improves its accuracy. For example, in case of hair segmentation it can even highlight separate locks of hairs that were missed by a neural network. We recommend to apply a slight spacial smoothing (about 0.5) of a mask before MaskSharpPlugin. If not provided by the Processor temporal smoothing of a segmentation mask can also improve final results. While improvement of mask quality is significant MaskSharpPlugin is really fast and can be used in mobile browsers without sacrificing frames per second.

Body Patching

In this release we reimplemented body patching processing and corresponding plugins. Partial body patching is now not limited in radius of search for the closes body part defined by “patch” and “keep” meshes. This means that no matter how long is the distance between a pixel of a segmentation mask and the closest mesh it’ll be correctly classified as “patch” or “keep” pixel. Moreover this stage of processing is now blazingly fast and processing time doesn’t depend on a search radius. This makes usage of body patching more convenient across all virtual try-on types. Additionally, the same algorithm is applied within patching procedure itself improving it’s performance, quality, and accuracy. The same improvements and changes are applied within full body patching.

On top of that, to improve quality of image processing and overcome floating point precision limits in shaders on mobile devices we implement multi-stage refinement of results of intermediate processing stages. Additionally, memory access pattern is optimized for all GPGPU kernels across Engeenee SDK.

Wrist Tracking

Hand pose detection was optimized to provide more precise and smooth tracking. We’ve implemented totally new approach for wrist detection that provides much more robust and stable results and overcomes major disadvantages of the previous algorithm. New wrist estimation borrows many principles from phalanx refinement procedure introduced in the previous release and bringing promising results. Wrist is approximated by an elliptical cylindroid with subpixel accuracy. New detector has much better performance, it’s at least twice faster, thus virtual try-on of watches and bracelets has gained significant boost in FPS. On top of that, GPGPU kernels of wrist detector were significantly optimized for real-time usage on mobile devices. Wrist detection can be enabled setting corresponding parameter. WristTrackPlugin may be utilized to align scene node with a wrist estimation.

HandAlignPlugin has been considerably reworked. Now we utilize minimization of back-projection error when fitting hand armature in detected keypoints and iteratively converge skeleton in approximated pose to keypoints at the same time preserving relative lengths and scales. We utilized similar approach when aligning a body armature. HandAlignPlugin still requires some work to be 100% ready for virtual try-on of gloves, we are planning to complete implementation within the next release cycle.

Pose Alignment

This release introduces number of improvements in alignment of armature and detected 3D pose including more natural kinematics, both forward and inverse. We improved fitting of a skeleton into detected points relaxing some hard requirements that when enforced lead to less natural deformations of a mesh. Kinematics of shoulder bones has been reworked to provide less deformations within shoulders and underarms areas. We also fixed rescaling issue near shoulders using the same approach we implemented for limbs in the previous version of the SDK.

Spine curve parameter of ClothAlignPlugin has been fixed and when not specified or set to null | undefined default 1.0 value is used. At the same time 0.0 value is now treated properly and disables curvature recovery. Behaviour of spine bones is also improved to reduce unnatural rescaling.

Image Texture

ImageTexture supports generation of mipmap levels via new convenient API. Mipmaps can considerably improve performance of certain image processing algorithms or visual effect shaders. Number of mipmap levels required to reach 1x1 resolution is evaluated internally. Mipmaps are generated automatically on texture update or upload. ShaderProgram provides an option to opt-in for output texture with auto-generated mipmaps that can be useful when chaining image processing stages.

New read() method of a ShaderProgram copies GPU texture data to the host memory providing access to raw pixel values. Helper TextureReader class provides functionality to get texture data on the client side and use it within image processing pipelines. The important utility provided by TextureReader is non-blocking asynchronous read of texture data using readAsync(). In scenarios where data is not needed immediately or can be accessed with a delay, usage of readAsync() can considerably improve performance as blocking reads stall the whole GPU pipeline and wait until all operations having the texture as a dependency are finished, this in turn significantly decreases performance and utilization of a GPU device. Implementation of readAsync() is thread-safe and correctly handles WebGL state recovery.

Generation of mipmaps and asynchronous read of textures allowed to reimplement BrightnessPlugin and derived LightsPlugin and make them very fast. Now they can be used in rendering pipelines without sacrificing frames per second. Additionally BrightnessPlugin was fine-tuned to react on illumination changes mush faster.

API Improvements

We’ve improved API of video capture setup and made it more flexible. Previously, we had two sets of capture options available. One option is to provide standard MediaStreamConstraints defining set of device capabilities and allowing complete control over selection of a video device. The second option is our simplified VideoParams that provide basic selection of a video device. The main downside was that advanced features like rotation or cropping of the input video streams were available only through VideoParams. To effectively resolve this limitation we’ve merged VideoParams and MediaStreamConstraints options. Now fine-grained specification of the device is available via opts field of VideoParams. This allows for example to request specific deviceId or even focus or white balance modes. Provided MediaStreamConstraints have higher priority than the rest or parameters of VideoParams.

SDK users can provide custom options to underlying babylon.js or three.js renderer when initializing BabylonUniRenderer or ThreeRenderer. This might be useful when more fine-grained control over rendering pipeline or canvas settings are required.

Other Changes

  • Documentation page of Engeenee SDK is now generated by Starlight/Astro.
  • Several fixes where made in saving state of rendering pipeline when WebGL context is shared between several clients.
  • User can provide custom options to underlying babylon.js or three.js renderer when constructing BabylonUniRenderer or ThreeRenderer.
  • Fine-tuning of temporal filter for 2D and 3D pose keypoints.
  • Optimization of memory access in GPGPU kernels with performance improvement.

Release v0.7.0

Segmentation Masks

Segmentation masks associated with tracked objects (face, head, body) now have the same BodyMask type and added to an object as mask field when segmentation is enabled for a processor/engine. Unification of mask type within tracked objects structure allows plugins that use segmentation mask to be compatible with any types of tracking. Most of mask plugins are now compatible with the SegmentationResult union type, so they can be used within face, body and hand based virtual try-ons.

Pipeline of segmentation masks processing was completely restructured to provide the best performance and more flexibility when using masks in rendering or video processing. All plugins processing or utilizing a mask now rely on the fact that it should be pre-uploaded to a texture. MaskUploadPlugin uploads segmentation mask to an image texture that can be reused by plugins next in the rendering pipeline. Uploaded mask is added to the corresponding tracked Pose, Face, or Mask as maskTex field of type BodyMaskTexture.

Next plugins update or use the current mask texture associated with a tracked object. This allows to apply chains of plugins post-processing the segmentation mask. For example, before doing background blur we can apply erosion operation on the mask to remove a couple of pixels on foreground boarders and smooth operator to remove sharp edges. This is a quite powerful approach allowing to combine atomic operations on a segmentation mask and fine-tune it for a particular use case or type of virtual try-on. We introduce a set of plugins that process a mask.

Mask Post-Processing

We introduce a set of plugins for mask post-processing:

  • MaskSmoothPlugin performs blurring of a segmentation mask to smooth sharp edges or propagate masked region outwards and inwards.
  • MaskDilationPlugin performs the dilation operation on a segmentation mask adding pixels to the boundaries of masked objects.
  • MaskErosionPlugin performs the erosion operation on a segmentation mask removing pixels on the boundaries of masked objects.
  • MaskMorphPlugin performs a morphological operation on a segmentation mask combining dilation and erosion in one plugin.

Following the principle of plugin chaining, now plugins update and have access to the same segmentation mask associated with a tracked object. This makes high resolution versions of plugins obsolete. BodyPatchHDPlugin, BodypartPatchHDPlugin, and OccluderMaskHDPlugin are removed from the Engeenee SDK as redundant. BodyPatchPlugin, BodypartPatchPlugin, and OccluderMaskPlugin become compatible with segmentation masks of any resolution uploaded to a texture.

While mask plugins automatically reconfigure themselves to the resolution of a segmentation mask their constructors have mask size parameter. It is recommended, though not required, to set it to the size that mask will have at the stage where the plugin will be attached to a renderer. This will allow to avoid unnecessary reconfiguration on runtime after renderer is initialized.

More information can be found here.

Mask Upscaler

Mask upscaling was updated for the new pipeline and now works as other mask plugins updating mask texture embedded into tracked objects by mask uploader. BodyMaskHDPlugin was renamed to MaskUpscalePlugin.

Performance and accuracy of MaskUpscalePlugin was significantly improved. Additionally, now it’s possible to choose number of x2 upscale levels and thus resolution of output mask that can help to fine-tune between performance and accuracy.

Temporal Mask Filter

Temporal filtering of segmentation masks provides for more stable results with less noise and smoother changes of a mask over time. Yet temporal filter doesn’t introduce any delay and reacts to changes immediately. The main goal of temporal filtering is to resolve ambiguity in regions that are hard to robustly classify as foreground. It’s available for all processors that output segmentation masks, namely PoseProcessor, FaceProcessor, and MaskProcessor. We’ve developed blazingly fast implementation that has minimal impact on performance and at the same time significantly increases accuracy. Temporal mask filtering can be enabled by providing { smooth: true } as mask parameter of processor settings when calling Engine’s init().

Resizable Segmentation

In this release we introduce resizable neural networks for body segmentation. Resolution of evaluated segmentation masks can be adjusted, for a particular use case one can choose between speed and accuracy. A lower resolution mask can be evaluated much faster, while higher resolution provides higher density of mask pixels for a region of interest. Smart upscaling of segmentation masks allows to achieve good segmentation results even when mask size is reduced.

Another important improvement is that trackers adjust size of segmentation mask according to aspect ratio of input video. When mask is evaluated for the whole image, size of required padding is significantly reduced. This improves both, accuracy of segmentation and density of mask pixels ratio to image pixels.

For face tracking it’s possible to choose the size of the segmentation mask by setting maskSize parameter. There are two modes of selecting region of segmentation in face tracker, first is to do segmentation for the whole image that is useful for selfie-like use cases, second - for an extended bounding box of the detected face, this option may work better when face tracking is used in full-body experiences on AR mirrors. One may choose between modes setting maskExt processor parameter.

Mask Plugins

  • BgBlurPlugin blurs background region of an image.
  • BodyPatchPlugin patches (inpaints / erases) foreground region of an image.
  • BodypartPatchPlugin conditionally patches (inpaints / erases) foreground regions of an image that are closer to virtual added objects than to the body.

All plugins that classify pixels as foreground or background are parametrized by foreground threshold.

Other Changes

  • Performance of BodypartPatchPlugin has been significantly improved.
  • Adjustable radius of patching in BodyPatchPlugin and BodypartPatchPlugin.
  • ShaderProgram supports changing uniforms on runtime using setUniform(). Previously it might be done only in constructor of process().
  • Support of vector uniforms.
  • ShaderProgram may output a grayscale image.
  • Rename BlurMaskPlugin -> BgBlurPlugin.
  • Optimization of WASM modules initialization.