Create augmented selfies with a real-time “green screen” effect. This sample app demonstrates how to use the Intel® RealSense™ SDK's User Segmentation module to replace the screen background with a static image selected by the user.
Learn how to use the Intel® RealSense™ SDK to scan and map a user’s face onto an existing 3D character model. The face mapping sample uses the 3D Scan module to scan the user's face and then map it onto an existing 3D head model. This technique does a "stone face" mapping that is not rigged or currently capable of animating. The code is written in C++ and uses DirectX*.
Use this code sample to learn how to use a native Unity plug-in to increase performance of displaying Intel® RealSense™ SDK image data by bypassing the C# layers of the SDK. Image data is uploaded to the GPU directly through the graphics API.
The Chat Heads code sample implements an application that uses the background segmentation feature of the Intel® RealSense™ SDK, multiplayer networking, and media encode and decode to create an improved e-sport experience.
This article provides an introduction to autonomous navigation and its use in augmented reality applications, with a focus on agents that move and navigate. Autonomous agents are entities that act independently using artificial intelligence, which defines the operational parameters and rules by which the agent must abide. The agent responds dynamically in real time to its environment, so even a...
This code sample allows the user to scan their face using a front-facing Intel® RealSense™ camera, project it onto a customizable head mesh, and apply post-processing effects on it. It’s an extension of a previous code sample titled Applying Intel® RealSense™ SDK Face Scans to a 3D Mesh by adding features and techniques to improve the quality of the final head result as well as provide a series...