This article and sample application show you how to use the Intel® RealSense™ camera (R200) and the Enhanced Photography functionality that is part of the Intel® RealSense™ SDK. The article separates the Intel RealSense SDK functionality from the GUI layer code to make it easier to focus on the R200 Enhanced Photography functionality.
The second part in the Intel® Software Guard Extensions (Intel® SGX) tutorial series is a high-level specification for the application we’ll be developing: a simple password manager that provides the same core functions as a commercial product while following good security practices, which we'll then use as a learning vehicle for designing for Intel SGX.
This first article in the Intel® Software Guard Extensions (Intel® SGX) tutorial series is a brief overview of the technology. Learn how Intel SGX can protect your application's private information, including passwords, account numbers, financial information, encryption keys, and health records.
Announcing a new multi-part tutorial series to help software developers integrate Intel® Software Guard Extensions (Intel® SGX) into their applications. The series will guide you through building an Intel SGX application, beginning at application design and running through development, testing, packaging, and deployment. This in-depth look at enabling Intel SGX in a single application provides developers with a hands-on and holistic view of the technology as it is woven into a real-world application.
This article provides an introduction to autonomous navigation and its use in augmented reality applications, with a focus on agents that move and navigate. Autonomous agents are entities that act independently using artificial intelligence, which defines the operational parameters and rules by which the agent must abide. The agent responds dynamically in real time to its environment, so even a simple design can result in complex behavior. An example is developed that uses the Intel RealSense camera R200 and the Unity* 3D Game Engine.
Virtual reality is rapidly gaining popularity, and may soon become a common way of viewing 3D environments. While stereo rendering has been performed on consumer grade graphics processors for a while now, the new wave of virtual reality display devices have two properties that typical applications have not needed to consider before. Pixels no longer appear on regular grids and the displays subtend a wide field-of-view. In this paper, we evaluate several techniques designed to efficiently render for head-mounted displays with such properties. We show that the amount of rendered pixels can be reduced down to 36% without compromising visual fidelity compared to traditional rendering, by rendering multiple optimized sub-projections.