So Intel was at Augmented World Expo 2015. Why? And what did we do there and what did we share? Read this blog to find out!
If you’re looking for that next new experience to put into your games, you should consider Intel® RealSense™ Technology.
This software example demonstrates how to use multi-core technologies to edit images. There are two parts to this project, a .NET Windows application front end written using C# and Windows Presentation Foundation (WPF) and a C++ DLL which is responsible for the actual manipulation of the image.
Attach this script as component to a gameobject in Unity. After setting up a Tracked Pose Driver component in the gameobject to the correct device and handedness the MRHand.cs script will provide you with public Input functions to obtain controller button presses to drive your game.
Attach this script as a component to any prefab you might want to lightmap. From the editor window inside Unity you can setup a scene with proper lighting and position your prefabs inside the scene as per your requirements.
Code Sample: Intel® Multi-Core Processor Enterprise VR Experience, Integrating Crowd Simulations into Mixed RealityCode Sample: Intel® Multi-Core Processor Enterprise VR Experience, Integrating Crowd Simulations into Mixed Reality