Tags

Zone

This article provides an introduction to autonomous navigation and its use in augmented reality applications, with a focus on agents that move and navigate. Autonomous agents are entities that act independently using artificial intelligence, which defines the operational parameters and rules by which the agent must abide. The agent responds dynamically in real time to its environment, so even a...

The idea behind this project was to provide a demonstration of parallel processing in gaming with Direct3D 12. It expands upon the results from the paper "A Comparison of the Intel® Core™ i5 Processor and Intel® Core™ i7 Processor with Visualizations in OpenGL* and Oculus* VR" (see References section) and extends the code there to contain a Direct3D 12 renderer. It also re-implements the previous...

This Code Sample show you step-by-step on building a useful tool for VR devs using Unreal Engine that leverages the power of intel CPU’s. Unreal Engine has a powerful virtual reality editor option, but something they did not include is the ability to edit and place sounds while inside VR. It can be troublesome constantly having to restart the editor after adjusting a sound to test what it sounds...

With multi-core processors now common place in PCs, and core counts continually climbing, software developers must adapt. By learning to tackle potential performance bottlenecks and issues with concurrency, engineers can future-proof their code to seamlessly handle additional cores as they are added to consumer systems.

This software example demonstrates how to use multi-core technologies to edit images. There are two parts to this project, a .NET Windows application front end written using C# and Windows Presentation Foundation (WPF) and a C++ DLL which is responsible for the actual manipulation of the image.

Developing experiences in virtual reality (VR) is a fun and exciting journey, and without the proper tools or a helping hand it can become quite daunting. To aid in this process, Intel looks to create a roadmap; a series of signposts and lessons learned in a way that helps bridge the gap into the world of VR. To illustrate many of the tools and practices required for this journey we collectively...

Attach this script as component to a gameobject in Unity. After setting up a Tracked Pose Driver component in the gameobject to the correct device and handedness the MRHand.cs script will provide you with public Input functions to obtain controller button presses to drive your game.

Attach this script as a component to any prefab you might want to lightmap. From the editor window inside Unity you can setup a scene with proper lighting and position your prefabs inside the scene as per your requirements.