On-Demand Webinars




Tool Type



Applied Filters

Use Intel® oneAPI DL Framework Developer Toolkit to create fast deep learning frameworks from one node to multiple nodes.

Use the Intel® DPC++ Compatibility Tool to perform a one-time migration that ports both kernels and API calls.

How to get started with your first projects on Intel® DevCloud.

How to take advantage of Intel® accelerators using the Offload Advisor.

A walkthrough for efficient Intel® oneAPI developer tool workflows specific to Intel® FPGAs.

Take advantage of a software model that is flexible, familiar, and portable.

Part 2 of this 3-part series, delivers insights into the latest optimizations for Intel® Optimization for TensorFlow* and PyTorch leveraging the new instructions on Intel® Xeon® Scalable Processors.

Part 3 of this 3-part series shifts to “hands-on”, with presenters demonstrating the steps needed to execute key machine learning end-to-end workflows using the Intel® AI Analytics Toolkit.

The Intel® Distribution of OpenVINO™ toolkit solves the cross-architecture inference execution on GPUs, VPUs, and FPGAs with its improved Model Server.

The Intel® Distribution of OpenVINO™ toolkit was designed specifically to help developers deploy AI-powered solutions across the heterogeneous landscape.

The Intel® oneAPI Math Kernel Library, has been optimized for cross-architecture performance, enabling complex math-processing routines to run on CPUs, GPUs, FPGAs, and other accelerators.

Intel® Open Image Denoise is an open source library of high-performance, high-quality, machine-learning-based denoising filters for images rendered with ray tracing.