Skip To Main Content
Intel logo - Return to the home page
My Tools

Select Your Language

  • Bahasa Indonesia
  • Deutsch
  • English
  • Español
  • Français
  • Português
  • Tiếng Việt
  • ไทย
  • 한국어
  • 日本語
  • 简体中文
  • 繁體中文
Sign In to access restricted content

Using Intel.com Search

You can easily search the entire Intel.com site in several ways.

  • Brand Name: Core i9
  • Document Number: 123456
  • Code Name: Emerald Rapids
  • Special Operators: “Ice Lake”, Ice AND Lake, Ice OR Lake, Ice*

Quick Links

You can also try the quick links below to see results for most popular searches.

  • Product Information
  • Support
  • Drivers & Software

Recent Searches

Sign In to access restricted content

Advanced Search

Only search in

Sign in to access restricted content.

The browser version you are using is not recommended for this site.
Please consider upgrading to the latest version of your browser by clicking one of the following links.

  • Safari
  • Chrome
  • Edge
  • Firefox

Optimize the Latest Deep Learning Workloads Using PyTorch* Optimized by Intel

@IntelDevTools

 

Subscribe Now

Stay in the know on all things CODE. Updates are delivered to your inbox.

Sign Up

Overview

For developers focused on deep learning use cases—predictive modeling, recommendation systems, natural language processing, object detection, and tons more—it is paramount to extract the most workload performance using newer technologies like bfloat16, graph-level optimizations, and custom kernels.

This session focuses on the performance and ease-of-use benefits for deep learning training and inference of large models like deep learning recommendation model (DLRM) using Intel® Extension for PyTorch* and Intel® oneAPI Deep Neural Network Library (oneDNN).

Join senior deep learning engineer, Eikan Wang to learn more about the following topics:

  • Using oneDNN to deliver optimal training and inference workload performance for the PyTorch* framework on Intel hardware
  • oneDNN-based graph optimizations and custom kernel implementations to boost performance of DLRM modules in PyTorch
  • How the extension library for PyTorch can be dynamically loaded as a Python module to offer a more modular design for custom compound operations that are critical to accelerating key deep learning modules, for example, the interaction module from DLRM.

 

Get the Software

  • Get the Intel Extension for PyTorch as part of the AI Frameworks and Tools.
  • Get oneDNN as part of the Intel® oneAPI Base Toolkit. (Want this tool stand-alone only? Get it here.)

 

Other Resources

  • Sign up for an Intel® Tiber™ AI Cloud account—a free development sandbox with access to the latest Intel hardware and oneAPI software.
  • Explore oneAPI, including developer opportunities and benefits
  • Subscribe to Code Together—an interview series that explores the challenges at the forefront of cross-architecture development. Each bi-weekly episode features industry VIPs who are blazing new trails through today’s data-centric world. Available wherever you get your podcasts.
 

Eikan Wang
Senior deep learning engineer, Intel Corporation

Eikan is part of the Graphics and Software group where he is the technical lead on PyTorch framework optimization for Intel architecture. He is also one of the major contributors to low-precision inference solutions on Intel architecture. He has four years of full-stack experience in AI from various AI applications to framework, library, and compiler optimizations. Eikan received his bachelor’s degree in mathematics from Huaiyin Institute of Technology.

Jump to:

You May Also Like
 

   

You May Also Like

Related Articles

Deliver Blazing-Fast Python Data Science and AI Performance on CPUs—with Minimal Code Changes

Use Intel Deep Learning Optimizations in TensorFlow*

Related Video

Accelerate AI Inferencing from Development to Deployment

  • Company Overview
  • Contact Intel
  • Newsroom
  • Investors
  • Careers
  • Corporate Responsibility
  • Inclusion
  • Public Policy
  • © Intel Corporation
  • Terms of Use
  • *Trademarks
  • Cookies
  • Privacy
  • Supply Chain Transparency
  • Site Map
  • Recycling
  • Your Privacy Choices California Consumer Privacy Act (CCPA) Opt-Out Icon
  • Notice at Collection

Intel technologies may require enabled hardware, software or service activation. // No product or component can be absolutely secure. // Your costs and results may vary. // Performance varies by use, configuration, and other factors. Learn more at intel.com/performanceindex. // See our complete legal Notices and Disclaimers. // Intel is committed to respecting human rights and avoiding causing or contributing to adverse impacts on human rights. See Intel’s Global Human Rights Principles. Intel’s products and software are intended only to be used in applications that do not cause or contribute to adverse impacts on human rights.

Intel Footer Logo