Skip To Main Content
Intel logo - Return to the home page
My Tools

Select Your Language

  • Bahasa Indonesia
  • Deutsch
  • English
  • Español
  • Français
  • Português
  • Tiếng Việt
  • ไทย
  • 한국어
  • 日本語
  • 简体中文
  • 繁體中文
Sign In to access restricted content

Using Intel.com Search

You can easily search the entire Intel.com site in several ways.

  • Brand Name: Core i9
  • Document Number: 123456
  • Code Name: Emerald Rapids
  • Special Operators: “Ice Lake”, Ice AND Lake, Ice OR Lake, Ice*

Quick Links

You can also try the quick links below to see results for most popular searches.

  • Product Information
  • Support
  • Drivers & Software

Recent Searches

Sign In to access restricted content

Advanced Search

Only search in

Sign in to access restricted content.

The browser version you are using is not recommended for this site.
Please consider upgrading to the latest version of your browser by clicking one of the following links.

  • Safari
  • Chrome
  • Edge
  • Firefox

PyTorch* Optimizations from Intel

Speed Up AI from Research to Production Deployment

  • Features
  • Documentation & Code Samples
  • Training
  • Specifications
  • Help

  

Maximize PyTorch Performance on Intel® Hardware

PyTorch* is an AI and machine learning framework popular for both research and production usage. This open source library is often used for deep learning applications whose compute-intensive training and inference test the limits of available hardware resources.

Intel releases its newest optimizations and features in Intel® Extension for PyTorch* before upstreaming them into open source PyTorch.

With a few lines of code, you can use Intel Extension for PyTorch to:

  • Take advantage of the most up-to-date Intel software and hardware optimizations for PyTorch.
  • Automatically mix different precision data types to reduce the model size and computational workload for inference.
  • Add your own performance customizations using APIs.

Intel also works closely with the open source PyTorch project to optimize the PyTorch framework for Intel hardware. These optimizations for PyTorch, along with the extension, are part of the end-to-end suite of Intel® AI and machine learning development tools and resources.

Download the Stand-Alone Versions

Stand-alone versions of PyTorch and Intel Extension for PyTorch are available. You can install them using a package manager or build from the source.

 

PyTorch | Intel Extension for PyTorch

Help Intel® Extension for PyTorch* Evolve

This open source component has an active developer community. We welcome you to participate.

Open Source Version (GitHub*)


Download the AI Tools

PyTorch and Intel Extension for PyTorch are available in the AI Tools Selector, which provides accelerated machine learning and data analytics pipelines with optimized deep learning frameworks and high-performing Python* libraries.

Get the Tools Now


Features

PyTorch and Intel Extension for PyTorch optimizations and features

Open Source PyTorch Powered by Optimizations from Intel

  • Get the best PyTorch training and inference performance on Intel CPU or GPU hardware through open source contributions from Intel.
  • PyTorch 2.5 introduces prototype support for Intel GPUs.
  • Take advantage of Intel® Deep Learning Boost, Intel® Advanced Vector Extensions (Intel® AVX-512), and Intel® Advanced Matrix Extensions (Intel® AMX) instruction set features to parallelize and accelerate PyTorch workloads.
  • Perform distributed training with oneAPI Collective Communications Library (oneCCL) bindings for PyTorch.

Intel Extension for PyTorch Optimizations and Features

  • Apply the newest performance optimizations not yet in PyTorch with minimal code changes.
  • Run PyTorch on Intel CPUs or GPUs.
  • Automatically mix operator datatype precision between float32 and bfloat16 to reduce computational workload and model size.
  • Control aspects of the thread runtime such as multistream inference and asynchronous task spawning.

Optimized Deployment with OpenVINO™ Toolkit

  • Import your PyTorch model into OpenVINO Runtime to compress model size and increase inference speed.
  • Instantly target Intel CPUs, GPUs (integrated or discrete), or NPUs.
  • Deploy with OpenVINO model server for optimized inference in microservice applications, container-based, or cloud environments. Scale using the same architecture API as KServe for inference execution and inference service provided in gRPC* or REST.

Access the latest AI benchmark performance data for PyTorch and OpenVINO toolkit when running on data center products from Intel. Performance Data.

Documentation & Code Samples

Documentation

  • PyTorch Documentation
  • PyTorch Prerequisites for Intel GPUs
  • Get Started with PyTorch on Intel GPUs
  • PyTorch Performance Tuning Guide
  • Intel Extension for PyTorch
    • Documentation & Tutorials
    • Installation Guide (All Operating Systems)
    • Performance Tuning Guide
    • Release Notes
    • System Requirements
  • TorchServe with Intel Extension for PyTorch
  • oneCCL Bindings for PyTorch

 

View All Documentation

Intel Extension for PyTorch Code Samples

  • Single-Instance Training
  • bfloat16 Inference—Imperative Mode
  • bfloat16 Inference—TorchScript Mode
  • int8 Deployment—Graph Mode
  • C++ Dynamic Library
  • GPU Single-Instance Training
  • GPU Inference
 

More Samples

Training & Tutorials

Accelerate PyTorch Training and Inference Performance Using Intel AMX

Llama 2 Inference with PyTorch on Intel® Arc™ A-series GPUs

Build an Interactive Chat-Generation Model Using DialoGPT and PyTorch

Stable Diffusion* with Intel Arc GPUs

Accelerated Image Segmentation Using PyTorch

Get Faster PyTorch Programs with TorchDynamo

How to Improve TorchServe Inference Performance with Intel Extension for PyTorch

Demonstrations

Optimize Text and Image Generation Using PyTorch

Learn how to speed up generative AI that runs on CPUs by setting key environment variables, by using ipex.llm.optimize() for a Llama 2 model and ipex.optimize() for a Stable Diffusion model.

Read

 

Build an End-to-End Language Identification with PyTorch

Follow along with a code sample that performs language identification from audio samples using the Hugging Face SpeechBrain* toolkit. Learn how to optimize the model for inference on CPU or GPU using Intel Extension for PyTorch.

Read

Predict Forest Fires Using Transfer Learning on a CPU

This application classifies aerial photos according to the fire danger they convey. It uses the MODIS fire dataset to adapt a pretrained ResNet-18 model.

Read

 

Use PyTorch for Monocular Depth Estimation

Learn how to use a model based on Hugging Face Transformers to produce a clipped image with background clutter removed, ultimately creating a depth estimate from a single image.

Watch

Case Studies

Intel and Microsoft Azure* Accelerate PadChest and fMRI Models

Using Intel Extension for PyTorch with the OpenVINO toolkit, this project optimized for deployment to Intel CPUs a chest X-ray image classification dataset and a brain functional magnetic resonance imaging (fMRI) resting-state classification model.

Learn More

 

L&T Technology Services Enhances Chest Radiology Outcomes

Chest-rAI* is a deep learning algorithm developed by L&T Technology Services (LTTS) to detect and isolate abnormalities in chest X-ray imagery. LTTS adopted the AI Tools and OpenVINO toolkit, reducing inference time by 46% and reducing their product development time from eight weeks to two weeks.

Learn More

Hugging Face* Accelerates Stable Diffusion* on CPUs

Automatic mixed precision in Intel Extension for PyTorch helped this application take advantage of Intel AMX to speed up inference latency by 2x.

Learn More

 

HippoScreen Improves AI Performance by 2.4x

The Taiwan-based neurotechnology startup used tools and frameworks in the Intel® oneAPI Base Toolkit and AI Tools to improve the efficiency and training times of deep learning models used in its Brain Waves AI system.

Learn More

News

Intel Boosts AI Development with Contributions to PyTorch 2.5

New features enhance the programming experience for AI developers across data center and client hardware, expanding support for Intel GPUs.

Learn More

Intel Joins the PyTorch Foundation

Intel is now a Premier member of the PyTorch Foundation, with four full-time PyTorch maintainers for CPU performance, and a seat on the PyTorch Foundation Governing Board.

Learn More

Intel and Microsoft Collaborate to Extend DirectML GPU Support

Intel has extended support for PyTorch with the DirectML back end to all Intel® Arc™ Graphics and Intel® Iris® Xe Graphics GPUs.

Learn More

 

 

Specifications

Processors:

  • Intel® Xeon® processor
  • Intel® Core™ processor
  • Intel Core Ultra Processors with Intel® Arc™ graphics
  • Intel Arc GPUs
  • Intel® Data Center GPU Max Series

Operating systems:

  • Linux* (Intel Extension for PyTorch is for Linux only)
  • Windows*

Languages:

  • Python
  • C++


Deploy PyTorch models to a variety of devices and operating systems with Intel Distribution of OpenVINO Toolkit.

 

Get Help

Your success is our success. Access these support resources when you need assistance.

  • AI Tools Support Forum
  • Intel® Optimized AI Frameworks Support Forum
  • Intel Extension for PyTorch: GitHub Issue Tickets

Stay Up to Date on AI Workload Optimizations

Sign up to receive hand-curated technical articles, tutorials, developer tools, training opportunities, and more to help you accelerate and optimize your end-to-end AI and data science workflows. Take a chance and subscribe. You can change your mind at any time.

All fields are required unless marked optional.

Intel strives to provide you with a great, personalized experience, and your data helps us to accomplish this.
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
By submitting this form, you are confirming you are age 18 years or older. Intel will process your Personal Data for the purpose of this business request. To learn more about Intel's practices, including how to manage your preferences and settings, visit Intel's Privacy Notice.
By submitting this form, you are confirming you are age 18 years or older. Intel may contact you for marketing-related communications. You can opt out at any time. To learn more about Intel's practices, including how to manage your preferences and settings, visit Intel's Privacy Notice.

You’re In!

Thank you for signing up. Watch for a welcome email to get you started.

  • Company Overview
  • Contact Intel
  • Newsroom
  • Investors
  • Careers
  • Corporate Responsibility
  • Inclusion
  • Public Policy
  • © Intel Corporation
  • Terms of Use
  • *Trademarks
  • Cookies
  • Privacy
  • Supply Chain Transparency
  • Site Map
  • Recycling
  • Your Privacy Choices California Consumer Privacy Act (CCPA) Opt-Out Icon
  • Notice at Collection

Intel technologies may require enabled hardware, software or service activation. // No product or component can be absolutely secure. // Your costs and results may vary. // Performance varies by use, configuration, and other factors. Learn more at intel.com/performanceindex. // See our complete legal Notices and Disclaimers. // Intel is committed to respecting human rights and avoiding causing or contributing to adverse impacts on human rights. See Intel’s Global Human Rights Principles. Intel’s products and software are intended only to be used in applications that do not cause or contribute to adverse impacts on human rights.

Intel Footer Logo