Build and Run a Sample Using the Command Line
Intel® AI Analytics Toolkit
In this section, you will run a simple "Hello World" project to familiarize yourself with the process of building projects, and then build your own project.
You can use either a terminal window or Visual Studio Code* when working from the command line. For details on how to use VS Code locally, see Basic Usage of Visual Studio Code with oneAPI on Linux*. To use VS Code remotely, see Remote Visual Studio Code Development with oneAPI on Linux*.
Build and Run a Sample Project
The samples below must be cloned to your system before you can build the sample project:
Name of Sample | Description | How to Clone and Build |
Intel Extension for PyTorch Getting Started, Intel oneCCL Bindings for PyTorch | Train a PyTorch model and run the inference with the Intel® Deep Neural Network Library (Intel® DNNL) enabled. Intel® Extension for PyTorch* extends PyTorch* with optimizations for extra performance boost on Intel hardware. |
Clone oneCCL Bindings for PyTorch or Intel Extension for PyTorch sample, then follow the directions in README.md to build and run the sample. |
TensorFlow Hello World, Intel Extension for TensorFlow Getting Started | TensorFlow optimized on Intel hardware enables Intel® DNNL calls by default. It implements an example neural network with one convolution layer and one ReLU layer. Intel® Extension for TensorFlow* is a heterogeneous, high performance deep learning extension plugin based on TensorFlow PluggableDevice interface. This extension plugin brings Intel XPU (GPU, CPU, etc) devices into the TensorFlow open source community for AI workload acceleration. |
Clone TensorFlow_HelloWorld or Intel Extension for TensorFlow sample, then follow the directions in README.md to build and run the sample. |
Intel® Distribution of Modin* Getting Started | This Getting Started sample code shows how to use distributed Pandas using the Modin package. | To get the Intel® Distribution of Modin*, you must install the AI Kit using the Conda* package manager. After the AI Kit is installed with Conda, clone Intel® Distribution of Modin* Getting Started, then follow the directions in README.md to build and run the sample. |
Model Zoo for Intel® Architecture |
|
Model Zoo for Intel® Architecture can be found in your installation of Intel® oneAPI AI Analytics Toolkit, typically found at /opt/intel/oneapi/modelzoo/latest/models. Instructions for navigating the zoo, using the samples, and running the benchmarks are here: https://github.com/IntelAI/models/blob/v2.4.0/docs/general/tensorflow/AIKit.md#navigate-to-the-model-zoo |
Intel® Neural Compressor | Intel® Neural Compressor is an open-source Python* library designed to help you quickly deploy low-precision inference solutions on popular deep-learning frameworks such as TensorFlow*, PyTorch*, MXNet*, and ONNX* (Open Neural Network Exchange) runtime. | Clone neural-compressor, then follow the directions in README.md to build and run the sample. |
Intel® Extension for Scikit-learn* | Provide a seamless way to speed up your Scikit-learn application using using of the Intel® oneAPI Data Analytics Library (oneDAL). | Clone Intel® Extension for Scikit-learn*, then follow the directions in the README.md to build and run the sample. |
For more samples, browse the full GitHub repository: Intel® oneAPI AI Analytics Toolkit Code Samples. |
To see a list of components that support CMake, see Use CMake to with oneAPI Applications.
Build Your Own Project
No special modifications to your existing Python projects are required to start using them with this toolkit. For new projects, the process closely follows the process used for creating sample Hello World projects. Refer to the Hello World README files for instructions.
Maximizing Performance
You can get documentation to help you maximize performance for either TensorFlow or PyTorch.
Configure Your Environment
If you are developing outside of a container, source the following script to use the Intel® Distribution for Python*:
. <install_dir>/setvars.sh
where <install_dir> is where you installed this toolkit. By default the install directory is:
Root or sudo installations: /opt/intel/oneapi
Local user installations: ~/intel/oneapi
To switch environments, you must first deactivate the active environment.
The following example demonstrates configuring the environment, activating TensorFlow*, and then returning to the Intel Distribution for Python:
. conda activate tensorflow conda deactivate conda activate root