Rate this Page

XPU Support#

PyTorch provides XPU support for Intel GPU-accelerated tensor operations. The XPU API allows you to manage Intel GPU devices, streams for asynchronous execution, and synchronization.

When to use XPU APIs:

  • When running on Intel GPUs (Data Center GPU Max, Arc, etc.)

  • When implementing custom XPU kernels or operations

  • When managing asynchronous execution with XPU streams

  • When writing device-portable code alongside CUDA

Basic usage:

#include <torch/torch.h>

// Check if XPU is available
if (torch::xpu::is_available()) {
    // Create tensor on XPU
    auto tensor = torch::randn({2, 3}, torch::device(torch::kXPU));

    // Move model to XPU
    model->to(torch::kXPU);
}

Header Files#

  • torch/xpu.h - High-level XPU utilities (device count, availability, seeding)

  • c10/xpu/XPUStream.h - XPU stream management

XPU Categories#