oneAPI | |
Operating System: | Cross-platform |
Platform: | Cross-platform |
Genre: | Open-source software specification for parallel programming |
oneAPI is an open standard, adopted by Intel, for a unified application programming interface (API) intended to be used across different computing accelerator (coprocessor) architectures, including GPUs, AI accelerators and field-programmable gate arrays. It is intended to eliminate the need for developers to maintain separate code bases, multiple programming languages, tools, and workflows for each architecture.[1] [2] [3] [4]
oneAPI competes with other GPU computing stacks: CUDA by Nvidia and ROCm by AMD.
The oneAPI specification extends existing developer programming models to enable multiple hardware architectures through a data-parallel language, a set of library APIs, and a low-level hardware interface to support cross-architecture programming. It builds upon industry standards and provides an open, cross-platform developer stack.[5] [6]
DPC++[7] [8] is a programming language implementation of oneAPI, built upon the ISO C++ and Khronos Group SYCL standards.[9] DPC++ is an implementation of SYCL with extensions that are proposed for inclusion in future revisions of the SYCL standard, including: unified shared memory, group algorithms, and sub-groups.[10] [11] [12]
The set of APIs spans several domains, including libraries for linear algebra, deep learning, machine learning, video processing, and others.
Library Name | ShortName | Description | |
---|---|---|---|
oneAPI DPC++ Library | oneDPL | Algorithms and functions to speed DPC++ kernel programming | |
oneAPI Math Kernel Library | oneMKL | Math routines including matrix algebra, FFT, and vector math | |
oneAPI Data Analytics Library | oneDAL | Machine learning and data analytics functions | |
oneAPI Deep Neural Network Library | oneDNN | Neural networks functions for deep learning training and inference | |
oneAPI Collective Communications Library | oneCCL | Communication patterns for distributed deep learning | |
oneAPI Threading Building Blocks | oneTBB | Threading and memory management template library | |
oneAPI Video Processing Library | oneVPL | Real-time video encode, decode, transcode, and processing |
The source code of parts of the above libraries is available on GitHub.[13]
The oneAPI documentation also lists the "Level Zero" API defining the low-level direct-to-metal interfaces and a set of ray tracing components with its own APIs.
oneAPI Level Zero,[14] [15] [16] the low-level hardware interface, defines a set of capabilities and services that a hardware accelerator needs to interface with compiler runtimes and other developer tools.
Intel has released oneAPI production toolkits that implement the specification and add CUDA code migration, analysis, and debug tools.[17] [18] [19] These include the Intel oneAPI DPC++/C++ Compiler,[20] Intel Fortran Compiler, Intel VTune Profiler[21] and multiple performance libraries.
Codeplay has released an open-source layer[22] [23] [24] to allow oneAPI and SYCL/DPC++ to run atop Nvidia GPUs via CUDA.
University of Heidelberg has developed a SYCL/DPC++ implementation for both AMD and Nvidia GPUs.[25]
Huawei released a DPC++ compiler for their Ascend AI Chipset
Fujitsu has created an open-source ARM version of the oneAPI Deep Neural Network Library (oneDNN)[26] for their Fugaku CPU.
Unified Acceleration Foundation (UXL) is a new technology consortium that are working on the contiuation of the OneAPI initiative, with the goal to create a new open standard accelerator software ecosystem, related open standards and specification projects through Working Groups and Special Interest Groups (SIGs). The goal will compete with Nvidia's CUDA. The main companies behind it are Intel, Google, ARM, Qualcomm, Samsung, Imagination, and VMware.[27]