跳转至主要内容
英特尔标志 - 返回主页
我的工具

选择您的语言

  • Bahasa Indonesia
  • Deutsch
  • English
  • Español
  • Français
  • Português
  • Tiếng Việt
  • ไทย
  • 한국어
  • 日本語
  • 简体中文
  • 繁體中文
登录 以访问受限制的内容

使用 Intel.com 搜索

您可以使用几种方式轻松搜索整个 Intel.com 网站。

  • 品牌名称: 酷睿 i9
  • 文件号: 123456
  • Code Name: Emerald Rapids
  • 特殊操作符: “Ice Lake”、Ice AND Lake、Ice OR Lake、Ice*

快速链接

您也可以尝试使用以下快速链接查看最受欢迎搜索的结果。

  • 产品信息
  • 支持
  • 驱动程序和软件

最近搜索

登录 以访问受限制的内容

高级搜索

仅搜索

Sign in to access restricted content.

不建议本网站使用您正在使用的浏览器版本。
请考虑通过单击以下链接之一升级到最新版本的浏览器。

  • Safari
  • Chrome
  • Edge
  • Firefox

Intel® oneAPI Deep Neural Network Library

Increase Deep Learning Framework Performance on CPUs and GPUs

 

  • Overview
  • Download
  • Documentation & Resources

Develop Faster Deep Learning Frameworks and Applications

The Intel® oneAPI Deep Neural Network Library (oneDNN) provides highly optimized implementations of deep learning building blocks. With this open source, cross-platform library, deep learning application and framework developers can use the same API for CPUs, GPUs, or both—it abstracts out instruction sets and other complexities of performance optimization.

Using this library, you can:

  • Improve performance of frameworks you already use, such as OpenVINO™ toolkit, AI Tools from Intel, PyTorch*, and TensorFlow*. 
  • Develop faster deep learning applications and frameworks using optimized building blocks.
  • Deploy applications optimized for Intel CPUs and GPUs without writing any target-specific code.
Download the Stand-Alone Version

A stand-alone download of oneDNN is available. You can download binaries from Intel or choose your preferred repository.

Download
Help oneDNN Evolve

oneDNN is an implementation of the oneAPI specification, part of the Unified Acceleration (UXL) Foundation. We welcome you to participate.

 

Specification

Open Source Version (GitHub*)

Download as Part of the Toolkit

oneDNN is included as part of the Intel® oneAPI Base Toolkit, which is a core set of tools and libraries for developing high-performance, data-centric applications across diverse architectures.

Get It Now

 oneDNN accelerates popular deep learning frameworks such as OpenVINO, PyTorch, and TensorFlow on Intel CPUs and GPUs.

Features

Automatic Optimization

  • Use existing deep learning frameworks
  • Develop and deploy platform-independent deep learning applications with automatic detection of instruction set architecture (ISA) and ISA-specific optimization

 

Network Optimization

  • Identify performance bottlenecks using Intel® VTune™ Profiler
  • Use automatic memory format selection and propagation based on hardware and convolutional parameters
  • Fuse primitives with operations applied to the primitive’s result, for instance, Conv+ReLU
  • Quantize primitives from FP32 to FP16, bf16, or int8 using Intel® Neural Compressor

Optimized Implementations of Key Building Blocks

  • Convolution
  • Matrix multiplication
  • Pooling
  • Batch normalization
  • Activation functions
  • Recurrent neural network (RNN) cells
  • Long short-term memory (LSTM) cells
     

Abstract Programming Model

  • Primitive: Any low-level operation from which more complex operations are constructed, such as convolution, data format reorder, and memory
  • Memory: Handles to memory allocated on a specific engine, tensor dimensions, data type, and memory format
  • Engine: A hardware processing unit, such as a CPU or GPU
  • Stream: A queue of primitive operations on an engine

Learn More


Benchmarks

 Matrix multiplication comparison chart between BF16, FP16, and int8

   Convolution comparison chart between BF16, FP16, and int8

matrix multiply for Intel Xeon 6900 processors with P-cores compared to 5th gen Intel Xeon processors running oneDNN version 3.7

convolution for Intel Xeon 6900 processors with P-cores compared to 5th gen Intel Xeon processors running oneDNN version 3.7

 Matrix multiplication comparison chart between BF16, FP16, and int8

   Convolution comparison chart between BF16, FP16, and int8

matrix multiply for Intel Xeon 6900 processors with P-cores compared to 5th gen Intel Xeon processors running oneDNN version 3.7

convolution for Intel Xeon 6900 processors with P-cores compared to 5th gen Intel Xeon processors running oneDNN version 3.7

Case Studies

Netflix* Improves Video Encoding Performance up to 2X

Netflix* engineers optimized cloud-based video encoding by using oneDNN to get the most out of the Intel® Advanced Vector Extensions 512 (Intel® AVX-512) instruction set.

Learn More

IBM* Research Improves Watson* NLP Performance up to 165%

Adding oneDNN optimizations improved text and sentiment classification task performance up to 35% on 3rd gen Intel® Xeon® Scalable processors. Upgrading to 4th gen Intel Xeon processors with oneDNN led to a 165% improvement.

Learn More

Demonstrations

Accelerate Inference on x86-64 Machines with oneDNN Graph

oneDNN Graph, included in PyTorch 2.0 and beyond, can help accelerate inference on x86-64 CPUs with float32 and bfloat16 datatypes.

Learn More

Optimize Transformer Model Inference on Intel Processors

Fusing TensorFlow operations helps take advantage of optimized oneDNN operations, reducing memory use and inference latency.

Learn More

News

2024 Tools from Intel Now Available with UXL Foundation

The 2024.1 release of oneDNN adds features that streamline storage efficiency, optimize performance on Intel® Xeon® processors, and improve development productivity.

Learn More

oneDNN 3.4 Adds Performance Improvements for CPUs and GPUs

Performance improvements target new and upcoming device support, including improved MATMUL performance for large language models and transformer-style models on Intel CPUs and GPUs.

Learn More

Documentation & Code Samples

Documentation

  • Installation Guide (All Operating Systems)
  • Get Started Guide
  • API Reference
  • Developer Guide and Reference
  • Release Notes
  • System Requirements


View All Documentation

Code Samples

Get Started

oneDNN: Get Started

Learn how to configure and compile oneDNN applications using prebuilt oneDNN binaries, and how to run these applications on different Intel architectures.

oneDNN with SYCL* Interoperability

Use this code sample to learn about programming for Intel CPU and GPU with SYCL* extensions API in oneDNN.

Tutorials

A oneDNN Library for Convolutional Neural Network (CNN) Inference (FP32)

Learn how oneDNN helps to build a neural network topology for forward-pass inference that implements topology layers as numbered primitives.

oneDNN Tutorials

Use these guided samples on a Jupyter* Notebook to examine oneDNN functionality for developing deep learning applications and neural networks, optimized for Intel CPUs and GPUs.

View All oneDNN Samples

View All oneAPI Samples
 

How to work with code samples:

  • Use a command-line interface: Windows* | Linux*
  • Use an IDE: Windows | Linux

Training

Understanding oneDNN

  • Optimized Deep Learning with oneDNN [3:41]
  • oneDNN Library Programming Model and Samples [4:33]

 

AI Model Performance

  • How to Debug AI Model Performance on Intel CPUs

 

🗐 View All Resources

🗗 Training & Events Calendar

Specifications

Processors:

  • Intel Atom® processors with Intel® Streaming SIMD Extensions
  • Intel® Core™ processors
  • Intel® Xeon® processors

 

GPUs:

  • Intel® Processor Graphics Gen9 and above
  • Iris® graphics
  • Intel® Data Center GPUs
  • Intel® Arc™ A-series graphics

 

Host & target operating systems:

  • Linux*
  • Windows*
  • macOS*

Languages:

  • SYCL
    Note Must have Intel oneAPI Base Toolkit installed
  • C and C++

 

Compilers:

  • Intel® oneAPI DPC++/C++ Compiler
  • Clang*
  • GNU C++ Compiler*
  • Microsoft Visual Studio*
  • LLVM* for Apple*

 

Threading runtimes:

  • Intel® oneAPI Threading Building Blocks
  • OpenMP*
  • SYCL

For more information, see the system requirements.

Get Help

Your success is our success. Access these resources when you need assistance.

  • Intel oneAPI Deep Neural Network Library

For additional help, see oneAPI Support.

Stay Up to Date on AI Workload Optimizations

Sign up to receive hand-curated technical articles, tutorials, developer tools, training opportunities, and more to help you accelerate and optimize your end-to-end AI and data science workflows. Take a chance and subscribe. You can change your mind at any time.

除非标为可选,否则所有字段均为必填。

英特尔致力于为您提供优质、个性化的体验,您的数据帮助我们实现这一目标。
本网站采用了 reCAPTCHA 保护机制,并且适用谷歌隐私政策和服务条款。
提交此表单,即表示您确认自己已经年满 18 周岁。英特尔将针对此业务请求处理您的个人数据。要详细了解英特尔的实践,包括如何管理您的偏好和设置,请访问英特尔的隐私声明。
提交此表单,即表示您确认自己已经年满 18 周岁。 英特尔可能会与您联系,以进行与营销相关的沟通。您可以随时选择退出。要详细了解英特尔的实践,包括如何管理您的偏好和设置,请访问英特尔的隐私声明。

You’re In!

Thank you for signing up. Watch for a welcome email to get you started.

  • 公司信息
  • 英特尔资本
  • 企业责任部
  • 投资者关系
  • 联系我们
  • 新闻发布室
  • 网站地图
  • 招贤纳士 (英文)
  • © 英特尔公司
  • 沪 ICP 备 18006294 号-1
  • 使用条款
  • *商标
  • Cookie
  • 隐私条款
  • 请勿分享我的个人信息 California Consumer Privacy Act (CCPA) Opt-Out Icon

英特尔技术可能需要支持的硬件、软件或服务激活。// 没有任何产品或组件能够做到绝对安全。// 您的成本和结果可能会有所不同。// 性能因用途、配置和其他因素而异。// 请参阅我们的完整法律声明和免责声明。// 英特尔致力于尊重人权,并避免成为侵犯人权行为的同谋。请参阅英特尔的《全球人权原则》。英特尔产品和软件仅可用于不会导致或有助于任何国际公认的侵犯人权行为的应用。

英特尔页脚标志