跳转至主要内容
英特尔标志 - 返回主页
我的工具

选择您的语言

  • Bahasa Indonesia
  • Deutsch
  • English
  • Español
  • Français
  • Português
  • Tiếng Việt
  • ไทย
  • 한국어
  • 日本語
  • 简体中文
  • 繁體中文
登录 以访问受限制的内容

使用 Intel.com 搜索

您可以使用几种方式轻松搜索整个 Intel.com 网站。

  • 品牌名称: 酷睿 i9
  • 文件号: 123456
  • Code Name: Emerald Rapids
  • 特殊操作符: “Ice Lake”、Ice AND Lake、Ice OR Lake、Ice*

快速链接

您也可以尝试使用以下快速链接查看最受欢迎搜索的结果。

  • 产品信息
  • 支持
  • 驱动程序和软件

最近搜索

登录 以访问受限制的内容

高级搜索

仅搜索

Sign in to access restricted content.

不建议本网站使用您正在使用的浏览器版本。
请考虑通过单击以下链接之一升级到最新版本的浏览器。

  • Safari
  • Chrome
  • Edge
  • Firefox

Introduction to Getting Faster PyTorch* Programs with TorchDynamo

@IntelDevTools


Subscribe Now

Stay in the know on all things CODE. Updates are delivered to your inbox.

Sign Up

Overview

Learn the principles and techniques for making your PyTorch* programs faster, more useable, and performant with the framework’s easy-to-use, just-in-time (JIT) compiler, TorchDynamo.

Introduced in PyTorch 2.0, TorchDynamo transforms a general Python* program into a computational graph. It works by hooking into the Python frame-evaluation process, and then fundamentally changing and supercharging PyTorch operations at the compiler level in the back end.

This session presents this JIT graph compiler and includes:

  • An overview of its design and use, including flexible support for graph acquisition with better performance and usability.
  • The novel TorchDynamo technique principles in PyTorch 2.0, including debugging.
  • How to use Intel® Extension for PyTorch* for TorchDynamo use.
  • Intel’s support for and contributions to the tools.

This session includes demos.

Skill level: Intermediate

 

Featured Software

Get the following stand-alone versions of the tools:

  • Intel Extension for PyTorch
  • TorchDynamo (GitHub*)

 

Download Code Samples

  • Intel Extension for PyTorch (GitHub)
  • TorchDynamo Benchmarks (GitHub)

See All Code Samples

 

More Resources

  • Go Deeper with TorchDynamo
  • What’s Behind PyTorch 2.0: TorchDynamo and TorchInductor (Especially for Developers)

 

 

Jump to:

You May Also Like
 

PyTorch* Optimizations from Intel

Intel is one of the largest contributors to PyTorch*, providing regular upstream optimizations to the PyTorch deep learning framework that provide superior performance on Intel® architectures. The AI Tools includes the latest binary version of PyTorch tested to work with the rest of the kit, along with Intel® Extension for PyTorch*, which adds the newest Intel optimizations and usability features.

 

Get as Part of the AI Tools

 

PyTorch

  

Intel Extension for PyTorch

 

See All Tools

 

   

You May Also Like

Related Articles

PyTorch 2.0 Release Includes AI Performance Features from Intel

Three Ways to Accelerate PyTorch Geometric on Intel CPUs

Improve Inference with int8 Quantization for x86 CPU Platforms in PyTorch

Use PyTorch and AI Everywhere: Intel Joins the PyTorch Foundation

Accelerate Inference on x86-64 Machines with a oneAPI Deep Neural Network Library (oneDNN) Graph

Related Videos

PyTorch 2.0 Overview and Technical Analysis of Its Deep Learning Compiler

Intel Extension for PyTorch: New Features on CPUs and GPUs

Get Started with Intel Extension for PyTorch on a GPU

  • 公司信息
  • 英特尔资本
  • 企业责任部
  • 投资者关系
  • 联系我们
  • 新闻发布室
  • 网站地图
  • 招贤纳士 (英文)
  • © 英特尔公司
  • 沪 ICP 备 18006294 号-1
  • 使用条款
  • *商标
  • Cookie
  • 隐私条款
  • 请勿分享我的个人信息 California Consumer Privacy Act (CCPA) Opt-Out Icon

英特尔技术可能需要支持的硬件、软件或服务激活。// 没有任何产品或组件能够做到绝对安全。// 您的成本和结果可能会有所不同。// 性能因用途、配置和其他因素而异。请访问 intel.cn/performanceindex 了解更多信息。// 请参阅我们的完整法律声明和免责声明。// 英特尔致力于尊重人权,并避免成为侵犯人权行为的同谋。请参阅英特尔的《全球人权原则》。英特尔产品和软件仅可用于不会导致或有助于任何国际公认的侵犯人权行为的应用。

英特尔页脚标志