- update to 2.3.1 with following summarized highlights:
* from 2.0.x:
- torch.compile is the main API for PyTorch 2.0, which wraps your model and
returns a compiled model. It is a fully additive (and optional) feature
and hence 2.0 is 100% backward compatible by definition
- Accelerated Transformers introduce high-performance support for training
and inference using a custom kernel architecture for scaled dot product
attention (SPDA). The API is integrated with torch.compile() and model
developers may also use the scaled dot product attention kernels directly
by calling the new scaled_dot_product_attention() operato
* from 2.1.x:
- automatic dynamic shape support in torch.compile,
torch.distributed.checkpoint for saving/loading distributed training jobs
on multiple ranks in parallel, and torch.compile support for the NumPy
API.
- In addition, this release offers numerous performance improvements (e.g.
CPU inductor improvements, AVX512 support, scaled-dot-product-attention
support) as well as a prototype release of torch.export, a sound
full-graph capture mechanism, and torch.export-based quantization.
* from 2.2.x:
- 2x performance improvements to scaled_dot_product_attention via
FlashAttention-v2 integration, as well as AOTInductor, a new
ahead-of-time compilation and deployment tool built for non-python
server-side deployments.
* from 2.3.x:
- support for user-defined Triton kernels in torch.compile, allowing for
users to migrate their own Triton kernels from eager without
experiencing performance complications or graph breaks. As well, Tensor
Parallelism improves the experience for training Large Language Models
using native PyTorch functions, which has been validated on training
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/python-torch?expand=0&rev=32
2024-07-19 14:15:19 +02:00
|
|
|
From 4f4138cfc5a4ea59384b613a4389bf388ebfa7bb Mon Sep 17 00:00:00 2001
|
|
|
|
From: Christian Goll <cgoll@suse.com>
|
|
|
|
Date: Fri, 5 Jul 2024 14:53:32 +0200
|
|
|
|
Subject: [PATCH] skip-third-party-check
|
|
|
|
|
|
|
|
Signed-off-by: Christian Goll <cgoll@suse.com>
|
|
|
|
---
|
|
|
|
cmake/public/utils.cmake | 3 ++-
|
|
|
|
setup.py | 2 +-
|
|
|
|
2 files changed, 3 insertions(+), 2 deletions(-)
|
|
|
|
|
|
|
|
diff --git a/cmake/public/utils.cmake b/cmake/public/utils.cmake
|
|
|
|
index 78a90dbc587..172f8ef422c 100644
|
|
|
|
--- a/cmake/public/utils.cmake
|
|
|
|
+++ b/cmake/public/utils.cmake
|
|
|
|
@@ -185,7 +185,8 @@ endfunction()
|
|
|
|
macro(caffe2_update_option variable value)
|
|
|
|
if(CAFFE2_CMAKE_BUILDING_WITH_MAIN_REPO)
|
|
|
|
get_property(__help_string CACHE ${variable} PROPERTY HELPSTRING)
|
|
|
|
- set(${variable} ${value} CACHE BOOL ${__help_string} FORCE)
|
|
|
|
+ #set(${variable} ${value} CACHE BOOL ${__help_string} FORCE)
|
|
|
|
+ set(${variable} ${value})
|
|
|
|
else()
|
|
|
|
set(${variable} ${value})
|
|
|
|
endif()
|
|
|
|
diff --git a/setup.py b/setup.py
|
|
|
|
index 6b0860a9ee1..8d1aaff5668 100644
|
|
|
|
--- a/setup.py
|
|
|
|
+++ b/setup.py
|
|
|
|
@@ -458,7 +458,7 @@ def mirror_files_into_torchgen():
|
|
|
|
def build_deps():
|
|
|
|
report("-- Building version " + version)
|
2020-02-21 16:50:33 +01:00
|
|
|
|
- update to 2.3.1 with following summarized highlights:
* from 2.0.x:
- torch.compile is the main API for PyTorch 2.0, which wraps your model and
returns a compiled model. It is a fully additive (and optional) feature
and hence 2.0 is 100% backward compatible by definition
- Accelerated Transformers introduce high-performance support for training
and inference using a custom kernel architecture for scaled dot product
attention (SPDA). The API is integrated with torch.compile() and model
developers may also use the scaled dot product attention kernels directly
by calling the new scaled_dot_product_attention() operato
* from 2.1.x:
- automatic dynamic shape support in torch.compile,
torch.distributed.checkpoint for saving/loading distributed training jobs
on multiple ranks in parallel, and torch.compile support for the NumPy
API.
- In addition, this release offers numerous performance improvements (e.g.
CPU inductor improvements, AVX512 support, scaled-dot-product-attention
support) as well as a prototype release of torch.export, a sound
full-graph capture mechanism, and torch.export-based quantization.
* from 2.2.x:
- 2x performance improvements to scaled_dot_product_attention via
FlashAttention-v2 integration, as well as AOTInductor, a new
ahead-of-time compilation and deployment tool built for non-python
server-side deployments.
* from 2.3.x:
- support for user-defined Triton kernels in torch.compile, allowing for
users to migrate their own Triton kernels from eager without
experiencing performance complications or graph breaks. As well, Tensor
Parallelism improves the experience for training Large Language Models
using native PyTorch functions, which has been validated on training
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/python-torch?expand=0&rev=32
2024-07-19 14:15:19 +02:00
|
|
|
- check_submodules()
|
|
|
|
+ #check_submodules()
|
|
|
|
check_pydep("yaml", "pyyaml")
|
2020-02-21 16:50:33 +01:00
|
|
|
|
- update to 2.3.1 with following summarized highlights:
* from 2.0.x:
- torch.compile is the main API for PyTorch 2.0, which wraps your model and
returns a compiled model. It is a fully additive (and optional) feature
and hence 2.0 is 100% backward compatible by definition
- Accelerated Transformers introduce high-performance support for training
and inference using a custom kernel architecture for scaled dot product
attention (SPDA). The API is integrated with torch.compile() and model
developers may also use the scaled dot product attention kernels directly
by calling the new scaled_dot_product_attention() operato
* from 2.1.x:
- automatic dynamic shape support in torch.compile,
torch.distributed.checkpoint for saving/loading distributed training jobs
on multiple ranks in parallel, and torch.compile support for the NumPy
API.
- In addition, this release offers numerous performance improvements (e.g.
CPU inductor improvements, AVX512 support, scaled-dot-product-attention
support) as well as a prototype release of torch.export, a sound
full-graph capture mechanism, and torch.export-based quantization.
* from 2.2.x:
- 2x performance improvements to scaled_dot_product_attention via
FlashAttention-v2 integration, as well as AOTInductor, a new
ahead-of-time compilation and deployment tool built for non-python
server-side deployments.
* from 2.3.x:
- support for user-defined Triton kernels in torch.compile, allowing for
users to migrate their own Triton kernels from eager without
experiencing performance complications or graph breaks. As well, Tensor
Parallelism improves the experience for training Large Language Models
using native PyTorch functions, which has been validated on training
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/python-torch?expand=0&rev=32
2024-07-19 14:15:19 +02:00
|
|
|
build_caffe2(
|
|
|
|
--
|
|
|
|
2.43.0
|
|
|
|
|