Commit Graph

3 Commits

Author SHA256 Message Date
9c8ce17a59 - update to 2.3.1 with following summarized highlights:
* from 2.0.x:
    - torch.compile is the main API for PyTorch 2.0, which wraps your model and
      returns a compiled model. It is a fully additive (and optional) feature
      and hence 2.0 is 100% backward compatible by definition
    - Accelerated Transformers introduce high-performance support for training
      and inference using a custom kernel architecture for scaled dot product
      attention (SPDA). The API is integrated with torch.compile() and model
      developers may also use the scaled dot product attention kernels directly
      by calling the new scaled_dot_product_attention() operato
  * from 2.1.x:
    - automatic dynamic shape support in torch.compile,
      torch.distributed.checkpoint for saving/loading distributed training jobs
      on multiple ranks in parallel, and torch.compile support for the NumPy
      API.
    - In addition, this release offers numerous performance improvements (e.g.
      CPU inductor improvements, AVX512 support, scaled-dot-product-attention
      support) as well as a prototype release of torch.export, a sound
      full-graph capture mechanism, and torch.export-based quantization.
  * from 2.2.x:
    - 2x performance improvements to scaled_dot_product_attention via
      FlashAttention-v2 integration, as well as AOTInductor, a new
      ahead-of-time compilation and deployment tool built for non-python
      server-side deployments.
  * from 2.3.x:
    - support for user-defined Triton kernels in torch.compile, allowing for
      users to migrate their own Triton kernels from eager without
      experiencing performance complications or graph breaks. As well, Tensor
      Parallelism improves the experience for training Large Language Models
      using native PyTorch functions, which has been validated on training

OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/python-torch?expand=0&rev=32
2024-07-19 12:15:19 +00:00
73b1680af3 OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/python-torch?expand=0&rev=31 2022-05-24 05:53:14 +00:00
0a90b0a626 - updated to stable release 1.4.0, which has as Highlights:
* Distributed Model Parallel Training
  * Pruning functionalities have been added to PyTorch 
- New Features:
  * torch.optim.lr_scheduler now support “chaining.”
  * torch.distributed.rpc is a newly introduced package
- full Changelog listed in relases file or under
  https://github.com/pytorch/pytorch/releases
- added files: 
  * skip-third-party-check.patch which is a patch to skip
    the check of disabled dependencies
  * QNNPACK-7d2a4e9931a82adc3814275b6219a03e24e36b4c.tar.gz
    which is part of pytorch but developed in different repo
- removed patch files:
  * fix-build-options.patch
  * honor-PSIMD-env.patch
  * removed-some-tests.patch
- Requires python-PeachPy on x86_64 only, as it is optional
  and available on x86_64 only

OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/python-torch?expand=0&rev=7
2020-02-21 15:50:33 +00:00