_            _    _        _         _
      /\ \         /\ \ /\ \     /\_\      / /\
      \_\ \       /  \ \\ \ \   / / /     / /  \
      /\__ \     / /\ \ \\ \ \_/ / /     / / /\ \__
     / /_ \ \   / / /\ \ \\ \___/ /     / / /\ \___\
    / / /\ \ \ / / /  \ \_\\ \ \_/      \ \ \ \/___/
   / / /  \/_// / /   / / / \ \ \        \ \ \
  / / /      / / /   / / /   \ \ \   _    \ \ \
 / / /      / / /___/ / /     \ \ \ /_/\__/ / /
/_/ /      / / /____\/ /       \ \_\\ \/___/ /
\_\/       \/_________/         \/_/ \_____\/
python-accelerate 1.12.0
Propagated dependencies: python-huggingface-hub@0.31.4 python-numpy@1.26.4 python-packaging@25.0 python-psutil@7.0.0 python-pytorch@2.8.0 python-pyyaml@6.0.2 python-safetensors@0.4.3
Channel: guix-science
Location: guix-science/packages/machine-learning.scm (guix-science packages machine-learning)
Home page: https://huggingface.co/docs/accelerate/
Licenses: ASL 2.0
Synopsis: Launch, train and use PyTorch models on any configuration
Description:

Accelerate was created for PyTorch users who like to write the training loop of PyTorch models but are reluctant to write and maintain the boilerplate code needed to use multi-GPUs/TPU/fp16. It abstracts exactly and only the boilerplate code related to multi-GPUs/TPU/fp16 and leaves the rest of your code unchanged.

Total results: 1