# torch-nccl **Repository Path**: mirrors_NVIDIA/torch-nccl ## Basic Information - **Project Name**: torch-nccl - **Description**: torch bindings for nccl - **Primary Language**: Unknown - **License**: Not specified - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2020-08-18 - **Last Updated**: 2026-03-29 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # nccl.torch Torch7 FFI bindings for NVidia NCCL library. # Installation - Install NCCL from https://github.com/NVIDIA/nccl - Have at least Cuda 7.0 - Have libnccl.so in your library path # Collective operations supported - allReduce - reduce - broadcast - allGather # Example usage Argument to the collective call should be a table of contiguous tensors located on the different devices. Example: perform in-place allReduce on the table of tensors: ```lua require 'nccl' nccl.allReduce(inputs) ``` where inputs is a table of contiguous tensors of the same size located on the different devices.