1. 08 Jun, 2021 1 commit
    • Enhance transpose operators · 936c351b
      Summary:
      This commit allows transpose to compute in-place by leveraging buffer.
      We also adds CRD mode for space-depth transpose (i.e., pixel shuffle).
      Ting PAN committed
  2. 31 May, 2021 1 commit
  3. 13 May, 2021 1 commit
  4. 07 May, 2021 1 commit
  5. 01 May, 2021 1 commit
  6. 28 Apr, 2021 1 commit
  7. 21 Apr, 2021 1 commit
    • Add GELU operator · bdf4e10f
      Summary:
      This commit adds GELU activation to compute output
      via approximate or naive mode.
      Ting PAN committed
  8. 14 Apr, 2021 1 commit
  9. 08 Apr, 2021 1 commit
    • Update with the new frontend API · f431756f
      Summary:
      The new frontend makes an union of two execution modes, while starts from
      a single tensor class. Besides, it emits the operator execution through
      a common path that works both for dragon and torch.
      Ting PAN committed
  10. 04 Feb, 2021 1 commit
    • Reimplement the general matrix multiplication · 6bfe3e73
      Summary:
      This commit generalizes the fully-connected operation into GEMM,
      and enhances the matmul operation via batched Dot, GEMV and GEMM.
      New representations and attributes have been consistent with ONNX.
      Ting PAN committed
  11. 25 Jan, 2021 1 commit
    • Remove support for CUDNN v6 · 73ed1b96
      Summary:
      For the purpose of consistency on getting CUDNN convolution algorithms,
      CUDNN v6 (mainly relied by CUDA 8.0) is now dropped.
      Ting PAN committed
  12. 20 Jan, 2021 1 commit
    • Add sysconfig module · bbfecf22
      Summary:
      This commit adds the sysconfig module to get the build information.
      Build information is helpful to select tests or report issues.
      Ting PAN committed
  13. 16 Jan, 2021 1 commit
  14. 29 Dec, 2020 1 commit
  15. 23 Dec, 2020 1 commit
  16. 15 Dec, 2020 1 commit
  17. 11 Dec, 2020 1 commit
  18. 10 Dec, 2020 1 commit
  19. 09 Dec, 2020 1 commit
    • Refactor ONNX frontends and backends · b93bde0d
      Summary:
      This commit redesigns the ``vm.onnx`` by referring the official repository.
      Frontends and backends are aligned with identical API for dragon, torch and tensorrt.
      Ting PAN committed
  20. 03 Dec, 2020 1 commit
  21. 02 Dec, 2020 1 commit
  22. 29 Nov, 2020 1 commit
  23. 05 Nov, 2020 1 commit
    • Use FP32 accumulator for FP16 ReduceSum · d56e67d1
      Summary:
      This commit adds a fallback with FP32 accumulator
      for FP16 ReduceSum to avoid dropping too many small values.
      Besides, FP16 kernels for arch < 530 are almost available.
      Ting PAN committed
  24. 24 Oct, 2020 1 commit
  25. 20 Oct, 2020 1 commit
  26. 14 Oct, 2020 1 commit
  27. 13 Oct, 2020 1 commit
    • Add LinSpace Operator · e83c407a
      Summary:
      This commit adds the linspace op for dragon, torch and tensorflow.
      And, a workaround for truncated int interval is made to range/linspace (up to 2**57).
      Ting PAN committed
  28. 08 Oct, 2020 1 commit
  29. 07 Oct, 2020 1 commit
    • Add Sort Operator · b4019faa
      Summary:
      This commit adds the sort op for dragon, torch and tensorflow.
      Besides, cuda implementation of topk op is now available.
      Ting PAN committed
  30. 27 Sep, 2020 1 commit
    • Use local workspace for Context · fdf26ef2
      Summary:
      This commit uses local(thread or stream) workspace for Context,
      which provides a more elegant way to dispatch kernels requiring scratch.
      Besides, TF32 math type is provided as a cuDNN option for Ampere device.
      Ting PAN committed
  31. 10 Sep, 2020 1 commit
    • Add Unique Operator · 1dd8aeef
      Summary:
      This commit adds the unique op for dragon, torch, tensorflow and onnx.
      Besides, fixes the bug that gets the wrong workspace size in cached cudnn convolution.
      Ting PAN committed
  32. 05 Sep, 2020 1 commit
    • Use sequential sampling as the default shuffle policy · 80267d8f
      Summary:
      This commit reimplements the default shuffle policy of data reader with
      sequential sampling (be consistent with DALI) instead of chunk permutation (MXNet solution).
      Sequential sampling is tuned by argument ``initial_fill`` only, and works both for HDD and SSD.
      Ting PAN committed
  33. 30 Aug, 2020 1 commit
  34. 23 Aug, 2020 1 commit
    • Fix the stream issue with NCCL2 on CUDA 9.2 and later · 58708021
      Summary:
      This commit enforces the stream synchronization before dispatching NCCL collectives.
      Otherwise, data corruption will happen due to the default value of ``NCCL_GROUP_CUDA_STREAM``
      changed to 0 since CUDA 9.2, i.e., no explicit event waiting for unfinished kernels.
      Ting PAN committed
  35. 12 Aug, 2020 1 commit
  36. 07 Aug, 2020 1 commit
  37. 05 Aug, 2020 1 commit
    • Remove the deprecated DALI API · 218796ed
      Summary:
      This commit removes the deprecated API for DALI 0.24.
      Besides, variable length keyword arguments are added for forward compatibility.
      Ting PAN committed
  38. 03 Aug, 2020 1 commit
  39. 30 Jul, 2020 1 commit
  40. 25 Jul, 2020 1 commit