Skip to content

Latest commit

 

History

History
122 lines (83 loc) · 3.28 KB

CHANGELOG.md

File metadata and controls

122 lines (83 loc) · 3.28 KB

Changelog for hipBLASLt

Full documentation for hipBLASLt is available at rocm.docs.amd.com/projects/hipBLASLt.

(Unreleased) hipBLASLt 0.8.0

Additions

  • Extension APIs:
    • hipblasltExtAMaxWithScale
  • GemmTuning extension parameter to set wgm by user
  • Support HIPBLASLT_MATMUL_DESC_AMAX_D_POINTER for FP8/BF8 datatype
  • Support for FP8/BF8 input, FP32/FP16/BF16/F8/BF8 output (only for gfx94x platform)
  • Support HIPBLASLT_MATMUL_DESC_COMPUTE_INPUT_TYPE_A_EXT and HIPBLASLT_MATMUL_DESC_COMPUTE_INPUT_TYPE_B_EXT for FP16 input datatype to use FP8/BF8 mfma
  • Support for gfx110x

Optimizations

  • Improve library loading time

(Unreleased) hipBLASLt 0.7.0

Additions

  • Extension APIs:
    • hipblasltExtSoftmax
    • hipblasltExtLayerNorm
    • hipblasltExtAMax
  • GemmTuning extension parameter to set split-k by user
  • Support for mixed-precision datatype: FP16/FP8 in with FP16 out
  • Add CMake support for documentation
  • Support for datatype: Int8 in Int8 out

Deprecations

  • algoGetHeuristic() ext API for GroupGemm will be deprecated in a future release of hipBLASLt

hipBLASLt 0.6.0

Additions

  • New UserArguments variable for GroupedGemm
  • Support for datatype: FP16 in with FP32 out
  • Support for datatype: Int8 in Int32 out
  • Support for gfx94x platform
  • Support for FP8/BF8 datatype (only for gfx94x platform)
  • Support scalar A,B,C,D for FP8/BF8 datatype
  • Added samples

Changes

  • Replaced hipblasDatatype_t with hipDataType
  • Replaced hipblasLtComputeType_t with hipblasComputeType_t

Removals

  • Deprecated HIPBLASLT_MATMUL_DESC_D_SCALE_VECTOR_POINTER

hipBLASLt 0.3.0

Additions

  • Added getAllAlgos extension APIs
  • TensileLite support for new epilogues: gradient gelu, gradient D, gradient A/B, aux
  • Added a sample package that includes three sample apps
  • Added a new C++ GEMM class in the hipBLASLt extension

Changes

  • Refactored GroupGemm APIs as C++ class in the hipBLASLt extension
  • Changed the scaleD vector enum to HIPBLASLT_MATMUL_DESC_D_SCALE_VECTOR_POINTER

Fixes

  • Enabled norm check validation for CI

Optimizations

  • GSU kernel: wider memory, PGR N
  • Updated logic yaml to improve some FP16 NN sizes
  • GroupGemm support for GSU kernel
  • Added grouped GEMM tuning for aldebaran

hipBLASLt 0.2.0

Additions

  • Added CI tests for TensileLite
  • Initialized extension group GEMM APIs (FP16 only)
  • Added a group GEMM sample app: example_hipblaslt_groupedgemm

Fixes

  • Fixed incorrect results for the ScaleD kernel

Optimizations

  • Tuned equality sizes for the HHS data type
  • Reduced host-side overhead for hipblasLtMatmul()
  • Removed unused kernel arguments
  • Schedule values setup before first s_waitcnt
  • Refactored TensileLite host codes
  • Optimized build time

hipBLASLt 0.1.0

Additions

  • Enabled hipBLASLt APIs
  • Support for gfx90a
  • Support for problem type: FP32, FP16, BF16
  • Support activation: relu, gelu
  • Support for bias vectors
  • Integrated with TensileLite kernel generator
  • Added Gtest: hipblaslt-test
  • Added the full function tool hipblaslt-bench
  • Added the sample app example_hipblaslt_preference

Optimizations

  • gridBase solution search algorithm for untuned size
  • Tuned 10k sizes for each problem type