Skip to content

Releases: B0-B/blowtorch-transformer-api

blowtorch 1.3.0

10 Oct 00:29
Compare
Choose a tag to compare

Release Updates

  • Added vLLM support for accelerated attention!
  • Better loading algorithms and post processing of outputs.
  • Included automatic stopping in generation.
  • New example files added

v1.2.2

03 Jul 13:57
Compare
Choose a tag to compare

blowtorch 1.2.2

Release Updates

  • Added automated context-length detection to stdout so the user is always aware about the current context length.
  • Automated context trimming: If the current aggregated context is too long for the context length the auto_trim=True argument (can be called in client init or config) will ensure the most recent context which does not overflow the allowed length. Otherwise users are prone to run into errors like this.
  • Added cut_unfinished argument to client. If enabled all outputs will be truncated to the last fulfilled sentence, unfinished sentences will be cutoff.

v1.2.1

07 May 13:36
Compare
Choose a tag to compare

blowtorch 1.2.1

Updates

  • adding llama.cpp support for cpu and optionally gpu
  • more robust code for parameter handling of various backends
  • improved setup with tested depenency versions (should work for all OS)