From 3c8fc57eaaac5420e35a9716dc6c5145d5797773 Mon Sep 17 00:00:00 2001 From: biswaroop1547 Date: Fri, 28 Jul 2023 11:38:18 +0000 Subject: [PATCH] update: build version --- cht-llama-v2/build.sh | 2 +- cht-llama-v2/models.py | 6 ++++-- 2 files changed, 5 insertions(+), 3 deletions(-) diff --git a/cht-llama-v2/build.sh b/cht-llama-v2/build.sh index c3f4de7..dd5d5e5 100755 --- a/cht-llama-v2/build.sh +++ b/cht-llama-v2/build.sh @@ -1,6 +1,6 @@ #!/bin/bash set -e -export VERSION=1.0.0 +export VERSION=1.0.1 IMAGE=ghcr.io/premai-io/chat-llama-2-7b-gpu docker buildx build ${@:1} \ diff --git a/cht-llama-v2/models.py b/cht-llama-v2/models.py index 2a09098..4fd6eca 100644 --- a/cht-llama-v2/models.py +++ b/cht-llama-v2/models.py @@ -32,7 +32,6 @@ def embeddings(cls, text) -> None: pass @abstractmethod - @staticmethod def stitch_prompt(messages: list) -> str: pass @@ -68,7 +67,10 @@ def generate( do_sample=kwargs.get("do_sample", True), stop_sequence=stop[0] if stop else None, stopping_criteria=cls.stopping_criteria(stop, prompt, cls.tokenizer), - )[0]["generated_text"].rstrip(stop[0] if stop else "") + )[0]["generated_text"] + .rstrip(stop[0] if stop else "") + .rsplit(".", 1)[0] + .strip() ] @classmethod