diff --git a/model_servers/whispercpp/Makefile b/model_servers/whispercpp/Makefile index 7486379b..2f639c8f 100644 --- a/model_servers/whispercpp/Makefile +++ b/model_servers/whispercpp/Makefile @@ -14,7 +14,7 @@ MODEL_NAME ?= ggml-small.bin .PHONY: all all: build download-model-whisper-small run -.PHONY: download-model-whisper-small # small .bin model type testing +.PHONY: download-model-whisper-small download-model-whisper-small: cd ../../models && \ make download-model-whisper-small diff --git a/models/Makefile b/models/Makefile index 76f2e669..92ff4a0d 100644 --- a/models/Makefile +++ b/models/Makefile @@ -34,8 +34,8 @@ download-model-mistral: $(MAKE) MODEL_NAME=mistral-7b-instruct-v0.2.Q4_K_M.gguf MODEL_URL=https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF/resolve/main/mistral-7b-instruct-v0.2.Q4_K_M.gguf download-model .PHONY: download-model-mistral-code -download-mode-mistral-code: - $(MAKE) MODEL_NAME=mistral-7b-code-16k-qlora.Q4_K_M.gguf MODEL_URL=https://huggingface.co/TheBloke/Mistral-7B-Code-16K-qlora-GGUF/resolve/main/mistral-7b-code-16k-qlora.Q4_K_M.gguf +download-model-mistral-code: + $(MAKE) MODEL_NAME=mistral-7b-code-16k-qlora.Q4_K_M.gguf MODEL_URL=https://huggingface.co/TheBloke/Mistral-7B-Code-16K-qlora-GGUF/resolve/main/mistral-7b-code-16k-qlora.Q4_K_M.gguf download-model .PHONY: clean clean: diff --git a/recipes/natural_language_processing/codegen/Makefile b/recipes/natural_language_processing/codegen/Makefile index c45d8a13..b355dc4a 100644 --- a/recipes/natural_language_processing/codegen/Makefile +++ b/recipes/natural_language_processing/codegen/Makefile @@ -8,3 +8,4 @@ RECIPE_BINARIES_PATH := $(shell realpath ../../common/bin) RELATIVE_MODELS_PATH := ../../../models RELATIVE_TESTS_PATH := ../tests MODEL_IMAGE := quay.io/ai-lab/mistral-7b-code-16k-qlora:latest +MODEL_NAME := mistral-7b-code-16k-qlora.Q4_K_M.gguf diff --git a/recipes/natural_language_processing/rag/README.md b/recipes/natural_language_processing/rag/README.md index 0f2487bf..2595dcb2 100644 --- a/recipes/natural_language_processing/rag/README.md +++ b/recipes/natural_language_processing/rag/README.md @@ -58,15 +58,9 @@ performant mid-sized model with an apache-2.0 license that has been quanitzed an The recommended model can be downloaded using the code snippet below: ```bash -<<<<<<< HEAD -cd models -wget https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF/resolve/main/mistral-7b-instruct-v0.2.Q4_K_M.gguf -cd ../ -======= cd ../../../models curl -sLO https://huggingface.co/instructlab/granite-7b-lab-GGUF/resolve/main/granite-7b-lab-Q4_K_M.gguf cd ../recipes/natural_language_processing/rag ->>>>>>> 4f8f404 (swapping standard mistral with granite as default) ``` _A full list of supported open models is forthcoming._