From baeaf3b8ab505d38d3fa5dd0a991241181ad0e00 Mon Sep 17 00:00:00 2001 From: Margaret Qian Date: Tue, 29 Aug 2023 17:37:52 -0700 Subject: [PATCH] llama2 lint --- examples/inference-deployments/llama2/README.md | 2 +- examples/inference-deployments/llama2/llama2_13b.yaml | 2 +- examples/inference-deployments/llama2/llama2_7b_chat.yaml | 2 +- examples/inference-deployments/llama2/requirements.txt | 1 + examples/inference-deployments/mpt/mpt_30b_chat.yaml | 2 +- 5 files changed, 5 insertions(+), 4 deletions(-) create mode 100644 examples/inference-deployments/llama2/requirements.txt diff --git a/examples/inference-deployments/llama2/README.md b/examples/inference-deployments/llama2/README.md index 37813e825..8ffbec30a 100644 --- a/examples/inference-deployments/llama2/README.md +++ b/examples/inference-deployments/llama2/README.md @@ -61,7 +61,7 @@ Once the deployment is ready, it's time to run inference! Detailed information a from mcli import predict prompt = """[INST] <> -You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. +You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature. If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information. <> diff --git a/examples/inference-deployments/llama2/llama2_13b.yaml b/examples/inference-deployments/llama2/llama2_13b.yaml index 140ff1909..a7fc1745e 100644 --- a/examples/inference-deployments/llama2/llama2_13b.yaml +++ b/examples/inference-deployments/llama2/llama2_13b.yaml @@ -10,4 +10,4 @@ compute: image: mosaicml/inference:0.1.37 cluster: r7z15 default_model: - model_type: llama2-13b \ No newline at end of file + model_type: llama2-13b diff --git a/examples/inference-deployments/llama2/llama2_7b_chat.yaml b/examples/inference-deployments/llama2/llama2_7b_chat.yaml index a304ed53c..c9f8047d9 100644 --- a/examples/inference-deployments/llama2/llama2_7b_chat.yaml +++ b/examples/inference-deployments/llama2/llama2_7b_chat.yaml @@ -10,4 +10,4 @@ compute: image: mosaicml/inference:0.1.37 cluster: r7z15 default_model: - model_type: llama2-7b-chat \ No newline at end of file + model_type: llama2-7b-chat diff --git a/examples/inference-deployments/llama2/requirements.txt b/examples/inference-deployments/llama2/requirements.txt new file mode 100644 index 000000000..422166bae --- /dev/null +++ b/examples/inference-deployments/llama2/requirements.txt @@ -0,0 +1 @@ +torch==1.13.1 diff --git a/examples/inference-deployments/mpt/mpt_30b_chat.yaml b/examples/inference-deployments/mpt/mpt_30b_chat.yaml index 9e37b2ef3..771eed921 100644 --- a/examples/inference-deployments/mpt/mpt_30b_chat.yaml +++ b/examples/inference-deployments/mpt/mpt_30b_chat.yaml @@ -5,4 +5,4 @@ compute: image: mosaicml/inference:0.1.37 replicas: 1 default_model: - model_type: mpt-30b-chat \ No newline at end of file + model_type: mpt-30b-chat