From 34e29f83104acc31ed74b3985edfcfd29eea0ec6 Mon Sep 17 00:00:00 2001 From: Amit Raj <168538872+quic-amitraj@users.noreply.github.com> Date: Tue, 10 Sep 2024 17:28:57 +0530 Subject: [PATCH] Updated README.md with latest changes (#105) Updated links of README.md Signed-off-by: Amit Raj --- README.md | 17 ++++++++++------- docs/source/introduction.md | 8 ++++++-- docs/source/validate.md | 6 ++++-- 3 files changed, 20 insertions(+), 11 deletions(-) diff --git a/README.md b/README.md index 62d02100..a3f823d9 100644 --- a/README.md +++ b/README.md @@ -7,8 +7,12 @@ *Latest news* :fire:
-- [coming soon] Support for more popular [models](#models-coming-soon) and inference optimization techniques like continuous batching and speculative decoding
-* [06/2024] Added support for [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b) +- [coming soon] Support for more popular [models](coming_soon_models) and inference optimization technique speculative decoding
+- [08/2024] Added Support for inference optimization technique ```continuous batching``` +- [08/2024] Added support for [Jais-adapted-70b](https://huggingface.co/inceptionai/jais-adapted-70b) +- [08/2024] Added support for [Jais-adapted-13b-chat](https://huggingface.co/inceptionai/jais-adapted-13b-chat) +- [08/2024] Added support for [Jais-adapted-7b](https://huggingface.co/inceptionai/jais-adapted-7b) +- [06/2024] Added support for [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b) - [06/2024] Added support for [Qwen2-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct) - [06/2024] Added support for [StarCoder2-15B](https://huggingface.co/bigcode/starcoder2-15b) - [06/2024] Added support for [Phi3-Mini-4K-Instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) @@ -17,7 +21,6 @@ - [05/2024] Added support for [Mixtral-8x7B](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) & [Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1). - [04/2024] Initial release of [efficient transformers](https://github.com/quic/efficient-transformers) for seamless inference on pre-trained LLMs. - # Overview ## Train anywhere, Infer on Qualcomm Cloud AI with a Developer-centric Toolchain @@ -60,15 +63,15 @@ pip install dist/QEfficient-0.0.1.dev0-py3-none-any.whl ``` -For more details about using ``QEfficient`` via Cloud AI 100 Apps SDK, visit [Linux Installation Guide](https://quic.github.io/efficient-transformers/source/Linux_installation.html) +For more details about using ``QEfficient`` via Cloud AI 100 Apps SDK, visit [Linux Installation Guide](https://quic.github.io/efficient-transformers/source/installation.html) ## Documentation * [Quick Start Guide](https://quic.github.io/efficient-transformers/source/quick_start.html#) -* [Python API](https://quic.github.io/efficient-transformers/source/high_level_api.html) -* [Validated Models](https://quic.github.io/efficient-transformers/source/Validate.html) -* [Models coming soon](https://quic.github.io/efficient-transformers/source/Validate.html#models-coming-soon) +* [Python API](https://quic.github.io/efficient-transformers/source/hl_api.html) +* [Validated Models](https://quic.github.io/efficient-transformers/source/validate.html) +* [Models coming soon](https://quic.github.io/efficient-transformers/source/validate.html#models-coming-soon) > Note: More details are here: https://quic.github.io/cloud-ai-sdk-pages/latest/Getting-Started/Model-Architecture-Support/Large-Language-Models/llm/ diff --git a/docs/source/introduction.md b/docs/source/introduction.md index f6a21410..2e72b97a 100644 --- a/docs/source/introduction.md +++ b/docs/source/introduction.md @@ -22,8 +22,12 @@ For other models, there is comprehensive documentation to inspire upon the chang ***Latest news*** :
-- [coming soon] Support for more popular [models](coming_soon_models) and inference optimization techniques like continuous batching and speculative decoding
-* [06/2024] Added support for [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b) +- [coming soon] Support for more popular [models](coming_soon_models) and inference optimization technique speculative decoding
+- [08/2024] Added Support for inference optimization technique ```continuous batching``` +- [08/2024] Added support for [Jais-adapted-70b](https://huggingface.co/inceptionai/jais-adapted-70b) +- [08/2024] Added support for [Jais-adapted-13b-chat](https://huggingface.co/inceptionai/jais-adapted-13b-chat) +- [08/2024] Added support for [Jais-adapted-7b](https://huggingface.co/inceptionai/jais-adapted-7b) +- [06/2024] Added support for [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b) - [06/2024] Added support for [Qwen2-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct) - [06/2024] Added support for [StarCoder2-15B](https://huggingface.co/bigcode/starcoder2-15b) - [06/2024] Added support for [Phi3-Mini-4K-Instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) diff --git a/docs/source/validate.md b/docs/source/validate.md index a367d99a..a95db12c 100644 --- a/docs/source/validate.md +++ b/docs/source/validate.md @@ -26,10 +26,12 @@ | [Codestral-22B-v0.1](https://huggingface.co/mistralai/Codestral-22B-v0.1) | ✔️ | | | [Falcon-40b](https://huggingface.co/tiiuae/falcon-40b) | ✔️ | | | [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b) | ✔️ | | +| [Jais-adapted-70b](https://huggingface.co/inceptionai/jais-adapted-70b) | ✔️ | ✔️ | +| [Jais-adapted-13b-chat](https://huggingface.co/inceptionai/jais-adapted-13b-chat) | ✔️ | ✔️ | +| [Jais-adapted-7b](https://huggingface.co/inceptionai/jais-adapted-7b) | ✔️ | ✔️ | + (coming_soon_models)= # Models Coming Soon -* [Jais-13b](https://huggingface.co/core42/jais-13b) -* [Jais-30b](https://huggingface.co/core42/jais-30b-chat-v1) * [Chatglm2-6b](https://huggingface.co/THUDM/chatglm2-6b) * [Baichuan2-7B-Base](https://huggingface.co/baichuan-inc/Baichuan2-7B-Base) \ No newline at end of file