Skip to content

Commit

Permalink
Updated README.md with latest changes (quic#105)
Browse files Browse the repository at this point in the history
Updated links of README.md

Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
  • Loading branch information
quic-amitraj authored Sep 10, 2024
1 parent 614c4cd commit 34e29f8
Show file tree
Hide file tree
Showing 3 changed files with 20 additions and 11 deletions.
17 changes: 10 additions & 7 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,12 @@

*Latest news* :fire: <br>

- [coming soon] Support for more popular [models](#models-coming-soon) and inference optimization techniques like continuous batching and speculative decoding <br>
* [06/2024] Added support for [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b)
- [coming soon] Support for more popular [models](coming_soon_models) and inference optimization technique speculative decoding <br>
- [08/2024] Added Support for inference optimization technique ```continuous batching```
- [08/2024] Added support for [Jais-adapted-70b](https://huggingface.co/inceptionai/jais-adapted-70b)
- [08/2024] Added support for [Jais-adapted-13b-chat](https://huggingface.co/inceptionai/jais-adapted-13b-chat)
- [08/2024] Added support for [Jais-adapted-7b](https://huggingface.co/inceptionai/jais-adapted-7b)
- [06/2024] Added support for [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b)
- [06/2024] Added support for [Qwen2-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct)
- [06/2024] Added support for [StarCoder2-15B](https://huggingface.co/bigcode/starcoder2-15b)
- [06/2024] Added support for [Phi3-Mini-4K-Instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct)
Expand All @@ -17,7 +21,6 @@
- [05/2024] Added support for [Mixtral-8x7B](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) & [Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1).
- [04/2024] Initial release of [efficient transformers](https://github.com/quic/efficient-transformers) for seamless inference on pre-trained LLMs.


# Overview

## Train anywhere, Infer on Qualcomm Cloud AI with a Developer-centric Toolchain
Expand Down Expand Up @@ -60,15 +63,15 @@ pip install dist/QEfficient-0.0.1.dev0-py3-none-any.whl

```

For more details about using ``QEfficient`` via Cloud AI 100 Apps SDK, visit [Linux Installation Guide](https://quic.github.io/efficient-transformers/source/Linux_installation.html)
For more details about using ``QEfficient`` via Cloud AI 100 Apps SDK, visit [Linux Installation Guide](https://quic.github.io/efficient-transformers/source/installation.html)


## Documentation

* [Quick Start Guide](https://quic.github.io/efficient-transformers/source/quick_start.html#)
* [Python API](https://quic.github.io/efficient-transformers/source/high_level_api.html)
* [Validated Models](https://quic.github.io/efficient-transformers/source/Validate.html)
* [Models coming soon](https://quic.github.io/efficient-transformers/source/Validate.html#models-coming-soon)
* [Python API](https://quic.github.io/efficient-transformers/source/hl_api.html)
* [Validated Models](https://quic.github.io/efficient-transformers/source/validate.html)
* [Models coming soon](https://quic.github.io/efficient-transformers/source/validate.html#models-coming-soon)

> Note: More details are here: https://quic.github.io/cloud-ai-sdk-pages/latest/Getting-Started/Model-Architecture-Support/Large-Language-Models/llm/
Expand Down
8 changes: 6 additions & 2 deletions docs/source/introduction.md
Original file line number Diff line number Diff line change
Expand Up @@ -22,8 +22,12 @@ For other models, there is comprehensive documentation to inspire upon the chang

***Latest news*** : <br>

- [coming soon] Support for more popular [models](coming_soon_models) and inference optimization techniques like continuous batching and speculative decoding <br>
* [06/2024] Added support for [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b)
- [coming soon] Support for more popular [models](coming_soon_models) and inference optimization technique speculative decoding <br>
- [08/2024] Added Support for inference optimization technique ```continuous batching```
- [08/2024] Added support for [Jais-adapted-70b](https://huggingface.co/inceptionai/jais-adapted-70b)
- [08/2024] Added support for [Jais-adapted-13b-chat](https://huggingface.co/inceptionai/jais-adapted-13b-chat)
- [08/2024] Added support for [Jais-adapted-7b](https://huggingface.co/inceptionai/jais-adapted-7b)
- [06/2024] Added support for [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b)
- [06/2024] Added support for [Qwen2-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct)
- [06/2024] Added support for [StarCoder2-15B](https://huggingface.co/bigcode/starcoder2-15b)
- [06/2024] Added support for [Phi3-Mini-4K-Instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct)
Expand Down
6 changes: 4 additions & 2 deletions docs/source/validate.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,10 +26,12 @@
| [Codestral-22B-v0.1](https://huggingface.co/mistralai/Codestral-22B-v0.1) | ✔️ | |
| [Falcon-40b](https://huggingface.co/tiiuae/falcon-40b) | ✔️ | |
| [GPT-J-6B](https://huggingface.co/EleutherAI/gpt-j-6b) | ✔️ | |
| [Jais-adapted-70b](https://huggingface.co/inceptionai/jais-adapted-70b) | ✔️ | ✔️ |
| [Jais-adapted-13b-chat](https://huggingface.co/inceptionai/jais-adapted-13b-chat) | ✔️ | ✔️ |
| [Jais-adapted-7b](https://huggingface.co/inceptionai/jais-adapted-7b) | ✔️ | ✔️ |

(coming_soon_models)=
# Models Coming Soon

* [Jais-13b](https://huggingface.co/core42/jais-13b)
* [Jais-30b](https://huggingface.co/core42/jais-30b-chat-v1)
* [Chatglm2-6b](https://huggingface.co/THUDM/chatglm2-6b)
* [Baichuan2-7B-Base](https://huggingface.co/baichuan-inc/Baichuan2-7B-Base)

0 comments on commit 34e29f8

Please sign in to comment.