Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
github-actions[bot] committed Aug 13, 2024
1 parent 7576ec4 commit ff767ad
Showing 1 changed file with 33 additions and 33 deletions.
66 changes: 33 additions & 33 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -32,11 +32,11 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| llama3.1 | 405b-instruct-awq-4bit-675e | [HF Link](https://huggingface.co/hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4) |
| llama3.1 | 70b-instruct-awq-4bit-28ed | [HF Link](https://huggingface.co/hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4) |
| llama3.1 | 70b-instruct-fp16-b66b | [HF Link](https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct) |
| llama3.1 | 8b-instruct-awq-4bit-5cb2 | [HF Link](https://huggingface.co/hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4) |
| llama3.1 | 8b-instruct-fp16-1c1c | [HF Link](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) |
| llama3.1 | 405b-instruct-awq-4bit-df2a | [HF Link](https://huggingface.co/hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4) |
| llama3.1 | 70b-instruct-awq-4bit-2988 | [HF Link](https://huggingface.co/hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4) |
| llama3.1 | 70b-instruct-fp16-ace8 | [HF Link](https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct) |
| llama3.1 | 8b-instruct-awq-4bit-fe8c | [HF Link](https://huggingface.co/hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4) |
| llama3.1 | 8b-instruct-fp16-2f36 | [HF Link](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) |

---

Expand All @@ -45,10 +45,10 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| llama3 | 70b-instruct-awq-4bit-9ceb | [HF Link](https://huggingface.co/casperhansen/llama-3-70b-instruct-awq) |
| llama3 | 70b-instruct-fp16-c3e4 | [HF Link](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) |
| llama3 | 8b-instruct-awq-4bit-1c94 | [HF Link](https://huggingface.co/casperhansen/llama-3-8b-instruct-awq) |
| llama3 | 8b-instruct-fp16-ba7c | [HF Link](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) |
| llama3 | 70b-instruct-awq-4bit-aebb | [HF Link](https://huggingface.co/casperhansen/llama-3-70b-instruct-awq) |
| llama3 | 70b-instruct-fp16-1315 | [HF Link](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) |
| llama3 | 8b-instruct-awq-4bit-3f34 | [HF Link](https://huggingface.co/casperhansen/llama-3-8b-instruct-awq) |
| llama3 | 8b-instruct-fp16-8f83 | [HF Link](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) |

---

Expand All @@ -57,8 +57,8 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| phi3 | 3.8b-instruct-fp16-37b9 | [HF Link](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) |
| phi3 | 3.8b-instruct-ggml-q4-cf55 | [HF Link](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf) |
| phi3 | 3.8b-instruct-fp16-166c | [HF Link](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) |
| phi3 | 3.8b-instruct-ggml-q4-8789 | [HF Link](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf) |

---

Expand All @@ -67,8 +67,8 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| mistral | 7b-instruct-awq-4bit-4406 | [HF Link](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-AWQ) |
| mistral | 7b-instruct-fp16-e3bd | [HF Link](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) |
| mistral | 7b-instruct-awq-4bit-01cd | [HF Link](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-AWQ) |
| mistral | 7b-instruct-fp16-e1cd | [HF Link](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) |

---

Expand All @@ -77,8 +77,8 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| gemma2 | 27b-instruct-fp16-9799 | [HF Link](https://huggingface.co/google/gemma-2-27b-it) |
| gemma2 | 9b-instruct-fp16-cb2b | [HF Link](https://huggingface.co/google/gemma-2-9b-it) |
| gemma2 | 27b-instruct-fp16-6b83 | [HF Link](https://huggingface.co/google/gemma-2-27b-it) |
| gemma2 | 9b-instruct-fp16-6e86 | [HF Link](https://huggingface.co/google/gemma-2-9b-it) |

---

Expand All @@ -87,13 +87,13 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| qwen2 | 0.5b-instruct-fp16-bca0 | [HF Link](https://huggingface.co/Qwen/Qwen2-0.5B-Instruct) |
| qwen2 | 1.5b-instruct-fp16-df66 | [HF Link](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct) |
| qwen2 | 57b-a14b-instruct-fp16-b847 | [HF Link](https://huggingface.co/Qwen/Qwen2-57B-A14B-Instruct) |
| qwen2 | 72b-instruct-awq-4bit-60b1 | [HF Link](https://huggingface.co/Qwen/Qwen2-72B-Instruct-AWQ) |
| qwen2 | 72b-instruct-fp16-ee8e | [HF Link](https://huggingface.co/Qwen/Qwen2-72B-Instruct) |
| qwen2 | 7b-instruct-awq-4bit-02f4 | [HF Link](https://huggingface.co/Qwen/Qwen2-7B-Instruct-AWQ) |
| qwen2 | 7b-instruct-fp16-761c | [HF Link](https://huggingface.co/Qwen/Qwen2-7B-Instruct) |
| qwen2 | 0.5b-instruct-fp16-33df | [HF Link](https://huggingface.co/Qwen/Qwen2-0.5B-Instruct) |
| qwen2 | 1.5b-instruct-fp16-7cda | [HF Link](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct) |
| qwen2 | 57b-a14b-instruct-fp16-365f | [HF Link](https://huggingface.co/Qwen/Qwen2-57B-A14B-Instruct) |
| qwen2 | 72b-instruct-awq-4bit-33fa | [HF Link](https://huggingface.co/Qwen/Qwen2-72B-Instruct-AWQ) |
| qwen2 | 72b-instruct-fp16-8cb4 | [HF Link](https://huggingface.co/Qwen/Qwen2-72B-Instruct) |
| qwen2 | 7b-instruct-awq-4bit-14aa | [HF Link](https://huggingface.co/Qwen/Qwen2-7B-Instruct-AWQ) |
| qwen2 | 7b-instruct-fp16-bbf2 | [HF Link](https://huggingface.co/Qwen/Qwen2-7B-Instruct) |

---

Expand All @@ -102,9 +102,9 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| gemma | 2b-instruct-fp16-f6ee | [HF Link](https://huggingface.co/google/gemma-2b-it) |
| gemma | 7b-instruct-awq-4bit-bdb5 | [HF Link](https://huggingface.co/casperhansen/gemma-7b-it-awq) |
| gemma | 7b-instruct-fp16-35e0 | [HF Link](https://huggingface.co/google/gemma-7b-it) |
| gemma | 2b-instruct-fp16-6ee7 | [HF Link](https://huggingface.co/google/gemma-2b-it) |
| gemma | 7b-instruct-awq-4bit-df0b | [HF Link](https://huggingface.co/casperhansen/gemma-7b-it-awq) |
| gemma | 7b-instruct-fp16-2297 | [HF Link](https://huggingface.co/google/gemma-7b-it) |

---

Expand All @@ -113,10 +113,10 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| llama2 | 13b-chat-fp16-a846 | [HF Link](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf) |
| llama2 | 70b-chat-fp16-fcef | [HF Link](https://huggingface.co/meta-llama/Llama-2-70b-chat-hf) |
| llama2 | 7b-chat-awq-4bit-753b | [HF Link](https://huggingface.co/TheBloke/Llama-2-7B-Chat-AWQ) |
| llama2 | 7b-chat-fp16-dc53 | [HF Link](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) |
| llama2 | 13b-chat-fp16-ef61 | [HF Link](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf) |
| llama2 | 70b-chat-fp16-16a0 | [HF Link](https://huggingface.co/meta-llama/Llama-2-70b-chat-hf) |
| llama2 | 7b-chat-awq-4bit-4f93 | [HF Link](https://huggingface.co/TheBloke/Llama-2-7B-Chat-AWQ) |
| llama2 | 7b-chat-fp16-21b9 | [HF Link](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) |

---

Expand All @@ -125,8 +125,8 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| mixtral | 8x7b-instruct-v0.1-awq-4bit-7bae | [HF Link](https://huggingface.co/casperhansen/mixtral-instruct-awq) |
| mixtral | 8x7b-instruct-v0.1-fp16-1c82 | [HF Link](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) |
| mixtral | 8x7b-instruct-v0.1-awq-4bit-06fd | [HF Link](https://huggingface.co/casperhansen/mixtral-instruct-awq) |
| mixtral | 8x7b-instruct-v0.1-fp16-e289 | [HF Link](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) |

---

Expand All @@ -135,8 +135,8 @@ openllm repo add nightly https://github.com/bentoml/openllm-models@nightly

| Model | Version | Huggingface Link |
| --- | --- | --- |
| mistral-large | 123b-instruct-awq-4bit-c380 | [HF Link](https://huggingface.co/casperhansen/mistral-large-instruct-2407-awq) |
| mistral-large | 123b-instruct-fp16-a203 | [HF Link](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407) |
| mistral-large | 123b-instruct-awq-4bit-1d37 | [HF Link](https://huggingface.co/casperhansen/mistral-large-instruct-2407-awq) |
| mistral-large | 123b-instruct-fp16-5c96 | [HF Link](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407) |

---

0 comments on commit ff767ad

Please sign in to comment.