Skip to content

Commit

Permalink
fix: comments
Browse files Browse the repository at this point in the history
  • Loading branch information
xianml committed Oct 8, 2024
1 parent 862f17f commit 0fada35
Show file tree
Hide file tree
Showing 106 changed files with 38 additions and 44 deletions.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

2 changes: 1 addition & 1 deletion src/llamacpp-chat/bentofile.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
service: "service:LlamaCppChat"
labels:
source: https://github.com/bentoml/openllm-models-feed/tree/main/src/llamacpp-chat
source: https://github.com/bentoml/openllm-models/tree/main/src/llamacpp-chat
include:
- "*.py"
- "*.yaml"
Expand Down
5 changes: 3 additions & 2 deletions src/recipe.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -769,6 +769,7 @@
engine_config:
model: microsoft/Phi-3-mini-4k-instruct-gguf
max_model_len: 2048
filename: "Phi-3-mini-4k-instruct-q4.gguf"
extra_labels:
openllm_alias: 3.8b-ggml-q4,3.8b-mini-instruct-4k-ggml-q4
model_name: microsoft/Phi-3-mini-4k-instruct-gguf
Expand Down Expand Up @@ -830,12 +831,12 @@
max_model_len: 2048
filename: "Llama-3.2-1B-Instruct-F16.gguf"
extra_labels:
openllm_alias: llama3.2,1b-instruct-fp16-ggml-mac
openllm_alias: llama3.2,1b-instruct-fp16-ggml-darwin
model_name: unsloth/Llama-3.2-1B-Instruct-GGUF
platforms: macos
extra_envs:
- name: CMAKE_ARGS
value: "-DGGML_CUDA=on"
value: "-DGGML_METAL=on"
'llama3.2:1b-instruct-fp16-ggml-linux':
project: llamacpp-chat
service_config:
Expand Down

0 comments on commit 0fada35

Please sign in to comment.