Skip to content

Actions: ggerganov/llama.cpp

Publish Docker image

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
9,963 workflow run results
9,963 workflow run results

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

Detokenizer fixes
Publish Docker image #14021: Pull request #8039 synchronize by jaime-m-p
June 23, 2024 19:12 3s jaime-m-p:detokenizer
June 23, 2024 19:12 3s
ggml : remove ggml_task_type and GGML_PERF
Publish Docker image #14020: Pull request #8017 synchronize by slaren
June 23, 2024 19:08 1h 28m 58s sl/remove-task-type
June 23, 2024 19:08 1h 28m 58s
disable publishing the full-rocm docker image
Publish Docker image #14019: Pull request #8083 opened by slaren
June 23, 2024 18:34 1h 43m 36s sl/disable-rocm-full-docker
June 23, 2024 18:34 1h 43m 36s
Detokenizer fixes
Publish Docker image #14018: Pull request #8039 synchronize by jaime-m-p
June 23, 2024 18:33 4s jaime-m-p:detokenizer
June 23, 2024 18:33 4s
llama : add support for BitnetForCausalLM (#7931)
Publish Docker image #14017: Commit e112b61 pushed by ggerganov
June 23, 2024 18:28 1h 12m 35s master
June 23, 2024 18:28 1h 12m 35s
ggml : remove ggml_task_type and GGML_PERF
Publish Docker image #14016: Pull request #8017 synchronize by slaren
June 23, 2024 18:11 33m 50s sl/remove-task-type
June 23, 2024 18:11 33m 50s
gfx908 optimizations
Publish Docker image #14015: Pull request #8082 opened by IMbackK
June 23, 2024 17:50 54m 10s IMbackK:gfx908_small_v2
June 23, 2024 17:50 54m 10s
Model conversion support for T5 and FLAN-T5 model variants
Publish Docker image #14014: Pull request #8055 synchronize by fairydreaming
June 23, 2024 17:36 51m 51s fairydreaming:t5-clean
June 23, 2024 17:36 51m 51s
ggml : remove ggml_task_type and GGML_PERF
Publish Docker image #14013: Pull request #8017 synchronize by slaren
June 23, 2024 17:30 31m 38s sl/remove-task-type
June 23, 2024 17:30 31m 38s
ggml : remove ggml_task_type and GGML_PERF
Publish Docker image #14012: Pull request #8017 synchronize by slaren
June 23, 2024 17:28 2m 10s sl/remove-task-type
June 23, 2024 17:28 2m 10s
ggml : remove ggml_task_type and GGML_PERF
Publish Docker image #14011: Pull request #8017 synchronize by slaren
June 23, 2024 17:23 5m 12s sl/remove-task-type
June 23, 2024 17:23 5m 12s
Add healthchecks to llama-server containers
Publish Docker image #14010: Pull request #8081 synchronize by codearranger
June 23, 2024 16:51 1h 48m 51s codearranger:healthcheck
June 23, 2024 16:51 1h 48m 51s
Model conversion support for T5 and FLAN-T5 model variants
Publish Docker image #14008: Pull request #8055 synchronize by fairydreaming
June 23, 2024 16:49 30m 56s fairydreaming:t5-clean
June 23, 2024 16:49 30m 56s
Add support for BitnetForCausalLM (new model / new datatype)
Publish Docker image #14006: Pull request #7931 synchronize by Eddie-Wang1120
June 23, 2024 15:58 49m 10s Eddie-Wang1120:bitnet
June 23, 2024 15:58 49m 10s
llama : reorganize source code + improve CMake
Publish Docker image #14003: Pull request #8006 synchronize by ggerganov
June 23, 2024 15:11 4s gg/reorganize-project
June 23, 2024 15:11 4s
server : fix JSON-Scheme typo (#7975)
Publish Docker image #14002: Commit 6a2f298 pushed by HanClinto
June 23, 2024 15:03 6h 5m 8s master
June 23, 2024 15:03 6h 5m 8s
Model conversion support for T5 and FLAN-T5 model variants
Publish Docker image #14001: Pull request #8055 synchronize by fairydreaming
June 23, 2024 13:59 46m 59s fairydreaming:t5-clean
June 23, 2024 13:59 46m 59s
Fix typo in llama_set_embeddings comment (#8077)
Publish Docker image #14000: Commit 11318d9 pushed by Galunid
June 23, 2024 13:39 34m 44s master
June 23, 2024 13:39 34m 44s
llama : fix typo in llama_set_embeddings comment
Publish Docker image #13997: Pull request #8077 opened by danbev
June 23, 2024 12:42 35m 36s danbev:llama_set_embeddings_comment
June 23, 2024 12:42 35m 36s
fix CI failures (#8066)
Publish Docker image #13996: Commit b6b9a8e pushed by slaren
June 23, 2024 11:14 25m 21s master
June 23, 2024 11:14 25m 21s
CUDA: use MMQ instead of cuBLAS by default
Publish Docker image #13995: Pull request #8075 opened by JohannesGaessler
June 23, 2024 10:47 40m 53s JohannesGaessler:cuda-mmq-default
June 23, 2024 10:47 40m 53s
ggml-cuda: Adding support for unified memory
Publish Docker image #13994: Pull request #8035 synchronize by matteoserva
June 23, 2024 10:41 7m 53s matteoserva:master
June 23, 2024 10:41 7m 53s
gguf-hash: model wide and per tensor hashing using xxhash and sha1
Publish Docker image #13993: Pull request #8048 synchronize by mofosyne
June 23, 2024 09:47 47m 15s mofosyne:gguf-hash
June 23, 2024 09:47 47m 15s