Skip to content

Actions: ggerganov/llama.cpp

Publish Docker image

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
9,249 workflow run results
9,249 workflow run results

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

Detokenizer fixes
Publish Docker image #14073: Pull request #8039 synchronize by jaime-m-p
June 24, 2024 18:56 4s jaime-m-p:detokenizer
June 24, 2024 18:56 4s
Streamline embeddings from "non-embedding" models
Publish Docker image #14072: Pull request #8087 synchronize by iamlemec
June 24, 2024 16:54 3h 3m 17s iamlemec:attention-type
June 24, 2024 16:54 3h 3m 17s
gfx1010 optimizations
Publish Docker image #14071: Pull request #8085 synchronize by daniandtheweb
June 24, 2024 16:49 2h 38m 40s daniandtheweb:gfx1010_optimizations
June 24, 2024 16:49 2h 38m 40s
CUDA: use MMQ instead of cuBLAS by default (#8075)
Publish Docker image #14070: Commit a818f30 pushed by JohannesGaessler
June 24, 2024 15:43 3h 34m 15s master
June 24, 2024 15:43 3h 34m 15s
nix: update flake.lock
Publish Docker image #14069: Pull request #8071 synchronize by philiptaron
June 24, 2024 15:26 2h 27m 25s update_flake_lock_action
June 24, 2024 15:26 2h 27m 25s
gfx1010 optimizations
Publish Docker image #14068: Pull request #8085 synchronize by daniandtheweb
June 24, 2024 14:47 2h 2m 18s daniandtheweb:gfx1010_optimizations
June 24, 2024 14:47 2h 2m 18s
llama : reorganize source code + improve CMake
Publish Docker image #14067: Pull request #8006 synchronize by ggerganov
June 24, 2024 13:54 3s gg/reorganize-project
June 24, 2024 13:54 3s
llama : reorganize source code + improve CMake
Publish Docker image #14066: Pull request #8006 synchronize by ggerganov
June 24, 2024 13:51 4s gg/reorganize-project
June 24, 2024 13:51 4s
[SYCL] Re-enabled mul_mat_batched_sycl
Publish Docker image #14065: Pull request #8095 opened by airMeng
June 24, 2024 12:57 2h 48m 55s sycl-mul-mat-batched
June 24, 2024 12:57 2h 48m 55s
CUDA: use MMQ instead of cuBLAS by default
Publish Docker image #14064: Pull request #8075 synchronize by JohannesGaessler
June 24, 2024 12:35 2h 28m 9s JohannesGaessler:cuda-mmq-default
June 24, 2024 12:35 2h 28m 9s
gguf-py : fix tensor groups for encoder-decoder models in gguf-dump.p…
Publish Docker image #14063: Commit d62e4aa pushed by fairydreaming
June 24, 2024 12:13 2h 4m 26s master
June 24, 2024 12:13 2h 4m 26s
llama : return nullptr from llama_grammar_init
Publish Docker image #14062: Pull request #8093 opened by danbev
June 24, 2024 11:25 1h 48m 21s danbev:grammar-init-return-null
June 24, 2024 11:25 1h 48m 21s
Fix tensor groups for encoder-decoder models in gguf-dump.py
Publish Docker image #14061: Pull request #8090 synchronize by mofosyne
June 24, 2024 10:58 1h 12m 57s fairydreaming:gguf-dump-grouping-fix
June 24, 2024 10:58 1h 12m 57s
CUDA: optimize MMQ int8 tensor core performance (#8062)
Publish Docker image #14059: Commit 9a590c8 pushed by JohannesGaessler
June 24, 2024 10:41 1h 48m 52s master
June 24, 2024 10:41 1h 48m 52s
CUDA: optimize MMQ int8 tensor core performance
Publish Docker image #14058: Pull request #8062 synchronize by JohannesGaessler
June 24, 2024 09:44 1h 50m 37s JohannesGaessler:cuda-mmq-2xa-3
June 24, 2024 09:44 1h 50m 37s
Option to split during conversion (#6942)
Publish Docker image #14057: Commit 52fc870 pushed by mofosyne
June 24, 2024 09:42 2h 15m 27s master
June 24, 2024 09:42 2h 15m 27s
Add chat template support for llama-cli
Publish Docker image #14056: Pull request #8068 synchronize by ngxson
June 24, 2024 09:07 1h 36m 33s ngxson:xsn/main_chat_template_2
June 24, 2024 09:07 1h 36m 33s
Add chat template support for llama-cli
Publish Docker image #14055: Pull request #8068 synchronize by ngxson
June 24, 2024 09:00 7m 12s ngxson:xsn/main_chat_template_2
June 24, 2024 09:00 7m 12s
Add chat template support for llama-cli
Publish Docker image #14054: Pull request #8068 synchronize by ngxson
June 24, 2024 08:57 2m 27s ngxson:xsn/main_chat_template_2
June 24, 2024 08:57 2m 27s
Add chat template support for llama-cli
Publish Docker image #14053: Pull request #8068 synchronize by ngxson
June 24, 2024 08:57 1m 0s ngxson:xsn/main_chat_template_2
June 24, 2024 08:57 1m 0s
Add chat template support for llama-cli
Publish Docker image #14052: Pull request #8068 synchronize by ngxson
June 24, 2024 08:52 4m 45s ngxson:xsn/main_chat_template_2
June 24, 2024 08:52 4m 45s
llama : reorganize source code + improve CMake
Publish Docker image #14049: Pull request #8006 synchronize by ggerganov
June 24, 2024 08:21 3s gg/reorganize-project
June 24, 2024 08:21 3s