Skip to content

Actions: ggerganov/llama.cpp

Server

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
8,977 workflow runs
8,977 workflow runs

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

GGUF: C++ refactor, backend support, misc fixes
Server #9481: Pull request #11030 synchronize by JohannesGaessler
January 3, 2025 14:47 In progress JohannesGaessler:gguf-refactor-7
January 3, 2025 14:47 In progress
[GGML][RPC] Support for models with non-512-aligned tensors over RPC.
Server #9479: Pull request #11047 synchronize by matt23654
January 3, 2025 14:39 Action required matt23654:rpc-experimental
January 3, 2025 14:39 Action required
llama : add support for Cohere2ForCausalLM
Server #9477: Pull request #10900 synchronize by dranger003
January 3, 2025 13:43 1h 0m 16s dranger003:cohere2
January 3, 2025 13:43 1h 0m 16s
llama : remove notion of CLS token
Server #9476: Pull request #11064 opened by ggerganov
January 3, 2025 12:54 1h 39m 11s gg/llama-refactor-5
January 3, 2025 12:54 1h 39m 11s
llama : update llama_model API names
Server #9475: Pull request #11063 opened by ggerganov
January 3, 2025 12:44 1h 35m 1s gg/llama-refactor-4
January 3, 2025 12:44 1h 35m 1s
llama : use LLAMA_TOKEN_NULL
Server #9474: Pull request #11062 opened by ggerganov
January 3, 2025 12:28 1h 29m 11s gg/llama-refactor-3
January 3, 2025 12:28 1h 29m 11s
common : disable KV cache shifting automatically for unsupported mode…
Server #9473: Commit 4b0c638 pushed by ggerganov
January 3, 2025 12:13 42m 19s master
January 3, 2025 12:13 42m 19s
llama : avoid hardcoded QK_K constant
Server #9472: Pull request #11061 opened by ggerganov
January 3, 2025 12:08 23m 16s gg/llama-refactor-2
January 3, 2025 12:08 23m 16s
llama : use _impl suffix instead of _internal
Server #9470: Pull request #11060 opened by ggerganov
January 3, 2025 11:55 9m 20s gg/llama-refactor-1
January 3, 2025 11:55 9m 20s
tokenize : escape the prompt
Server #9468: Pull request #11058 opened by ggerganov
January 3, 2025 10:07 6m 0s gg/tokenize-escape
January 3, 2025 10:07 6m 0s
metal : avoid uint (#11019)
Server #9467: Commit e7da954 pushed by ggerganov
January 3, 2025 09:26 15m 2s master
January 3, 2025 09:26 15m 2s
llama : rename missed batch params/vars to ubatch
Server #9464: Pull request #10059 synchronize by danbev
January 3, 2025 06:54 5m 41s danbev:ubatch-leftovers
January 3, 2025 06:54 5m 41s
Add support for DeepSeek V3
Server #9461: Pull request #11049 synchronize by fairydreaming
January 2, 2025 20:40 5m 33s fairydreaming:deepseek-v3
January 2, 2025 20:40 5m 33s
Add support for DeepSeek V3
Server #9460: Pull request #11049 opened by fairydreaming
January 2, 2025 20:10 5m 28s fairydreaming:deepseek-v3
January 2, 2025 20:10 5m 28s
llama : refactor src/llama.cpp
Server #9459: Pull request #10902 synchronize by ggerganov
January 2, 2025 19:58 4m 50s gg/llama-refactor-0
January 2, 2025 19:58 4m 50s
llama : refactor src/llama.cpp
Server #9458: Pull request #10902 synchronize by ggerganov
January 2, 2025 19:40 6m 10s gg/llama-refactor-0
January 2, 2025 19:40 6m 10s
llama : refactor src/llama.cpp
Server #9457: Pull request #10902 synchronize by ggerganov
January 2, 2025 19:37 2m 19s gg/llama-refactor-0
January 2, 2025 19:37 2m 19s