COMMITS
July 3, 2025
A
chore: Bump version
Andrei Betlen committed
A
docs: Add Qwen2.5-VL to README
Andrei Betlen committed
A
fix: Use num_threads from llama model for mtmd
Andrei Betlen committed
A
feat: Add support for new mtmd api, add Qwen2.5-VL chat handler
Andrei Betlen committed
July 1, 2025
A
fix: Fix missing deprecated symbols on windows with missing LLAMA_API prefix in header file
Andrei Betlen committed
A
fix(minor): Fix type hint for older versions of python
Andrei Betlen committed
A
misc: Fix support for new parameters, deprecate rpc_servers parameter
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
May 8, 2025
A
hotfix: Disable curl support
Andrei Betlen committed
A
chore: Bump version
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
April 11, 2025
A
feat: Update llama.cpp
Andrei Betlen committed
March 12, 2025
A
chore: Bump version
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
January 29, 2025
A
chore: Bump version
Andrei Betlen committed
S
fix: error showing time spent in llama perf context print (#1898)
Shaka Huang committed
O
fix(ci): Fix the CUDA workflow (#1894)
oobabooga committed
A
feat: Update llama.cpp
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
January 8, 2025
A
chore: Bump version
Andrei Betlen committed
G
fix: streaming resource lock (#1879)
Graeme Power committed
A
feat: Update llama.cpp
Andrei Betlen committed
December 30, 2024
A
feat: Update llama.cpp
Andrei Betlen committed
December 19, 2024
A
feat: Update llama.cpp
Andrei Betlen committed
December 9, 2024
A
chore: Bump version
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
G
fix: add missing await statements for async exit_stack handling (#1858)
Graeme Power committed
A
fix(ci): update macos runner image to non-deprecated version
Andrei Betlen committed
A
chore: Bump version
Andrei Betlen committed