COMMITS
July 5, 2025
A
fix(ci): Remove macos-13 builds to fix cross compilation error
Andrei Betlen committed
A
fix(ci): Add git to package list
Andrei Betlen committed
A
fix(ci): Update cuda build action to use ubuntu 22.04
Andrei Betlen committed
A
fix: Update reference to in Llama.embed. Closes #2037
Andrei Betlen committed
July 3, 2025
A
chore: Bump version
Andrei Betlen committed
A
docs: Add Qwen2.5-VL to README
Andrei Betlen committed
A
fix: Use num_threads from llama model for mtmd
Andrei Betlen committed
A
feat: Add support for new mtmd api, add Qwen2.5-VL chat handler
Andrei Betlen committed
July 1, 2025
A
fix: Fix missing deprecated symbols on windows with missing LLAMA_API prefix in header file
Andrei Betlen committed
A
fix(minor): Fix type hint for older versions of python
Andrei Betlen committed
A
misc: Fix support for new parameters, deprecate rpc_servers parameter
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
May 8, 2025
A
hotfix: Disable curl support
Andrei Betlen committed
A
chore: Bump version
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
April 11, 2025
A
feat: Update llama.cpp
Andrei Betlen committed
March 12, 2025
A
chore: Bump version
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
January 29, 2025
A
chore: Bump version
Andrei Betlen committed
S
fix: error showing time spent in llama perf context print (#1898)
Shaka Huang committed
O
fix(ci): Fix the CUDA workflow (#1894)
oobabooga committed
A
feat: Update llama.cpp
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
January 8, 2025
A
chore: Bump version
Andrei Betlen committed
G
fix: streaming resource lock (#1879)
Graeme Power committed
A
feat: Update llama.cpp
Andrei Betlen committed
December 30, 2024
A
feat: Update llama.cpp
Andrei Betlen committed
December 19, 2024
A
feat: Update llama.cpp
Andrei Betlen committed
December 9, 2024
A
chore: Bump version
Andrei Betlen committed