COMMITS
May 5, 2024
O
feat(ci): Add docker checks and check deps more frequently (#1426)
Olivier DEBAUCHE committed
A
feat: Update llama.cpp
Andrei Betlen committed
May 4, 2024
N
J
feat: Implement streaming for Functionary v2 + Bug fixes (#1419)
Jeffrey Fong committed
May 3, 2024
A
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
Andrei Betlen committed
A
fix: Use memmove to copy str_value kv_override. Closes #1417
Andrei Betlen committed
A
feat(server): Remove temperature bounds checks for server. Closes #1384
Andrei Betlen committed
D
fix(server): Propagate `flash_attn` to model load. (#1424)
Daniel Thuerck committed
May 2, 2024
A
chore: Bump version
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
A
feat: Add llama-3-vision-alpha chat format
Andrei Betlen committed
April 30, 2024
A
A
feat: Update llama.cpp
Andrei Betlen committed
A
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
Andrei Betlen committed
A
J
fix: UTF-8 handling with grammars (#1415)
Jonathan Soma committed
A
docs: Change all examples from interpreter style to script style.
Andrei Betlen committed
A
docs: Update README.md
Andrei Betlen committed
A
chore: Bump version
Andrei Betlen committed
A
fix: wrong parameter for flash attention in pickle __getstate__
Andrei Betlen committed
A
feat: Add option to enable `flash_attn` to Lllama params and ModelSettings
Andrei Betlen committed
A
feat: Update llama.cpp
Andrei Betlen committed
O
fix(ci): Fix build-and-release.yaml (#1413)
Olivier DEBAUCHE committed
A
docs: Update README to include CUDA 12.4 wheels
Andrei Betlen committed
A
chore: Bump version
Andrei Betlen committed
A
fix: Ensure image renders before text in chat formats regardless of message content order.
Andrei Betlen committed
A
A
chore: Bump version
Andrei Betlen committed
A
A
feat: Update llama.cpp
Andrei Betlen committed