COMMITS
/ llama_cpp/llama.py October 15, 2023
E
Add validation for tensor_split size exceeding LLAMA_MAX_DEVICES (#820)
Eric Liu committed
September 30, 2023
A
Fix logits_all bug
Andrei Betlen committed
A
Fix bug in embedding
Andrei Betlen committed
September 29, 2023
A
Configurable Chat Formats (#711)
Andrei committed
J
Fix rope scaling defaults (#767)
Josh XT committed
A
Update llama.cpp
Andrei Betlen committed
September 18, 2023
A
Update llama.cpp
Andrei Betlen committed
September 14, 2023
A
A
Reorder init params to match llama.cpp order
Andrei Betlen committed
A
Explicitly make all init params other than model_path into keyword only params
Andrei Betlen committed
A
Add kwargs to init to catch extra params
Andrei Betlen committed
A
remove print
Andrei Betlen committed
A
Convert missed llama.cpp constants into standard python types
Andrei Betlen committed
A
Fix tensor_split cli option
Andrei Betlen committed
September 12, 2023
A
Merge branch 'main' into v0.2-wip
Andrei Betlen committed
August 29, 2023
A
cjk pr minor cleanup
Andrei Betlen committed
A
Merge pull request #309 from MeouSker77/fix-CJK
Andrei committed
August 27, 2023
A
Update llama.cpp
Andrei Betlen committed
A
Update llama.cpp
Andrei Betlen committed
August 25, 2023
A
Merge branch 'main' into v0.2-wip
Andrei Betlen committed
A
Use _with_model variants for tokenization
Andrei Betlen committed
A
Strip leading space when de-tokenizing.
Andrei Betlen committed
August 24, 2023
A
Remove deprecated params
Andrei Betlen committed
A
Merge branch 'main' into v0.2-wip
Andrei Betlen committed
A
Update llama.cpp
Andrei Betlen committed
August 15, 2023
A
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
Andrei Betlen committed
A
Remove unnused import
Andrei Betlen committed
August 13, 2023
B
make n_gpu_layers=-1 offload all layers
Billy Cao committed
August 12, 2023
B
Add doc string for n_gpu_layers argument
Billy Cao committed
August 9, 2023
M
fix CJK output again
MeouSker77 committed
August 8, 2023
A
Move grammar to function call argument
Andrei Betlen committed
A
Merge branch 'main' into c0sogi/main
Andrei Betlen committed
A
Add mul_mat_q option
Andrei Betlen committed
August 7, 2023
C
reset grammar for every generation
c0sogi committed
August 6, 2023
C
Added grammar based sampling
c0sogi committed
July 28, 2023
A
Suppress llama.cpp output when loading model.
Andrei Betlen committed
A
fix: annoying bug where attribute exceptions were droining out file not found exceptions
Andrei Betlen committed
July 25, 2023
S
Change tensor_split from array to pointer
Shouyi Wang committed
July 24, 2023
A
Merge branch 'main' into v0.2-wip
Andrei Betlen committed
A
Add temporary rms_norm_eps parameter
Andrei Betlen committed
A
Add rms_eps_norm
Andrei Betlen committed
B
add support for llama2 70b
bretello committed
July 20, 2023
A
Merge branch 'main' into v0.2-wip
Andrei Betlen committed
A
Merge pull request #481 from c0sogi/main
Andrei committed
C
Now the last token sent when `stream=True`
Carlos Tejada committed
July 19, 2023
A
Add functions parameters
Andrei Betlen committed
July 18, 2023
A
Use numpy arrays for logits_processors and stopping_criteria. Closes #491
Andrei Betlen committed
July 16, 2023
C
Added `RouteErrorHandler` for server
c0sogi committed