COMMITS
/ llama_cpp/server/__main__.py December 22, 2023
D
[Feat] Multi model support (#931)
Dave committed
December 12, 2023
R
Add support for running the server with SSL (#994)
Radoslav Gerganov committed
September 15, 2023
A
Remove confusing helpstring from server cli args. Closes #719
Andrei Betlen committed
September 14, 2023
A
Fix boolean env vars and cli arguments
Andrei Betlen committed
A
Fix boolean cli flags
Andrei Betlen committed
A
Fix tensor_split cli option
Andrei Betlen committed
September 13, 2023
A
Fix issue with Literal and Optional cli arguments not working. Closes #702
Andrei Betlen committed
July 9, 2023
A
bugfix: pydantic settings missing / changed fields
Andrei Betlen committed
July 8, 2023
A
Upgrade fastapi to 0.100.0 and pydantic v2
Andrei Betlen committed
June 11, 2023
G
correction to add back environment variable support <3 docker
Gabor committed
June 10, 2023
G
fixes abetlen/llama-cpp-python #358
Gabor committed
May 8, 2023
A
Bugfix: not falling back to environment variables when default is value is set.
Andrei Betlen committed
A
Show default value when --help is called
Andrei Betlen committed
A
Use environment variable if parsed cli arg is None
Andrei Betlen committed
May 5, 2023
A
Add cli options to server. Closes #37
Andrei Betlen committed
May 2, 2023
A
Refactor server to use factory
Andrei Betlen committed
April 29, 2023
L
llama_cpp server: slight refactor to init_llama function
Lucas Doyle committed
L
llama_cpp server: app is now importable, still runnable as a module
Lucas Doyle committed
April 25, 2023
A
Update n_batch for server
Andrei Betlen committed
April 19, 2023
A
Add use_mmap flag to server
Andrei Betlen committed
April 15, 2023
A
Add experimental cache
Andrei Betlen committed
April 14, 2023
A
Fix completion request
Andrei Betlen committed
April 13, 2023
A
Formatting
Andrei Betlen committed
April 12, 2023
A
Enable logprobs on completion endpoint
Andrei Betlen committed
A
Fix threading bug. Closes #62
Andrei Betlen committed
A
Implement logprobs parameter for text completion. Closes #2
Andrei Betlen committed
April 11, 2023
A
Remove async from function signature to avoid blocking the server
Andrei Betlen committed
April 8, 2023
A
Formatting
Andrei Betlen committed
April 7, 2023
A
Safer calculation of default n_threads
Andrei Betlen committed
A
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
Andrei Betlen committed
A
Handle prompt list
Andrei Betlen committed
April 5, 2023
M
Set n_batch to default values and reduce thread count:
MillionthOdin16 committed
M
A
Add server as a subpackage
Andrei Betlen committed