Commit Graph

  • 1848afebe0 Added dependencies. Maximilian-Winter 2023-05-31 22:41:35 +02:00
  • 29f9c9cca3 Added both LlamaChache classes Disk and RAM. Maximilian-Winter 2023-05-31 22:33:56 +02:00
  • 217d78320f Added paramterised search and d/l for Hugging Face. Updated README.md Gary Mulder 2023-05-31 16:00:31 +00:00
  • 483b6ba53a Updated README.md instructions on how to use *_simple/Dockerfiles Gary Mulder 2023-05-31 15:16:32 +00:00
  • 9ea7a379d3 Merge branch 'abetlen:main' into main Maximilian Winter 2023-05-31 12:55:51 +02:00
  • 232880cbc6 Merge pull request #297 from gjmulder/update-issue-tmpl-llamacpp-repro Andrei 2023-05-30 14:48:14 -04:00
  • 9dd8cf3472 Update bug_report.md Gary Mulder 2023-05-30 08:20:34 +01:00
  • cb0bcdbbb7 Bump version v0.1.56 Andrei Betlen 2023-05-30 03:07:36 -04:00
  • f4ff8a03c4 Add numpy dependency to pyproject Andrei Betlen 2023-05-30 03:06:57 -04:00
  • 828f9ec015 Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-05-29 21:39:40 -04:00
  • b1daf568e3 Update changelog Andrei Betlen 2023-05-29 21:39:19 -04:00
  • 825912a3a5 Merge pull request #293 from abetlen/dependabot/pip/typing-extensions-4.6.2 Andrei 2023-05-29 21:28:27 -04:00
  • f4fc126a00 Bump typing-extensions from 4.5.0 to 4.6.2 dependabot[bot] 2023-05-30 01:09:10 +00:00
  • 213549787a Merge pull request #294 from abetlen/dependabot/pip/uvicorn-0.22.0 Andrei 2023-05-29 21:07:41 -04:00
  • 8dfb0816df Bump uvicorn from 0.21.1 to 0.22.0 dependabot[bot] 2023-05-30 01:04:25 +00:00
  • f5d136d0aa Merge pull request #295 from abetlen/dependabot/pip/mkdocs-material-9.1.15 Andrei 2023-05-29 21:02:27 -04:00
  • 49fe9395a1 Merge pull request #277 from abetlen/add-numpy-support Andrei 2023-05-29 20:59:30 -04:00
  • 38b918503f Bump mkdocs-material from 9.1.14 to 9.1.15 dependabot[bot] 2023-05-30 00:40:52 +00:00
  • b61b016585 Merge pull request #296 from abetlen/dependabot/pip/scikit-build-0.17.5 Andrei 2023-05-29 20:39:52 -04:00
  • fa79484a29 Bump scikit-build from 0.13.0 to 0.17.5 dependabot[bot] 2023-05-30 00:29:33 +00:00
  • 41e2184431 Merge pull request #292 from abetlen/dependabot/pip/mkdocstrings-0.22.0 Andrei 2023-05-29 20:28:02 -04:00
  • 62ac7c3761 Bump mkdocstrings from 0.21.2 to 0.22.0 dependabot[bot] 2023-05-29 21:03:33 +00:00
  • 719c3eae0a Diskcache implementation for llama state. Maximilian-Winter 2023-05-28 15:56:38 +02:00
  • 80066f0b80 Use async routes Andrei Betlen 2023-05-27 09:12:58 -04:00
  • c2b59a5f59 Import unnused import Andrei Betlen 2023-05-26 22:59:29 -04:00
  • 8f2b4456ad Format Andrei Betlen 2023-05-26 22:04:31 -04:00
  • 84e313bd6e Align dtype to match c structs Andrei Betlen 2023-05-26 22:02:16 -04:00
  • b0b154cfa6 Add changelog message for numpy Andrei Betlen 2023-05-26 20:26:08 -04:00
  • 66bcb8d70d Merge branch 'main' into add-numpy-support Andrei Betlen 2023-05-26 20:25:03 -04:00
  • 8f35bddd7e Fix stop sequence performance bug. Andrei Betlen 2023-05-26 20:23:49 -04:00
  • 7fc7bc30e7 Remove usage of eval_tokens for cache check Andrei Betlen 2023-05-26 20:12:05 -04:00
  • fe331ec589 Replace eval_logits and eval_tokens with numpy arrays Andrei Betlen 2023-05-26 20:03:31 -04:00
  • 00ea3af51b Add makefile Andrei Betlen 2023-05-26 17:56:20 -04:00
  • 447a3d249e Merge branch 'main' into setup Andrei Betlen 2023-05-26 17:53:58 -04:00
  • efb763bcdc Merge branch 'main' into add-numpy-support Andrei Betlen 2023-05-26 17:37:33 -04:00
  • 030fafe901 Add project changelog Andrei Betlen 2023-05-26 17:32:34 -04:00
  • 878c7d363f Merge branch 'main' into add-numpy-support Andrei Betlen 2023-05-26 17:22:29 -04:00
  • 6075e17cb6 Bump version v0.1.55 Andrei Betlen 2023-05-26 17:21:51 -04:00
  • bd4b95da45 Reduce numpy version dependency Andrei Betlen 2023-05-26 16:38:21 -04:00
  • 8eb9769f78 Add support for numpy Andrei Betlen 2023-05-26 16:12:45 -04:00
  • 2adf6f3f9a Merge pull request #265 from dmahurin/fix-from-bytes-byteorder Andrei 2023-05-26 12:53:06 -04:00
  • 34ad71f448 Merge pull request #274 from dmahurin/fix-missing-antiprompt Andrei 2023-05-26 12:52:34 -04:00
  • d78453c045 Merge pull request #264 from dmahurin/fix-min-keep Andrei 2023-05-26 12:52:05 -04:00
  • 4c1b7f7a76 Bugfix for logits_processor and stopping_criteria Andrei Betlen 2023-05-26 10:25:28 -04:00
  • 0fa2ec4903 low_level_api_chat_cpp.py: Fix missing antiprompt output in chat. Don Mahurin 2023-05-26 06:35:15 -07:00
  • 433a2e3e8a Add extra logits_processor and stopping_criteria Andrei Betlen 2023-05-26 03:13:24 -04:00
  • 30bf8ec557 Update llama.cpp Andrei Betlen 2023-05-26 03:03:11 -04:00
  • f74b90ed67 Fix streaming hang on last token when cache is on. Andrei Betlen 2023-05-26 03:03:01 -04:00
  • 5be8354e11 Added tokenizer Andrei Betlen 2023-05-26 03:00:51 -04:00
  • 8fa2ef1959 Format Andrei Betlen 2023-05-26 03:00:35 -04:00
  • 6bd1075291 Merge branch 'Maximilian-Winter/main' into main Andrei Betlen 2023-05-26 02:56:11 -04:00
  • ca01f98e09 Add LlamaTokenizer class Andrei Betlen 2023-05-25 14:11:33 -04:00
  • 1d247e0f35 Add StoppingCriteria and LogitsProcessor to generate to match huggingface API Andrei Betlen 2023-05-25 14:04:54 -04:00
  • c6a9659972 Merge branch 'abetlen:main' into main Maximilian Winter 2023-05-25 17:09:19 +02:00
  • de8d9a810b Merge pull request #270 from gjmulder/auto-docker Andrei 2023-05-25 09:30:13 -04:00
  • 0e0c9bb978 Merge branch 'auto-docker' of github.com:gjmulder/llama-cpp-python-gary into auto-docker Gary Mulder 2023-05-25 11:50:34 +00:00
  • 0d2cc21202 Fixed repeated imports Gary Mulder 2023-05-25 11:50:02 +00:00
  • c2585b6889 Fixed list elements typing Maximilian-Winter 2023-05-25 10:54:08 +02:00
  • da463e6c8c Added types to logit processor list and stop criteria list Maximilian-Winter 2023-05-25 09:07:16 +02:00
  • c05fcdf42f Fixed none value of logits processors. Maximilian-Winter 2023-05-24 22:02:06 +02:00
  • 5bb780d455 Implemented logit processors and stop criteria's Maximilian-Winter 2023-05-24 21:55:44 +02:00
  • fab064ded9 Remove unnecessary ffi calls Andrei Betlen 2023-05-23 17:56:21 -04:00
  • ec44bdad61 Update README.md Gary Mulder 2023-05-23 20:50:39 +01:00
  • ed19071ef8 Renamed and moved old Dockerfiles Gary Mulder 2023-05-23 19:38:37 +00:00
  • 70f629a72f Update README.md Gary Mulder 2023-05-23 20:36:21 +01:00
  • eaff7a8678 Initial commit of auto docker Gary Mulder 2023-05-23 19:26:40 +00:00
  • d6a7adb17a fix "missing 1 required positional argument: 'min_keep'" Don Mahurin 2023-05-22 23:54:57 -07:00
  • 327eedbfe1 fix "from_bytes() missing required argument 'byteorder'" Don Mahurin 2023-05-22 23:56:25 -07:00
  • e5d596e0e9 Bump version v0.1.54 Andrei Betlen 2023-05-22 23:50:58 -04:00
  • c41b1ebca7 Update llama.cpp Andrei Betlen 2023-05-22 23:50:35 -04:00
  • aa3d7a6299 Merge pull request #263 from abetlen/dependabot/pip/mkdocs-material-9.1.14 Andrei 2023-05-22 23:44:51 -04:00
  • 2240b949ae Bump mkdocs-material from 9.1.12 to 9.1.14 dependabot[bot] 2023-05-22 21:18:57 +00:00
  • 01c79e7bf1 Merge pull request #258 from Pipboyguy/main Andrei 2023-05-22 17:17:17 -04:00
  • c3e80b1714 Merge pull request #262 from abetlen/dependabot/pip/httpx-0.24.1 Andrei 2023-05-22 17:16:16 -04:00
  • 8e41d724ab Bump httpx from 0.24.0 to 0.24.1 dependabot[bot] 2023-05-22 21:05:39 +00:00
  • e6639e6620 Change docker build dynamic param to image instead of cuda version Marcel Coetzee 2023-05-22 10:10:14 +02:00
  • 4f7a6daa25 Merge pull request #248 from localagi/main Andrei 2023-05-22 03:15:15 -04:00
  • 0adb9ec37a Use model_name and index in response Andrei Betlen 2023-05-21 21:30:03 -04:00
  • 922b5b2bfd Merge branch 'main' into server-embedding Andrei Betlen 2023-05-21 21:21:38 -04:00
  • 2c45255a0a Bump version v0.1.53 Andrei Betlen 2023-05-21 19:24:20 -04:00
  • cd102e9da1 Cache shared library function calls for static tokens Andrei Betlen 2023-05-21 19:18:56 -04:00
  • b895511cca Fix penalize_nl Andrei Betlen 2023-05-21 18:38:06 -04:00
  • 03e2947b03 Fix unnecessary memory allocation while sampling Andrei Betlen 2023-05-21 18:36:34 -04:00
  • fafe47114c Update llama.cpp Andrei Betlen 2023-05-21 17:47:21 -04:00
  • 8f49ca0287 Bump version v0.1.52 Andrei Betlen 2023-05-20 08:53:40 -04:00
  • 76b1d2cd20 Change properties to functions to match token functions Andrei Betlen 2023-05-20 08:24:06 -04:00
  • a7ba85834f Add n_ctx, n_vocab, and n_embd properties Andrei Betlen 2023-05-20 08:13:41 -04:00
  • 0b079a658c make git module accessible anonymously Your Name 2023-05-20 02:25:59 +01:00
  • e783f1c191 feat: make embedding support list of string as input Simon Chabot 2023-05-20 01:23:32 +02:00
  • 01a010be52 Fix llama_cpp and Llama type signatures. Closes #221 Andrei Betlen 2023-05-19 11:59:33 -04:00
  • fb57b9470b Merge branch 'main' of github.com:abetlen/llama_cpp_python into main v0.1.51 Andrei Betlen 2023-05-19 03:19:32 -04:00
  • f82d85fbee Bump version Andrei Betlen 2023-05-19 03:19:27 -04:00
  • c7788c85ab Add Guidance example Andrei Betlen 2023-05-19 03:16:58 -04:00
  • a8cd169251 Bugfix: Stop sequences can be strings Andrei Betlen 2023-05-19 03:15:08 -04:00
  • f0812c4d8c Add upgrade instructions to the README Andrei Betlen 2023-05-19 02:20:41 -04:00
  • 17d4271b04 Fix logprobs for completions and implement for streaming logprobs. Andrei Betlen 2023-05-19 02:20:27 -04:00
  • a634a2453b Allow first logprob token to be null to match openai api Andrei Betlen 2023-05-19 02:04:57 -04:00
  • dc39cc0fa4 Use server sent events function for streaming completion Andrei Betlen 2023-05-19 02:04:30 -04:00
  • 69f9d50090 Merge pull request #235 from Pipboyguy/main Andrei 2023-05-18 13:42:04 -04:00
  • f0ec6e615e Stream tokens instead of text chunks Andrei Betlen 2023-05-18 11:35:59 -04:00