Commit Graph

  • f1edc66b21 Update llama.cpp Andrei Betlen 2023-12-11 10:21:35 -05:00
  • 40f2293ff0 Refactor _create_completion Andrei Betlen 2023-11-30 12:24:54 -05:00
  • 7ae9a3e614 Add vocab utils Andrei Betlen 2023-11-29 18:55:48 -05:00
  • 85caba5a63 Merge branch 'main' into batch-processing Andrei Betlen 2023-11-29 05:43:36 -05:00
  • f3b844ed0a Update llama.cpp Andrei Betlen 2023-11-29 05:40:22 -05:00
  • b069d06346 Fix #891 (#952) kddubey 2023-11-29 02:39:52 -08:00
  • 3a1ba77462 Add sampling context Andrei Betlen 2023-11-28 06:30:03 -05:00
  • b5c33da5a8 Merge branch 'main' into batch-processing Andrei Betlen 2023-11-28 04:59:35 -05:00
  • ad963a0961 Bump version v0.2.21 Andrei Betlen 2023-11-28 04:58:20 -05:00
  • e3941d9c67 Make building llava optional Andrei Betlen 2023-11-28 04:55:21 -05:00
  • 74f1949206 Update llama.cpp Andrei Betlen 2023-11-28 04:54:51 -05:00
  • 135b300b85 docs: re-order api reference Andrei Betlen 2023-11-28 04:12:32 -05:00
  • 3c13436e5e Clean up llama.py into seperate modules. Andrei Betlen 2023-11-28 04:09:13 -05:00
  • fb32f9d438 docs: Update README Andrei Betlen 2023-11-28 03:15:01 -05:00
  • 43e006a291 docs: Remove divider Andrei Betlen 2023-11-28 02:41:50 -05:00
  • 2cc6c9ae2f docs: Update README, add FAQ Andrei Betlen 2023-11-28 02:37:34 -05:00
  • 7f3704b896 Bump version v0.2.20 Andrei Betlen 2023-11-27 19:14:25 -05:00
  • f99b2385ee Update llama.cpp Andrei Betlen 2023-11-27 19:03:10 -05:00
  • 396dbf0b2b docs: Improve low-level docstrings Andrei Betlen 2023-11-27 19:03:02 -05:00
  • 9c68b1804a docs: Add api reference links in README Andrei Betlen 2023-11-27 18:54:07 -05:00
  • 174ef3ddf6 docs: Add headings to API reference Andrei Betlen 2023-11-27 18:42:15 -05:00
  • 41428244f0 docs: Fix README indentation Andrei Betlen 2023-11-27 18:29:13 -05:00
  • 1539146a5e docs: Fix README docs link Andrei Betlen 2023-11-27 18:21:00 -05:00
  • a928893d03 Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-11-26 15:57:13 -05:00
  • 6308f21d5e docs: Update Llama docs Andrei Betlen 2023-11-26 15:56:40 -05:00
  • aa5a7a1880 Update README.md (#940) Anton Vice 2023-11-27 04:39:38 +08:00
  • c2d63a7148 fix: Typo in the Open Orca chat format #874 (#947) Gardner Bickford 2023-11-27 09:39:18 +13:00
  • f03a38e62a Update llama.cpp Andrei Betlen 2023-11-26 15:38:22 -05:00
  • 1a7bf2037b docs: Update openapi endpoint names Andrei Betlen 2023-11-24 03:39:29 -05:00
  • 4026166e68 docs: Update completion and chat_completion parameter docstrings Andrei Betlen 2023-11-24 03:24:19 -05:00
  • 945e20fa2c docs: update link Andrei Betlen 2023-11-24 00:18:32 -05:00
  • e6a36b840e docs: edit function calling docs Andrei Betlen 2023-11-24 00:17:54 -05:00
  • 8c3aa7858b Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-11-24 00:15:36 -05:00
  • 19e02f1f87 docs: Add link to function calling notebook Andrei Betlen 2023-11-24 00:15:02 -05:00
  • de2e2bc083 misc fix verbose printing in functionary model Andrei Betlen 2023-11-23 20:14:23 -05:00
  • 36048d46af Update llama.cpp Andrei Betlen 2023-11-23 16:26:00 -05:00
  • d68fc07b1b Add Zephyr format (#937) mrfakename 2023-11-22 22:20:08 -08:00
  • 4184835078 Add chat format to support baichuan (#938) caiyesd 2023-11-23 14:19:50 +08:00
  • 4474157949 ci: tag built docker images with current version Andrei Betlen 2023-11-23 01:06:47 -05:00
  • 21abefa488 docs: Add grammar and types to api reference Andrei Betlen 2023-11-23 00:27:41 -05:00
  • 6aab77de04 docs: Fix module import bug Andrei Betlen 2023-11-23 00:27:22 -05:00
  • c647f01609 Add from_json_schema to LlamaGrammar Andrei Betlen 2023-11-23 00:27:00 -05:00
  • be1f64d569 docs: Add docstrings from llama.cpp Andrei Betlen 2023-11-23 00:26:26 -05:00
  • 31cf0ec680 docs: Fix mkdocstrings heading level Andrei Betlen 2023-11-22 23:45:19 -05:00
  • e349f314b4 docs: Fix API Reference page Andrei Betlen 2023-11-22 23:45:02 -05:00
  • b6bb7ac76a docs: Add Llama class example Andrei Betlen 2023-11-22 23:10:04 -05:00
  • c5173b0fb3 docs: Configure mkdocstrings Andrei Betlen 2023-11-22 23:09:42 -05:00
  • 3303ebe92b docs: Add dark mode and pymarkdown extensions Andrei Betlen 2023-11-22 22:47:22 -05:00
  • abb1976ad7 docs: Add n_ctx not for multimodal models Andrei Betlen 2023-11-22 21:07:00 -05:00
  • 36679a58ef Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-11-22 19:49:59 -05:00
  • bd43fb2bfe docs: Update high-level python api examples in README to include chat formats, function calling, and multi-modal models. Andrei Betlen 2023-11-22 19:49:56 -05:00
  • d977b44d82 docs: Add links to server functionality Andrei Betlen 2023-11-22 18:21:02 -05:00
  • aa815d580c docs: Link to langchain docs Andrei Betlen 2023-11-22 18:17:49 -05:00
  • 357e4dd69f docs: Use nav for better site layout control Andrei Betlen 2023-11-22 18:16:30 -05:00
  • 602ea64ddd docs: Fix whitespace Andrei Betlen 2023-11-22 18:09:31 -05:00
  • 971864ce92 docs: Watch README for changes during docs development Andrei Betlen 2023-11-22 18:08:17 -05:00
  • f336eebb2f docs: fix 404 to macos installation guide. Closes #861 Andrei Betlen 2023-11-22 18:07:30 -05:00
  • 1ff2c92720 docs: minor indentation fix Andrei Betlen 2023-11-22 18:04:18 -05:00
  • 68238b7883 docs: setting n_gqa is no longer required Andrei Betlen 2023-11-22 18:01:54 -05:00
  • 198178225c docs: Remove stale warning Andrei Betlen 2023-11-22 17:59:16 -05:00
  • 5a9770a56b Improve documentation for server chat formats (#934) Juraj Bednar 2023-11-22 12:10:03 +01:00
  • b8f29f4bf0 Add baichuan-2 chat format (#936) caiyesd 2023-11-22 19:08:06 +08:00
  • 9515467439 tests: add mock_kv_cache placeholder functions Andrei Betlen 2023-11-22 06:02:21 -05:00
  • 0ea244499e tests: avoid constantly reallocating logits Andrei Betlen 2023-11-22 04:31:05 -05:00
  • 0a7e05bc10 tests: don't mock sampling functions Andrei Betlen 2023-11-22 04:12:32 -05:00
  • d7388f1ffb Use mock_llama for all tests Andrei Betlen 2023-11-21 18:13:19 -05:00
  • dbfaf53fe0 Update llama.cpp Andrei Betlen 2023-11-21 18:12:38 -05:00
  • 8b6ca22846 Fix type warnings for json schema grammar converter Andrei Betlen 2023-11-21 13:32:00 -05:00
  • 230fc8b535 Bump version v0.2.19 Andrei Betlen 2023-11-21 05:04:55 -05:00
  • 128dc4731f Fix #569 Andrei Betlen 2023-11-21 04:39:05 -05:00
  • 7a3f87846b Format Andrei Betlen 2023-11-21 04:02:20 -05:00
  • 422ebc89ce Fix: Add logit_bias to all completion api methods Andrei Betlen 2023-11-21 04:01:36 -05:00
  • 79efc85206 Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-11-21 03:59:49 -05:00
  • 07e47f55ba Add support for logit_bias outside of server api. Closes #827 Andrei Betlen 2023-11-21 03:59:46 -05:00
  • 23a221999f Documenting server usage (#768) James Braza 2023-11-20 21:24:22 -08:00
  • c21edb6908 Do not set grammar to None for new LlamaGrammar objects (#834) Maarten ter Huurne 2023-11-21 06:23:18 +01:00
  • ef65fc5ff4 Add MistralLite, Intel, and OpenChat prompt formats (#927) mrfakename 2023-11-20 21:19:25 -08:00
  • 9d7c8307cd Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-11-20 23:23:20 -05:00
  • 3dc21b2557 tests: Improve llama.cpp mock Andrei Betlen 2023-11-20 23:23:18 -05:00
  • b8438f70b5 Added support for min_p (#921) TK-Master 2023-11-21 06:21:33 +02:00
  • 63fe1370ed Update llama.cpp Andrei Betlen 2023-11-20 23:19:47 -05:00
  • a34d480141 Fix #929 Andrei Betlen 2023-11-20 22:50:59 -05:00
  • 2c2afa320f Update llama.cpp Andrei Betlen 2023-11-20 14:11:33 -05:00
  • 6dde6bd09c bug fixing (#925) zocainViken 2023-11-20 18:31:52 +01:00
  • f3117c0cf6 fix: openblas_simple missing pkg-config to build (#920) ZisisTsatsas 2023-11-20 17:31:02 +00:00
  • 55efc9e6b2 Update llama.cpp Andrei Betlen 2023-11-16 18:51:55 -05:00
  • 96a377648c Merge tag 'v0.2.18' into main Andrei Betlen 2023-11-14 15:32:15 -05:00
  • ca30d898e9 Merge tag 'v0.2.17' into main Andrei Betlen 2023-11-14 15:32:12 -05:00
  • 3af167d8db Merge tag 'v0.2.16' into main Andrei Betlen 2023-11-14 15:32:08 -05:00
  • cc0fe43849 Disable opencl test Andrei Betlen 2023-11-14 14:59:08 -05:00
  • f2901d840e Bump version v0.2.18 Andrei Betlen 2023-11-14 14:10:00 -05:00
  • a14f46a720 Update llama.cpp Andrei Betlen 2023-11-14 14:08:52 -05:00
  • 020945049a Update llama.cpp Andrei Betlen 2023-11-13 22:20:19 -05:00
  • 01846a76b9 Bump version v0.2.17 Andrei Betlen 2023-11-10 16:36:12 -05:00
  • 7e3e85b53d Update llama.cpp Andrei Betlen 2023-11-10 16:33:55 -05:00
  • 4388f33414 Set CUDA_ARCHITECTURES=OFF for windows Andrei Betlen 2023-11-10 16:32:36 -05:00
  • 74167bdfb2 Update Functions notebook Andrei Betlen 2023-11-10 13:02:30 -05:00
  • 85ead98a3e Update Functions notebook example Andrei Betlen 2023-11-10 12:49:14 -05:00
  • b7e60b66f4 Bump version v0.2.16 Andrei Betlen 2023-11-10 06:21:24 -05:00
  • fb743f6c87 Update llama.cpp Andrei Betlen 2023-11-10 06:21:14 -05:00