Commit Graph

  • 22bc1e8a36 Use official cibuildwheel action Andrei Betlen 2024-03-04 13:13:06 -05:00
  • af42fb7ab6 Update download label Andrei Betlen 2024-03-04 12:49:42 -05:00
  • 5c501c00b2 Add total release downloads badge Andrei Betlen 2024-03-04 12:42:03 -05:00
  • 3dd2616d60 Generate binary wheel index on release Andrei Betlen 2024-03-03 11:43:37 -05:00
  • 87a6e5797e feat: Update llama.cpp Andrei Betlen 2024-03-03 11:27:04 -05:00
  • 13177aae0f chore: Bump version v0.2.55 Andrei Betlen 2024-03-02 22:46:40 -05:00
  • 663659f730 docs: fix small typo in README: 'model know how' -> 'model knows how' (#1244) Kenneth Hoste 2024-03-03 04:20:41 +01:00
  • 0e70984fb6 feat: Update llama.cpp Andrei Betlen 2024-03-02 22:20:04 -05:00
  • d5df431278 chore: Bump version v0.2.54 Andrei Betlen 2024-03-01 13:15:16 -05:00
  • 97aa3a153d docs: Add information re: auto chat formats. Closes #1236 Andrei Betlen 2024-03-01 13:10:25 -05:00
  • f062a7f51d feat: Update llama.cpp Andrei Betlen 2024-03-01 12:57:16 -05:00
  • cf1fdd8a9a docs: fix typo in README.md embeddings example. (#1232) Douglas Hanley 2024-02-29 12:55:50 -06:00
  • 8c71725d53 fix: Remove deprecated cfg sampling functions Andrei Betlen 2024-02-28 14:37:07 -05:00
  • 727d60c28a misc: Format Andrei Betlen 2024-02-28 14:27:40 -05:00
  • 0d37ce52b1 feat: Update llama.cpp Andrei Betlen 2024-02-28 14:27:16 -05:00
  • ffcd4b2636 chore: Bump version v0.2.53 Andrei Betlen 2024-02-28 01:38:32 -05:00
  • c36ab15e68 fix: eos/bos_token set correctly for Jinja2ChatFormatter and automatic chat formatter (#1230) Sigbjørn Skjæret 2024-02-28 07:30:31 +01:00
  • fea33c9b94 feat: Update llama.cpp Andrei Betlen 2024-02-27 12:22:17 -05:00
  • 4d574bd765 feat(server): Add support for pulling models from Huggingface Hub (#1222) Andrei 2024-02-26 14:35:08 -05:00
  • b3e358dee4 docs: Add example of local image loading to README Andrei Betlen 2024-02-26 11:58:33 -05:00
  • afe1e445c9 chore: Bump version v0.2.52 Andrei Betlen 2024-02-26 11:43:24 -05:00
  • 9558ce7878 feat: Update llama.cpp Andrei Betlen 2024-02-26 11:40:58 -05:00
  • a57d5dff86 feat: Update llama.cpp Andrei Betlen 2024-02-26 11:37:43 -05:00
  • 79c649c2d1 docs: Update multimodal example Andrei Betlen 2024-02-26 11:34:45 -05:00
  • bf315ee7a9 docs: Update multimodal example Andrei Betlen 2024-02-26 11:32:11 -05:00
  • dbaba3059d fix: positional arguments only for low-level api Andrei Betlen 2024-02-26 11:31:11 -05:00
  • 78e536dcfe fix: typo Andrei Betlen 2024-02-26 11:14:26 -05:00
  • 44558cbd7a misc: llava_cpp use ctypes function decorator for binding Andrei Betlen 2024-02-26 11:07:33 -05:00
  • 8383a9e562 fix: llava this function takes at least 4 arguments (0 given) Andrei Betlen 2024-02-26 11:03:20 -05:00
  • 34111788fe feat: Update llama.cpp Andrei Betlen 2024-02-26 10:58:41 -05:00
  • 5fc4c1efb6 Merge branch 'main' of https://github.com/abetlen/llama-cpp-python into main v0.2.51 Andrei Betlen 2024-02-25 21:15:54 -05:00
  • 8e03fd9957 chore: Bump version Andrei Betlen 2024-02-25 21:14:01 -05:00
  • e857c133fb feat: Update llama.cpp Andrei Betlen 2024-02-25 21:14:01 -05:00
  • 252e1ff2b4 docs(examples): Add huggingface pull example Andrei Betlen 2024-02-25 21:09:41 -05:00
  • bd4ec2e612 docs(examples): Add gradio chat example Andrei Betlen 2024-02-25 21:09:13 -05:00
  • dcf38f6141 fix: remove prematurely commited change Andrei Betlen 2024-02-25 21:00:37 -05:00
  • cbbcd888af feat: Update llama.cpp Andrei Betlen 2024-02-25 20:52:14 -05:00
  • 19234aa0db fix: Restore type hints for low-level api Andrei Betlen 2024-02-25 16:54:37 -05:00
  • 2292af5796 feat: Update llama.cpp Andrei Betlen 2024-02-25 16:53:58 -05:00
  • 221edb9ef1 feat: Update llama.cpp Andrei Betlen 2024-02-24 23:47:29 -05:00
  • 20ea6fd7d6 chore: Bump version v0.2.50 Andrei Betlen 2024-02-23 12:38:36 -05:00
  • b681674bf2 docs: Fix functionary repo_id Andrei Betlen 2024-02-23 12:36:13 -05:00
  • f94faab686 Merge branch 'main' of https://github.com/abetlen/llama-cpp-python into main Andrei Betlen 2024-02-23 12:34:03 -05:00
  • 702306b381 docs: Restore functionary docs in README Andrei Betlen 2024-02-23 12:34:02 -05:00
  • bce6dc0ac2 docs: Update Functionary OpenAI Server Readme (#1193) Jeffrey Fong 2024-02-24 01:24:10 +08:00
  • 47bad30dd7 fix: LlamaHFTokenizer now receives pre_tokens Andrei Betlen 2024-02-23 12:23:24 -05:00
  • ded5d627a5 chore: Bump version v0.2.49 Andrei Betlen 2024-02-23 11:32:43 -05:00
  • 858496224e feat: Auto detect Mixtral's slightly different format (#1214) Luke Stanley 2024-02-23 16:27:38 +00:00
  • db776a885c fix: module 'llama_cpp.llama_cpp' has no attribute 'c_uint8' Andrei Betlen 2024-02-23 11:24:53 -05:00
  • 427d816ebf chore: Bump version v0.2.48 Andrei Betlen 2024-02-23 04:54:08 -05:00
  • 52d9d70076 docs: Update README.md to fix pip install llama cpp server (#1187) Aditya Purandare 2024-02-23 15:11:22 +05:30
  • 251a8a2cad feat: Add Google's Gemma formatting via chat_format="gemma" (#1210) Alvaro Bartolome 2024-02-23 18:40:52 +09:00
  • eebb102df7 feat: Update llama.cpp Andrei Betlen 2024-02-23 03:42:08 -05:00
  • 5f96621e92 misc: only search tests folder for tests Andrei Betlen 2024-02-23 03:40:25 -05:00
  • b9aca612af misc: use typesafe byref for internal classes Andrei Betlen 2024-02-23 03:40:07 -05:00
  • a0ce429dc0 misc: use decorator to bind low level api functions, fixes docs Andrei Betlen 2024-02-23 03:39:38 -05:00
  • 410e02da51 docs: Fix typo Andrei Betlen 2024-02-23 00:43:31 -05:00
  • eb56ce2e2a docs: fix low-level api example Andrei Betlen 2024-02-22 11:33:05 -05:00
  • 0f8cad6cb7 docs: Update README Andrei Betlen 2024-02-22 11:31:44 -05:00
  • 045cc12670 docs: Update README Andrei Betlen 2024-02-22 03:53:52 -05:00
  • e10af30cf1 fix: TypeAlias import error Andrei Betlen 2024-02-22 03:27:28 -05:00
  • 3561ebf536 Merge branch 'main' of https://github.com/abetlen/llama-cpp-python into main Andrei Betlen 2024-02-22 03:25:13 -05:00
  • 32efed7b07 docs: Update README Andrei Betlen 2024-02-22 03:25:11 -05:00
  • d80c5cf29d docs: fix indentation for mkdocs-material Andrei Betlen 2024-02-22 02:30:24 -05:00
  • aefcb8f71a misc: additional type annotations for low level api Andrei Betlen 2024-02-22 02:00:09 -05:00
  • 3921e10770 feat: support minItems/maxItems in JSON grammar converter (by @nopperl) Andrei Betlen 2024-02-22 00:17:06 -05:00
  • e6d6260a91 fix: Update from_pretrained defaults to match hf_hub_download Andrei Betlen 2024-02-22 00:10:23 -05:00
  • dd22010e85 fix: Raise exceptions when llama model or context fails to load Andrei Betlen 2024-02-22 00:09:45 -05:00
  • 3632241e98 chore: Bump version v0.2.47 Andrei Betlen 2024-02-21 23:09:13 -05:00
  • 0653e15c20 feat: Update llama.cpp Andrei Betlen 2024-02-21 23:04:52 -05:00
  • 7981e9ce1e chore: Bump version v0.2.46 Andrei Betlen 2024-02-21 16:30:59 -05:00
  • 7f3962e11c feat: Update llama.cpp Andrei Betlen 2024-02-21 16:27:56 -05:00
  • 14191e9036 docs: Add create_chat_completion_openai_v1 to api reference Andrei Betlen 2024-02-21 16:26:49 -05:00
  • fe5626cd40 misc: add .local pattern to gitignore Andrei Betlen 2024-02-21 16:26:30 -05:00
  • 7f51b6071f feat(low-level-api): Improve API static type-safety and performance (#1205) Andrei 2024-02-21 16:25:38 -05:00
  • 0f8aa4ab5c feat: Pull models directly from huggingface (#1206) Andrei 2024-02-21 16:25:10 -05:00
  • e42f62c247 chore: Bump version v0.2.45 Andrei Betlen 2024-02-21 11:09:40 -05:00
  • 4edde21b3d feat: Update llama.cpp Andrei Betlen 2024-02-21 11:05:58 -05:00
  • f57b01ac9b ci: add debug build to dev makefile Andrei Betlen 2024-02-21 11:04:30 -05:00
  • 04fe33b999 feat: Update llama.cpp Andrei Betlen 2024-02-20 02:59:02 -05:00
  • d122bd7858 feat: Update llama.cpp Andrei Betlen 2024-02-19 22:10:16 -05:00
  • 6225f027e5 feat: Update llama.cpp Andrei Betlen 2024-02-19 04:11:34 -05:00
  • 748c0ce057 feat: Update llama.cpp Andrei Betlen 2024-02-18 21:30:36 -05:00
  • 53f6f5f415 fix: self.numa missing Andrei Betlen 2024-02-17 01:02:33 -05:00
  • fdce078cb9 feat: Update llama.cpp Andrei Betlen 2024-02-17 00:37:51 -05:00
  • c2a234a086 docs: Add embeddings section Andrei Betlen 2024-02-15 23:15:50 -05:00
  • f736827b9b chore: Bump version v0.2.44 Andrei Betlen 2024-02-15 23:10:50 -05:00
  • 0ce66bc080 fix: create_embedding broken response for input type str Andrei Betlen 2024-02-15 16:09:48 -05:00
  • ea1f88dd29 fix: Use '\n' seperator for EventSourceResponse (#1188) khimaros 2024-02-15 20:20:13 +00:00
  • a5cfeb7763 feat: Update llama.cpp Andrei Betlen 2024-02-15 15:17:30 -05:00
  • 7bb91f025f fix: Incorporate embedding pooling layer fixes (#1194) Douglas Hanley 2024-02-15 14:16:30 -06:00
  • ae71ad1a14 Bump version v0.2.43 Andrei Betlen 2024-02-14 04:31:42 -05:00
  • f300d4310a Merge branch 'main' of https://github.com/abetlen/llama-cpp-python into main Andrei Betlen 2024-02-14 04:27:33 -05:00
  • c336f78269 Update llama.cpp Andrei Betlen 2024-02-14 04:27:30 -05:00
  • d7a67917ba feat: Support batch embeddings (#1186) Douglas Hanley 2024-02-14 03:26:09 -06:00
  • 36b843228f misc: fix makefile build commands Andrei Betlen 2024-02-14 03:47:40 -05:00
  • 7b9960d1cb Update llama.cpp Andrei Betlen 2024-02-14 03:47:21 -05:00
  • 6943bab6d8 fix: destructor exception where internal classes are missing some uninitialized attributes Andrei Betlen 2024-02-14 03:38:41 -05:00
  • 07a783779a fix: Update openbuddy prompt format. Closes #1155 Andrei Betlen 2024-02-13 23:57:10 -05:00
  • 7a79e5ac49 Update llama.cpp Andrei Betlen 2024-02-13 23:54:05 -05:00