You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+5-5
Original file line number
Diff line number
Diff line change
@@ -44,7 +44,7 @@ RAGLite is a Python toolkit for Retrieval-Augmented Generation (RAG) with Postgr
44
44
> 🚀 If you want to use local models, it is recommended to install [an accelerated llama-cpp-python precompiled binary](https://github.com/abetlen/llama-cpp-python?tab=readme-ov-file#supported-backends) with:
45
45
> ```sh
46
46
># Configure which llama-cpp-python precompiled binary to install (⚠️ not every combination is available):
> 🧠 RAGLite extends [LiteLLM](https://github.com/BerriAI/litellm) with support for [llama.cpp](https://github.com/ggerganov/llama.cpp) models using [llama-cpp-python](https://github.com/abetlen/llama-cpp-python). To select a llama.cpp model (e.g., from [bartowski's collection](https://huggingface.co/bartowski)), use a model identifier of the form `"llama-cpp-python/<hugging_face_repo_id>/<filename>@<n_ctx>"`, where `n_ctx` is an optional parameter that specifies the context size of the model.
95
+
> 🧠 RAGLite extends [LiteLLM](https://github.com/BerriAI/litellm) with support for [llama.cpp](https://github.com/ggerganov/llama.cpp) models using [llama-cpp-python](https://github.com/abetlen/llama-cpp-python). To select a llama.cpp model (e.g., from [Unsloth's collection](https://huggingface.co/unsloth)), use a model identifier of the form `"llama-cpp-python/<hugging_face_repo_id>/<filename>@<n_ctx>"`, where `n_ctx` is an optional parameter that specifies the context size of the model.
96
96
97
97
> [!TIP]
98
98
> 💾 You can create a PostgreSQL database in a few clicks at [neon.tech](https://neon.tech).
@@ -112,7 +112,7 @@ my_config = RAGLiteConfig(
112
112
# Example 'local' config with a SQLite database and a llama.cpp LLM:
Copy file name to clipboardExpand all lines: src/raglite/_chatml_function_calling.py
+25-8
Original file line number
Diff line number
Diff line change
@@ -6,6 +6,9 @@
6
6
b. ✨ Add function descriptions to the system message so that tool use is better informed (fixes https://github.com/abetlen/llama-cpp-python/issues/1869).
7
7
c. ✨ Replace `print` statements relating to JSON grammars with `RuntimeWarning` warnings.
8
8
d. ✅ Add tests with fairly broad coverage of the different scenarios.
9
+
e. 🐛 Fix a 'content' KeyError in the prompt template.
10
+
f. ✨ Add support for Qwen3's <|endoftext|> separator.
11
+
g. ✨ Add support for Qwen3's <think>...</think> mode to (auto and fixed) function calling.
9
12
4. Case "Tool choice by user":
10
13
a. ✨ Add support for more than one function call by making this a special case of "Automatic tool choice" with a single tool (subsumes https://github.com/abetlen/llama-cpp-python/pull/1503).
11
14
5. Case "Automatic tool choice -> respond with a message":
0 commit comments