You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: boost/README.md
+24-5
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
-
> Handle: `boost`
1
+
> Handle: `boost`<br/>
2
2
> URL: <http://localhost:34131/>
3
3
4
-
 <small>`g1` and `rcn` optimizer modules compared to original LLMs. [BBH256](https://gist.github.com/av/18cc8138a0acbe1b30f51e8bb19add90) task, run with [Harbor Bench](../docs/5.1.-Harbor-Bench)</small>
4
+

5
5
6
6
`boost` is a service that acts as an optimizing LLM proxy. It takes your inputs, and pre-processes them before sending them to the downstream API.
7
7
@@ -12,14 +12,21 @@ Features that make Harbor's `boost` special:
12
12
* 🎭 `boost` can serve as a plain LLM proxy (multiple downstream APIs behind a single endpoint)
13
13
* ✍️ `boost` is scriptable, you can write your own modules
14
14
15
+

16
+
15
17
The main focus, of course are the workflows that can help improve the LLM output in specific scenarios. Here are some examples of what's possible with `boost`:
16
18
17
-
* When "random" is mentioned in the message, `klmbr` will rewrite 35% of message characters to increase the entropy and produce more diverse completion
18
-
* Launch self-reflection reasoning chain when the message ends with a question mark
19
+
* Add R1-like reasoning to [any LLM](https://www.reddit.com/r/LocalLLaMA/comments/1ixckba/making_older_llms_llama_2_and_gemma_1_reason/)
20
+
* When "random" is mentioned in the message, [`klmbr`](#klmbr---boost-llm-creativity) will rewrite 35% of message characters to increase the entropy and produce more diverse completion
21
+
* Launch [self-reflection reasoning](#rcn---recursive-certainty-validation) chain when the message ends with a question mark
19
22
* Expand the conversation context with the "inner monologue" of the model, where it can iterate over your question a few times before giving the final answer
20
23
* Apply a specific LLM personality if the message contains a specific keyword
24
+
* Add external memory to your interactions with a specific model
25
+
* Make your LLM [pass a skill check](https://www.reddit.com/r/LocalLLaMA/comments/1jaqylp/llm_must_pass_a_skill_check_to_talk_to_me/) before replying to you
26
+
27
+
Boost is scriptable, you can provision your own modules with the workflows suitable for your needs. See [Custom Modules](#custom-modules) section for more information.
21
28
22
-
Moreover, boost is scriptable, you can provision your own modules with the workflows suitable for your needs. See [Custom Modules](#custom-modules) section for more information.
29
+
 <small>`g1` and `rcn` optimizer modules compared to original LLMs. [BBH256](https://gist.github.com/av/18cc8138a0acbe1b30f51e8bb19add90) task, run with [Harbor Bench](../docs/5.1.-Harbor-Bench)</small>
23
30
24
31
`boost` operates at the OpenAI-compatible API level, so can be used with any LLM backend that accepts OpenAI API requests. You can also plug `boost` into the UIs that are compatible with OpenAI API.
25
32
@@ -43,6 +50,7 @@ Moreover, boost is scriptable, you can provision your own modules with the workf
43
50
*[`supersummer` - Super Summarization](#supersummer---super-summarization)
⚠️ This module is experimental and only compatible with Open WebUI as a client due to its support of custom artifacts.
524
+
525
+
When serving the completion, LLM will first invent a skill check it must pass to address your message. Then, the workflow will roll a dice determining if the model passes the check or not and will guide the model to respond accordingly.
526
+
527
+
```bash
528
+
# Enable the module
529
+
harbor boost modules add dnd
530
+
```
531
+
513
532
### API
514
533
515
534
`boost` works as an OpenAI-compatible API proxy. It'll query configured downstream services for which models they serve and provide "boosted" wrappers in its own API.
widely adopted and feature rich web interface for interacting with LLMs. Supports OpenAI-compatible and Ollama backends, multi-users, multi-model chats, custom prompts, TTS, Web RAG, RAG, and much much more.
0 commit comments