You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The LLM will run with your provided prompt, and the LLM's response to your prompt will be printed to the screen. You can replace the `"Hello, my thoughts are"` with any prompt you like.
@@ -97,6 +97,9 @@ You can also replace the `facebook/opt-125m` with any Hugging Face checkpoint yo
97
97
98
98
You can also set the `--device` argument in `oga-load` and `huggingface-load` to load your LLM on a different device.
99
99
100
+
The `-t` (or `--template`) flag instructs lemonade to insert the prompt string into the model's chat template.
101
+
This typically results in the model returning a higher quality response.
102
+
100
103
Run `lemonade huggingface-load -h` and `lemonade llm-prompt -h` to learn more about these tools.
Copy file name to clipboardExpand all lines: examples/lemonade/server/continue.md
+15-14Lines changed: 15 additions & 14 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -31,22 +31,23 @@ This will add a Continue tab to your VS Code Activity Bar.
31
31
> Note: The following instructions are based on instructions from Continue found [here](https://docs.continue.dev/customize/model-providers/openai#openai-compatible-servers--apis)
32
32
33
33
1. Open the Continue tab in your VS Code Activity Bar.
34
-
1. Click the gear icon at the top to open Settings.
35
-
1. Under "Configuration", click "Open Config File".
36
-
1. Replace the "models" key in the `config.json` with the following and save:
37
-
38
-
```json
39
-
"models": [
40
-
{
41
-
"title": "Lemonade",
42
-
"provider": "openai",
43
-
"model": "Qwen-1.5-7B-Chat-Hybrid",
44
-
"apiKey": "-",
45
-
"apiBase": "http://localhost:8000/api/v0"
46
-
}
47
-
],
34
+
1. Click the chat box. Some buttons will appear at the bottom of the box, including `Select model`.
35
+
1. Click `Select model`, then `+ Add Chat model` to open the new model dialog box.
36
+
1. Click the `config file` link at the very bottom of the dialog to open `config.yaml`.
37
+
1. Replace the "models" key in the `config.yaml` with the following and save:
38
+
39
+
```yaml
40
+
models:
41
+
- name: Lemonade
42
+
provider: openai
43
+
model: Qwen-1.5-7B-Chat-Hybrid
44
+
apiBase: http://localhost:8000/api/v0
45
+
apiKey: none
48
46
```
49
47
48
+
6. Close the dialog box.
49
+
7. Click the chat box again. You should see `Lemonade` where you used to see `Select model`. Ready!
50
+
50
51
## Usage
51
52
52
53
> Note: see the Continue [user guide](https://docs.continue.dev/) to learn about all of their features.
LangString MUI_TEXT_LICENSE_SUBTITLE ${LANG_ENGLISH}"Please review the license terms before installing AMD Ryzen AI Hybrid Execution Mode."
461
-
LangString DESC_SEC01 ${LANG_ENGLISH}"The minimum set of dependencies for a lemonade server that runs LLMs on CPU."
393
+
LangString DESC_SEC01 ${LANG_ENGLISH}"The minimum set of dependencies for a lemonade server that runs LLMs on CPU (includes Python)."
462
394
LangString DESC_HybridSec ${LANG_ENGLISH}"Add support for running LLMs on Ryzen AI hybrid execution mode. Only available on Ryzen AI 300-series processors."
463
395
LangString DESC_ModelsSec ${LANG_ENGLISH}"Select which models to install"
0 commit comments