Update README.md
Browse files
README.md
CHANGED
|
@@ -26,6 +26,24 @@ The GRaPE Family was trained on about **14 billion** tokens of data after pre-tr
|
|
| 26 |
|
| 27 |
GRaPE Flash and Nano are monomodal models, only accepting text. GRaPE Mini being trained most recently supports image and video inputs.
|
| 28 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
# How to Run
|
| 30 |
|
| 31 |
I recommend using **LM Studio** for running GRaPE Models, and have generally found these sampling parameters to work best:
|
|
|
|
| 26 |
|
| 27 |
GRaPE Flash and Nano are monomodal models, only accepting text. GRaPE Mini being trained most recently supports image and video inputs.
|
| 28 |
|
| 29 |
+
***
|
| 30 |
+
|
| 31 |
+
## Reasoning Modes
|
| 32 |
+
|
| 33 |
+
As GRaPE Mini is the only model that thinks, it has *some* support for reasoning modes. In testing, these modes sometimes work. Likely due to an innefficient dataset formatting for it.
|
| 34 |
+
|
| 35 |
+
To use thinking modes, you need an XML tag, `<thinking_mode>`, which can equal these values:
|
| 36 |
+
|
| 37 |
+
- **Minimal**: Skip thinking *(does not work most of the time, you'll have to be careful with this one)*
|
| 38 |
+
- **Low**: Think Below 1024 tokens
|
| 39 |
+
- **Medium**: Think between 1024 and 8192 tokens
|
| 40 |
+
- **High**: Think for any amount above 8192 tokens
|
| 41 |
+
|
| 42 |
+
In your prompt, place the thinking mode at the *end* of your prompt, like this:
|
| 43 |
+
```
|
| 44 |
+
Build me a website called "Aurora Beats." <thinking_mode=medium
|
| 45 |
+
```
|
| 46 |
+
|
| 47 |
# How to Run
|
| 48 |
|
| 49 |
I recommend using **LM Studio** for running GRaPE Models, and have generally found these sampling parameters to work best:
|