Update README.md
Browse files
README.md
CHANGED
|
@@ -2,57 +2,30 @@
|
|
| 2 |
license: mit
|
| 3 |
---
|
| 4 |
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
## gte-large.Q6_K.gguf
|
| 8 |
-
|
| 9 |
-
A Q6_K quantized version of [General Text Embeddings (GTE) model](https://huggingface.co/thenlper/gte-large) under MIT License. Used for all things RAG in the library.
|
| 10 |
|
| 11 |
-
|
| 12 |
|
| 13 |
-
|
| 14 |
|
| 15 |
-
|
| 16 |
|
| 17 |
-
|
| 18 |
|
| 19 |
-
Powerful, object-oriented, and highly configurable, general purpose library used to connect a local back-end running a Large Language Model (LLM) to a front-end. This library offers many tools and features for those who want to code their own C# front-end or LLM-powered tools without having to do all the heavy lifting. It can happily connect itself to the most popular backends (the program loading the LLM proper) and allows your code to "speak" with the LLM is a few function calls. It includes easy-to-use (and easy to build upon) systems to handle most of the operations you'd wish to do with a LLM, alongside many advanced features like RAG, agentic systems, web search, text to speech, semantic similarity testing, and prompt manipulation.
|
| 20 |
|
| 21 |
-
##
|
| 22 |
-
- **Kobold API:** Used by [KoboldCpp](https://github.com/LostRuins/koboldcpp). This is the recommended backend with the most features.
|
| 23 |
-
- **OpenAI API:** Used by [LM Studio](https://lmstudio.ai/), [Text Generation WebUI](https://github.com/oobabooga/text-generation-webui), and others. Less features.
|
| 24 |
|
| 25 |
-
|
| 26 |
|
| 27 |
-
## ⭐ Main Features
|
| 28 |
-
- Easy to use classes for bot personas, system prompts, instruction formats, and inference settings
|
| 29 |
-
- Session-based chatlog with automated summaries for past sessions
|
| 30 |
-
- Streamed (or not) inference / reroll / and impersonate functions
|
| 31 |
-
- Support CoT / "thinking" models out of the box
|
| 32 |
-
- GBNF grammar generation directly from a class's structure for structured output
|
| 33 |
-
- Basic support for VLM (visual language models) depending on the back-end
|
| 34 |
-
- Tools for reliable Web Search (DuckDuckGo and Brave)
|
| 35 |
-
- Text To Speech support (through the *Kobold API* only)
|
| 36 |
-
- Many useful tools to manipulate text, count tokens, and more
|
| 37 |
|
| 38 |
-
|
| 39 |
-
- Keyword-triggered text insertions (also known as "world info" in many frontends)
|
| 40 |
-
- Customizable RAG System using the Small World implementation
|
| 41 |
-
- Automatic (optional) and configurable insertion of relevant past chat sessions into the context
|
| 42 |
-
|
| 43 |
-
## 🧠 Agentic and Brain Module for personas
|
| 44 |
-
- Background agent system (bot can run tasks in the background)
|
| 45 |
-
- Analyze past chat sessions, run relevant web searches and mention results in next session
|
| 46 |
-
- Mood tracking + drift system (personality coloring over time)
|
| 47 |
-
- Goal‑driven behaviors (long‑term projects, self‑seeding topics of interest)
|
| 48 |
|
| 49 |
-
##
|
| 50 |
-
- Group chat functionalities (one user and multiple AI characters)
|
| 51 |
-
- Sentiment analysis
|
| 52 |
|
| 53 |
-
|
| 54 |
|
| 55 |
-
|
| 56 |
|
| 57 |
-
|
| 58 |
|
|
|
|
|
|
| 2 |
license: mit
|
| 3 |
---
|
| 4 |
|
| 5 |
+
# Lethe AI Sharp
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
|
| 7 |
+
Misc useful content for the [Lethe AI Sharp](https://github.com/SerialKicked/Lethe-AI-Sharp/) library. It is a modular, object‑oriented C# library that connects local or remote Large Language Model (LLM) backends to your applications (desktop tools, game engines, services). It also comes with its own light backend, allowing you to run a local LLM in the GGUF format directly without even having to rely on anything else.
|
| 8 |
|
| 9 |
+
It unifies: chat personas, conversation/session management, streaming inference, long‑term memory, RAG (retrieval augmented generation), background agentic tasks, web search tools, TTS, and structured output generation. It is extensible, documented, and backend-agnostic (you write the same code no matter which backend is being used)
|
| 10 |
|
| 11 |
+
**No Python Dependencies:** Pure .NET 10 C# implementation. No Python runtime, no conda environments, no pip hell.
|
| 12 |
|
| 13 |
+
**Self-Contained:** Built-in LlamaSharp backend means you can distribute a single executable that runs LLMs locally. No external server required, but external servers are supported too.
|
| 14 |
|
|
|
|
| 15 |
|
| 16 |
+
## Fixed ChatML Jinja Templates for Qwen 3.5
|
|
|
|
|
|
|
| 17 |
|
| 18 |
+
This repo also contains fixed Jinja templates for Qwen 3.5 models. This one allows for system messages mid conversations (requirement for LetheAI) while removing an error that would trigger (at least) on LM Studio.
|
| 19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
|
| 21 |
+
Two models required by the [**Lethe-AI C# Middleware Library**](https://github.com/SerialKicked/Lethe-AI-Sharp)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
|
| 23 |
+
## gte-large.Q6_K.gguf
|
|
|
|
|
|
|
| 24 |
|
| 25 |
+
A Q6_K quantized version of [General Text Embeddings (GTE) model](https://huggingface.co/thenlper/gte-large) under MIT License. Used for all things RAG in the library.
|
| 26 |
|
| 27 |
+
## emotion-bert-classifier.gguf
|
| 28 |
|
| 29 |
+
A quantized version of [Emotions Analyzer](https://huggingface.co/logasanjeev/emotions-analyzer-bert) under MIT License, a fine-tuned BERT-base-uncased on GoEmotions for multi-label classification (28 emotions). Used for sentiment analysis tasks.
|
| 30 |
|
| 31 |
+
---
|