File size: 2,738 Bytes
d392d41 ad559a7 d392d41 90a47cb d392d41 90a47cb d804a91 90a47cb d392d41 8141634 d392d41 8141634 3d89d33 6f8578c 3d89d33 90a47cb 3d89d33 90a47cb 3d89d33 90a47cb 3d89d33 5bc68a4 3d89d33 d392d41 ad559a7 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 | ---
license: mit
tags:
- jinja
- templates
---
# Repo Information
This repo contains useful content for the [Lethe AI Sharp](https://github.com/SerialKicked/Lethe-AI-Sharp/) library: classification models, fixed jinja templates, and other files. More information about the library below.
## Fixed Jinja Templates
This repo also contains fixed (and more permissive) Jinja templates for Mistral (Tekken7) and Qwen 3.5 (ChatML) models. They allow for system messages mid conversations (requirement for LetheAI), and fixes other so-called errors that could trigger when using those LLM in way not initially intended.
## gte-large.Q6_K.gguf
A Q6_K quantized version of [General Text Embeddings (GTE) model](https://huggingface.co/thenlper/gte-large) under MIT License. Used for all things RAG in the library.
## emotion-bert-classifier.gguf
A quantized version of [Emotions Analyzer](https://huggingface.co/logasanjeev/emotions-analyzer-bert) under MIT License, a fine-tuned BERT-base-uncased on GoEmotions for multi-label classification (28 emotions). Used for experimental sentiment analysis tasks. Optional.
# What is Lethe AI Sharp?
[Lethe AI Sharp](https://github.com/SerialKicked/Lethe-AI-Sharp/) is a modular, object‑oriented C# library that connects local or remote Large Language Model (LLM) backends to your applications (desktop tools, game engines, services). It also comes with its own light backend, allowing you to run a local LLM in the GGUF format directly without even having to rely on anything else.
It unifies: chat personas, conversation/session management, streaming inference, long‑term memory, RAG (retrieval augmented generation), background agentic tasks, web search tools, TTS, and structured output generation. It is extensible, documented, and backend-agnostic (you write the same code no matter which backend is being used)
**No Python Dependencies:** Pure .NET 10 C# implementation. No Python runtime, no conda environments, no pip hell.
**Self-Contained:** Built-in LlamaSharp backend means you can distribute a single executable that runs LLMs locally. No external server required, but external servers (KoboldAPI and OpenAI API) are supported too.
# Any program using Lethe AI Sharp?
Sure there's [Lethe AI Chat](https://github.com/SerialKicked/Lethe-AI-Chat/), it's a native Windows 10 frontend chat application. Compiled binaries are [available here](https://github.com/SerialKicked/Lethe-AI-Chat/releases).

While still in early beta, it's pretty much on par with other modern frontends, and contains many unique features.
--- |