--- license: apache-2.0 tags: - text-generation - instruction-tuned - llama - gguf - chatbot library_name: llama.cpp language: en datasets: - custom model-index: - name: Corelyn NeoMini results: [] base_model: - mistralai/Ministral-3-3B-Base-2512 --- ![logo](./images/neospecyn.png) # Corelyn NeoMini GGUF Model ## Specifications : - Model Name: Corelyn NeoMini - Base Name: NeoMini-3B - Type: Instruct / Fine-tuned - Architecture: Ministral-3 - Size: 3B parameters - Organization: Corelyn ## Model Overview Corelyn NeoMini is a 3-billion parameter LLaMA-based instruction-tuned model, designed for general-purpose assistant tasks and knowledge extraction. It is a fine-tuned variant optimized for instruction-following use cases. - Fine-tuning type: Instruct - Base architecture: Ministral-3 - Parameter count: 3B ### This model is suitable for applications such as: - Chatbots and conversational AI - Knowledge retrieval and Q&A - Code and text generation - Instruction-following tasks ## Usage Download from : [NeoMini3.2](https://huggingface.co/CorelynAI/NeoMini/resolve/main/NeoMini_3B.gguf) ```python # pip install pip install llama-cpp-python from llama_cpp import Llama # Load the model (update the path to where your .gguf file is) llm = Llama(model_path="path/to/the/file/NeoMini_3B.gguf") # Create chat completion response = llm.create_chat_completion( messages=[{"role": "user", "content": "Create a Haiku about AI"}] ) # Print the generated text print(response.choices[0].message["content"]) ```