How to use from
llama.cppInstall from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf c2p-cmd/google_gemma_guff# Run inference directly in the terminal:
llama-cli -hf c2p-cmd/google_gemma_guffUse pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf c2p-cmd/google_gemma_guff# Run inference directly in the terminal:
./llama-cli -hf c2p-cmd/google_gemma_guffBuild from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf c2p-cmd/google_gemma_guff# Run inference directly in the terminal:
./build/bin/llama-cli -hf c2p-cmd/google_gemma_guffUse Docker
docker model run hf.co/c2p-cmd/google_gemma_guffQuick Links
Gemma Model Card
Model Page: Gemma
This model card corresponds to the 2B and 7B Instruct versions of the Gemma model's Guff.
Terms of Use: Terms
Description
Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
Model Usage
Since this is a guff, it can be run locally using
- Ollama
- Llama.cpp
- LM Studio
- And Many More
- I have provided GemmaModelFile that can be used with ollama by:
- Download the model:
pip install huggingface_hub from huggingface_hub import hf_hub_download model_id="c2p-cmd/google_gemma_guff" hf_hub_download(repo_id=model_id, local_dir="gemma_snapshot", local_dir_use_symlinks=False, filename="gemma_snapshot/gemma-2b-it.gguf") - Load the model file to ollama
ollama create gemma -f GemmaModelFile - You change the model name based on needs
- Downloads last month
- 14
Hardware compatibility
Log In to add your hardware
We're not able to determine the quantization variants.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf c2p-cmd/google_gemma_guff# Run inference directly in the terminal: llama-cli -hf c2p-cmd/google_gemma_guff