MISHANM/google-gemma-2-2b-it.gguf

This model is a GGUF version of the Google gemma-2-2b-it model, optimized for use with the llama.cpp framework. It is designed to run efficiently on CPUs and can be used for various natural language processing tasks.

Model Details

  1. Language: English
  2. Tasks: Text generation
  3. Base Model: google/gemma-2-2b-it

Building and Running the Model

To build and run the model using llama.cpp, follow these steps:

Build llama.cpp Locally

git clone https://github.com/ggerganov/llama.cpp  
cd llama.cpp  
cmake -B build  
cmake --build build --config Release  

Run the Model

Navigate to the build directory and run the model with a prompt:

cd llama.cpp/build/bin   

Inference with llama.cpp

./llama-cli -m /path/to/model/ -p "Your prompt here" -n 128  

Citation Information

@misc{MISHANM/google-gemma-2-2b-it.gguf,
  author = {Mishan Maurya},
  title = {Introducing Google gemma-2-2b-it GGUF Model},
  year = {2025},
  publisher = {Hugging Face},
  journal = {Hugging Face repository},
  
}
Downloads last month
25
GGUF
Model size
2.61B params
Architecture
gemma2
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for MISHANM/google-gemma-2-2b-it.gguf

Base model

google/gemma-2-2b
Quantized
(151)
this model