gemma2b-nirf-lookup-gguf

This is a GGUF conversion of coderop12/gemma2b-nirf-lookup-2025.

Model Details

  • Original Model: coderop12/gemma2b-nirf-lookup-2025
  • Format: GGUF (F16 precision)
  • File Size: ~4.9 GB
  • Architecture: Gemma 2B
  • Specialization: NIRF (National Institutional Ranking Framework) lookup and ranking queries

Usage

With llama.cpp

./llama-cli -m gemma2b-nirf-lookup-gguf.gguf -p "What is the NIRF ranking methodology?"

With Python (llama-cpp-python)

from llama_cpp import Llama

llm = Llama(model_path="gemma2b-nirf-lookup-gguf.gguf")
response = llm("What are the top NIRF ranked engineering colleges?")
print(response['choices'][0]['text'])

With Ollama

# First, create a Modelfile
echo 'FROM ./gemma2b-nirf-lookup-gguf.gguf' > Modelfile
ollama create gemma2b-nirf-lookup-gguf -f Modelfile
ollama run gemma2b-nirf-lookup-gguf "Explain NIRF ranking parameters"

Model Capabilities

This model is specifically fine-tuned for:

  • NIRF ranking information and queries
  • Indian higher education institutional data
  • University and college ranking explanations
  • Educational policy and framework questions

Technical Details

  • Quantization: F16 (16-bit floating point)
  • Context Length: 2048 tokens
  • License: Follow original model license terms
  • Converted using: llama.cpp conversion tools

Original Model License

Please refer to the original model repository for license information.

Downloads last month
9
GGUF
Model size
3B params
Architecture
gemma2
Hardware compatibility
Log In to view the estimation

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for coderop12/gemma2b-nirf-lookup-gguf

Base model

google/gemma-2-2b
Quantized
(1)
this model