How to use from the
Use from the
llama-cpp-python library
# !pip install llama-cpp-python

from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="ping98k/gemma-han-2b",
	filename="gemma-han-2b.Q4_K_M.gguf",
)
llm.create_chat_completion(
	messages = [
		{
			"role": "user",
			"content": "What is the capital of France?"
		}
	]
)

for test unsloth finetune process and Inference API

this model overfit with train data so it cannot answer anything not in han dataset

prompt

Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
จงแต่งบทกวีเกี่ยวกับสายฝนที่ผ่านมา

### Response:
Downloads last month
35
Safetensors
Model size
3B params
Tensor type
BF16
·
Inference Providers NEW
Input a message to start chatting with ping98k/gemma-han-2b.

Model tree for ping98k/gemma-han-2b

Base model

unsloth/gemma-2b
Quantized
(4)
this model
Quantizations
1 model

Dataset used to train ping98k/gemma-han-2b