Text Generation
GGUF
How to use from
llama.cpp
Install from brew
brew install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf dranger003/starcoder2-15b-GGUF:
# Run inference directly in the terminal:
llama-cli -hf dranger003/starcoder2-15b-GGUF:
Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf dranger003/starcoder2-15b-GGUF:
# Run inference directly in the terminal:
llama-cli -hf dranger003/starcoder2-15b-GGUF:
Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases
# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf dranger003/starcoder2-15b-GGUF:
# Run inference directly in the terminal:
./llama-cli -hf dranger003/starcoder2-15b-GGUF:
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli
# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf dranger003/starcoder2-15b-GGUF:
# Run inference directly in the terminal:
./build/bin/llama-cli -hf dranger003/starcoder2-15b-GGUF:
Use Docker
docker model run hf.co/dranger003/starcoder2-15b-GGUF:
Quick Links

GGUF quants for https://huggingface.co/bigcode/starcoder2-15b

StarCoder2-15B model is a 15B parameter model trained on 600+ programming languages from The Stack v2, with opt-out requests excluded. The model uses Grouped Query Attention, a context window of 16,384 tokens with a sliding window attention of 4,096 tokens, and was trained using the Fill-in-the-Middle objective on 4+ trillion tokens.

The model was trained on GitHub code as well as additional selected data sources such as Arxiv and Wikipedia. As such it is not an instruction model and commands like "Write a function that computes the square root." do not work well.

Layers Context Template (None/Base Model)
40
16384
{prompt}
Downloads last month
57
GGUF
Model size
16B params
Architecture
starcoder2
Hardware compatibility
Log In to add your hardware

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support