StarCoder2-15B-Instruct is a self-aligned model optimizing code generation without human annotations.
Version:
15B GPU: BF16
This version is not quantized and a GPU is recommended.
Install our magic
package manager:
curl -ssL https://magic.modular.com/ | bash
Then run the source
command that's printed in your terminal.
Install Max Pipelines in order to run this model.
magic global install max-pipelines && magic global update
Start a local endpoint for starcoder2-instruct-v0.1/15B:
max-pipelines serve --huggingface-repo-id=bigcode/starcoder2-15b-instruct-v0.1
The endpoint is ready when you see the URI printed in your terminal:
Server ready on http://0.0.0.0:8000 (Press CTRL+C to quit)
Now open another terminal to send a request using curl
:
curl -N http://0.0.0.0:8000/v1/chat/completions -H "Content-Type: application/json" -d '{
"model": "bigcode/starcoder2-15b-instruct-v0.1",
"stream": true,
"messages": [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Who won the World Series in 2020?"}
]
}' | grep -o '"content":"[^"]*"' | sed 's/"content":"//g' | sed 's/"//g' | tr -d '
' | sed 's/\n//g'
🎉 Hooray! You’re running Generative AI. Our goal is to make this as easy as possible.
DETAILS
MAX GITHUB
Modular / MAX
MODEL
bigcode
bigcode/starcoder2-15b-instruct-v0.1
QUESTIONS ABOUT THIS MODEL?
Leave a comment
PROBLEMS WITH THE CODE?
File an Issue
TAGS
ENTERPRISES
@ Copyright - Modular Inc - 2025