models

/

c4ai-command-r-08-2024-35B

Version:

35B GPU: F16

PyTorch

This version is not quantized and a GPU is recommended.

  1. Install our magic package manager:

    curl -ssL https://magic.modular.com/ | bash

    Then run the source command that's printed in your terminal.

  2. Install Max Pipelines in order to run this model.

    magic global install max-pipelines && magic global update
  3. Start a local endpoint for c4ai-command-r-08-2024/35B:

    max-pipelines serve --huggingface-repo-id=CohereForAI/c4ai-command-r-plus-08-2024

    The endpoint is ready when you see the URI printed in your terminal:

    Server ready on http://0.0.0.0:8000 (Press CTRL+C to quit)
  4. Now open another terminal to send a request using curl:

    curl -N http://0.0.0.0:8000/v1/chat/completions -H "Content-Type: application/json" -d '{
        "model": "CohereForAI/c4ai-command-r-plus-08-2024",
        "stream": true,
        "messages": [
            {"role": "system", "content": "You are a helpful assistant."},
            {"role": "user", "content": "Who won the World Series in 2020?"}
        ]
    }' | grep -o '"content":"[^"]*"' | sed 's/"content":"//g' | sed 's/"//g' | tr -d '
    ' | sed 's/\n//g'
  5. 🎉 Hooray! You’re running Generative AI. Our goal is to make this as easy as possible.

Deploy this model to cloud

DETAILS

ChatMODEL CLASS
PyTorch
HARDWARE
GPU
QUANTIZATION
F16
ARCHITECTURE
PyTorch

MAX GITHUB

Modular / MAX

MODEL

CohereForAI

CohereForAI/c4ai-command-r-plus-08-2024

QUESTIONS ABOUT THIS MODEL?

Leave a comment

PROBLEMS WITH THE CODE?

File an Issue

TAGS

transformers

/

safetensors

/

cohere

/

text-generation

/

conversational

/

en

/

fr

/

de

/

es

/

it

/

pt

/

ja

/

ko

/

zh

/

ar

/

doi:10.57967/hf/3136

/

license:cc-by-nc-4.0

/

autotrain_compatible

/

text-generation-inference

/

region:us

Resources & support for
running c4ai-command-r-08-2024-35B

@ Copyright - Modular Inc - 2025