models

/

llava-v1.5-7B

Version:

7B GPU:

PyTorch

This version is not quantized and a GPU is recommended.

  1. Install our magic package manager:

    curl -ssL https://magic.modular.com/ | bash

    Then run the source command that's printed in your terminal.

  2. Install Max Pipelines in order to run this model.

    magic global install max-pipelines && magic global update
  3. Start a local endpoint for llava-v1.5/7B:

    max-pipelines serve --huggingface-repo-id=liuhaotian/llava-v1.5-7b

    The endpoint is ready when you see the URI printed in your terminal:

    Server ready on http://0.0.0.0:8000 (Press CTRL+C to quit)
  4. Now open another terminal to send a request using curl:

    curl -N http://0.0.0.0:8000/v1/chat/completions -H "Content-Type: application/json" -d '{
        "model": "liuhaotian/llava-v1.5-7b",
        "stream": true,
        "messages": [
            {
              "role": "user",
              "content": [
                {
                  "type": "text",
                  "text": "What is in this image?"
                },
                {
                  "type": "image_url",
                  "image_url": {
                    "url": "http://images-assets.nasa.gov/image/iss072e571418/iss072e571418~orig.jpg"
                  }
                }
              ]
            }
        ]
    
    }' | grep -o '"content":"[^"]*"' | sed 's/"content":"//g' | sed 's/"//g' | tr -d '
    ' | sed 's/\n//g'
  5. 🎉 Hooray! You’re running Generative AI. Our goal is to make this as easy as possible.

Deploy this model to cloud

DETAILS

VisionMODEL CLASS
PyTorch
HARDWARE
GPU
QUANTIZATION
ARCHITECTURE
PyTorch

MAX GITHUB

Modular / MAX

MODEL

liuhaotian

liuhaotian/llava-v1.5-7b

QUESTIONS ABOUT THIS MODEL?

Leave a comment

PROBLEMS WITH THE CODE?

File an Issue

TAGS

transformers

/

pytorch

/

llava

/

text-generation

/

image-text-to-text

/

autotrain_compatible

/

region:us

Resources & support for
running llava-v1.5-7B

@ Copyright - Modular Inc - 2025