Skip to content

Latest commit

 

History

History
136 lines (100 loc) · 4.89 KB

vllm_cpu_docker_quickstart.md

File metadata and controls

136 lines (100 loc) · 4.89 KB

vLLM Serving with IPEX-LLM on Intel CPU via Docker

This guide demonstrates how to run vLLM serving with ipex-llm on Intel CPU via Docker.

Install docker

Follow the instructions in this guide to install Docker on Linux.

Pull the latest image

Note: For running vLLM serving on Intel CPUs, you can currently use either the intelanalytics/ipex-llm-serving-cpu:latest or intelanalytics/ipex-llm-serving-vllm-cpu:latest Docker image.

# This image will be updated every day
docker pull intelanalytics/ipex-llm-serving-cpu:latest

Start Docker Container

To fully use your Intel CPU to run vLLM inference and serving, you should

#/bin/bash
export DOCKER_IMAGE=intelanalytics/ipex-llm-serving-cpu:latest
export CONTAINER_NAME=ipex-llm-serving-cpu-container
sudo docker run -itd \
        --net=host \
        --cpuset-cpus="0-47" \
        --cpuset-mems="0" \
        -v /path/to/models:/llm/models \
        -e no_proxy=localhost,127.0.0.1 \
        --memory="64G" \
        --name=$CONTAINER_NAME \
        --shm-size="16g" \
        $DOCKER_IMAGE

After the container is booted, you could get into the container through docker exec.

docker exec -it ipex-llm-serving-cpu-container /bin/bash

Running vLLM serving with IPEX-LLM on Intel CPU in Docker

We have included multiple vLLM-related files in /llm/:

  1. vllm_offline_inference.py: Used for vLLM offline inference example
  2. benchmark_vllm_throughput.py: Used for benchmarking throughput
  3. payload-1024.lua: Used for testing request per second using 1k-128 request
  4. start-vllm-service.sh: Used for template for starting vLLM service

Before performing benchmark or starting the service, you can refer to this section to setup our recommended runtime configurations.

Service

A script named /llm/start-vllm-service.sh have been included in the image for starting the service conveniently.

Modify the model and served_model_name in the script so that it fits your requirement. The served_model_name indicates the model name used in the API.

Then start the service using bash /llm/start-vllm-service.sh, the following message should be print if the service started successfully.

If the service have booted successfully, you should see the output similar to the following figure:

Verify

After the service has been booted successfully, you can send a test request using curl. Here, YOUR_MODEL should be set equal to served_model_name in your booting script, e.g. Qwen1.5.

curl http://localhost:8000/v1/completions \
-H "Content-Type: application/json" \
-d '{
  "model": "YOUR_MODEL",
  "prompt": "San Francisco is a",
  "max_tokens": 128,
  "temperature": 0
}' | jq '.choices[0].text'

Below shows an example output using Qwen1.5-7B-Chat with low-bit format sym_int4:

Tuning

You can tune the service using these four arguments:

  • --max-model-len
  • --max-num-batched-token
  • --max-num-seq

You can refer to this doc for a detailed explaination on these parameters.

Benchmark

Online benchmark throurgh api_server

We can benchmark the api_server to get an estimation about TPS (transactions per second). To do so, you need to start the service first according to the instructions mentioned above.

Then in the container, do the following:

  1. modify the /llm/payload-1024.lua so that the "model" attribute is correct. By default, we use a prompt that is roughly 1024 token long, you can change it if needed.
  2. Start the benchmark using wrk using the script below:
cd /llm
# warmup
wrk -t4 -c4 -d3m -s payload-1024.lua http://localhost:8000/v1/completions --timeout 1h
# You can change -t and -c to control the concurrency.
# By default, we use 8 connections to benchmark the service.
wrk -t8 -c8 -d15m -s payload-1024.lua http://localhost:8000/v1/completions --timeout 1h

Offline benchmark through benchmark_vllm_throughput.py

cd /llm
wget https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json

source ipex-llm-init -t
export MODEL="YOUR_MODEL"

python3 ./benchmark_vllm_throughput.py \
    --backend vllm \
    --dataset ./ShareGPT_V3_unfiltered_cleaned_split.json \
    --model $MODEL \
    --num-prompts 1000 \
    --seed 42 \
    --trust-remote-code \
    --enforce-eager \
    --dtype bfloat16 \
    --device cpu \
    --load-in-low-bit bf16