Skip to content

Commit a286a9c

Browse files
authored
[CI][CPU] adding numa node number as container name suffix (vllm-project#10441)
Signed-off-by: Yuan Zhou <yuan.zhou@intel.com>
1 parent 24025b5 commit a286a9c

File tree

1 file changed

+8
-8
lines changed

1 file changed

+8
-8
lines changed

.buildkite/run-cpu-test.sh

+8-8
Original file line numberDiff line numberDiff line change
@@ -13,26 +13,26 @@ numactl -C "$CORE_RANGE" -N "$NUMA_NODE" docker build -t cpu-test -f Dockerfile.
1313
numactl -C "$CORE_RANGE" -N "$NUMA_NODE" docker build --build-arg VLLM_CPU_DISABLE_AVX512="true" -t cpu-test-avx2 -f Dockerfile.cpu .
1414

1515
# Setup cleanup
16-
remove_docker_container() { docker rm -f cpu-test cpu-test-avx2 || true; }
16+
remove_docker_container() { docker rm -f cpu-test-"$NUMA_NODE" cpu-test-avx2-"$NUMA_NODE" || true; }
1717
trap remove_docker_container EXIT
1818
remove_docker_container
1919

2020
# Run the image, setting --shm-size=4g for tensor parallel.
2121
docker run -itd --entrypoint /bin/bash -v ~/.cache/huggingface:/root/.cache/huggingface --cpuset-cpus="$CORE_RANGE" \
22-
--cpuset-mems="$NUMA_NODE" --privileged=true --network host -e HF_TOKEN --env VLLM_CPU_KVCACHE_SPACE=4 --shm-size=4g --name cpu-test cpu-test
22+
--cpuset-mems="$NUMA_NODE" --privileged=true --network host -e HF_TOKEN --env VLLM_CPU_KVCACHE_SPACE=4 --shm-size=4g --name cpu-test-"$NUMA_NODE" cpu-test
2323
docker run -itd --entrypoint /bin/bash -v ~/.cache/huggingface:/root/.cache/huggingface --cpuset-cpus="$CORE_RANGE" \
24-
--cpuset-mems="$NUMA_NODE" --privileged=true --network host -e HF_TOKEN --env VLLM_CPU_KVCACHE_SPACE=4 --shm-size=4g --name cpu-test-avx2 cpu-test-avx2
24+
--cpuset-mems="$NUMA_NODE" --privileged=true --network host -e HF_TOKEN --env VLLM_CPU_KVCACHE_SPACE=4 --shm-size=4g --name cpu-test-avx2-"$NUMA_NODE" cpu-test-avx2
2525

2626
function cpu_tests() {
2727
set -e
2828

2929
# offline inference
30-
docker exec cpu-test-avx2 bash -c "
30+
docker exec cpu-test-avx2-"$NUMA_NODE" bash -c "
3131
set -e
3232
python3 examples/offline_inference.py"
3333

3434
# Run basic model test
35-
docker exec cpu-test bash -c "
35+
docker exec cpu-test-"$NUMA_NODE" bash -c "
3636
set -e
3737
pip install pytest pytest-asyncio \
3838
decord einops librosa peft Pillow sentence-transformers soundfile \
@@ -45,20 +45,20 @@ function cpu_tests() {
4545
pytest -v -s tests/models/decoder_only/vision_language -m cpu_model"
4646

4747
# Run compressed-tensor test
48-
docker exec cpu-test bash -c "
48+
docker exec cpu-test-"$NUMA_NODE" bash -c "
4949
set -e
5050
pytest -s -v \
5151
tests/quantization/test_compressed_tensors.py::test_compressed_tensors_w8a8_static_setup \
5252
tests/quantization/test_compressed_tensors.py::test_compressed_tensors_w8a8_dynamic_per_token"
5353

5454
# Run AWQ test
55-
docker exec cpu-test bash -c "
55+
docker exec cpu-test-"$NUMA_NODE" bash -c "
5656
set -e
5757
pytest -s -v \
5858
tests/quantization/test_ipex_quant.py"
5959

6060
# online inference
61-
docker exec cpu-test bash -c "
61+
docker exec cpu-test-"$NUMA_NODE" bash -c "
6262
set -e
6363
export VLLM_CPU_KVCACHE_SPACE=10
6464
export VLLM_CPU_OMP_THREADS_BIND=$1

0 commit comments

Comments
 (0)