|
set -x |
|
|
|
MODEL_PATH=$1 |
|
ANSWER_PATH=$2 |
|
METRIC=$3 |
|
PORT=8800 |
|
|
|
export VLLM_ENGINE_ITERATION_TIMEOUT_S=60 |
|
nohup vllm serve ${MODEL_PATH} \ |
|
--trust-remote-code \ |
|
--served-model-name server_model \ |
|
--max-num-seqs 256 \ |
|
--max-model-len 4096 \ |
|
--port 8000 \ |
|
> vllm_server.log & |
|
|
|
|
|
|
|
if [[ "${METRIC}" == "prob" ]]; then |
|
nohup python server.py \ |
|
--tokenizer_path ${MODEL_PATH} \ |
|
--answer_path ${ANSWER_PATH} \ |
|
--normalize_reward \ |
|
--port ${PORT} \ |
|
--prob_reward \ |
|
--vllm_url "http://localhost:8000/v1" \ |
|
--vllm_model server_model \ |
|
> reward_server.log & |
|
else |
|
nohup python server.py \ |
|
--tokenizer_path ${MODEL_PATH} \ |
|
--answer_path ${ANSWER_PATH} \ |
|
--normalize_reward \ |
|
--port ${PORT} \ |
|
--vllm_url "http://localhost:8000/v1" \ |
|
--vllm_model server_model \ |
|
> reward_server.log & |
|
fi |
|
|