Skip to content

Commit

Permalink
Removed explicit mode for multi-lora (#45)
Browse files Browse the repository at this point in the history
  • Loading branch information
oandreeva-nv authored Jul 5, 2024
1 parent 18a96e3 commit db3d794
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 5 deletions.
4 changes: 0 additions & 4 deletions ci/L0_multi_gpu/multi_lora/multi_lora_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -119,7 +119,6 @@ def _test_vllm_model(
self.triton_client.stop_stream()

def test_multi_lora_requests(self):
self.triton_client.load_model(self.vllm_model_name)
sampling_parameters = {"temperature": "0", "top_p": "1"}
# make two requests separately to avoid the different arrival of response answers
prompt_1 = ["Instruct: What do you think of Computer Science?\nOutput:"]
Expand Down Expand Up @@ -151,10 +150,8 @@ def test_multi_lora_requests(self):
exclude_input_in_output=True,
expected_output=expected_output,
)
self.triton_client.unload_model(self.vllm_model_name)

def test_none_exist_lora(self):
self.triton_client.load_model(self.vllm_model_name)
prompts = [
"Instruct: What is the capital city of France?\nOutput:",
]
Expand All @@ -169,7 +166,6 @@ def test_none_exist_lora(self):
exclude_input_in_output=True,
expected_output=None, # this request will lead to lora not supported error, so there is no expected output
)
self.triton_client.unload_model(self.vllm_model_name)

def tearDown(self):
self.triton_client.close()
Expand Down
2 changes: 1 addition & 1 deletion ci/L0_multi_gpu/multi_lora/test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ source ../../common/util.sh
TRITON_DIR=${TRITON_DIR:="/opt/tritonserver"}
SERVER=${TRITON_DIR}/bin/tritonserver
BACKEND_DIR=${TRITON_DIR}/backends
SERVER_ARGS="--model-repository=`pwd`/models --backend-directory=${BACKEND_DIR} --model-control-mode=explicit --log-verbose=1"
SERVER_ARGS="--model-repository=`pwd`/models --backend-directory=${BACKEND_DIR} --log-verbose=1"
SERVER_LOG="./multi_lora_server.log"
CLIENT_LOG="./multi_lora_client.log"
TEST_RESULT_FILE='test_results.txt'
Expand Down

0 comments on commit db3d794

Please sign in to comment.