Skip to content

Commit

Permalink
Work around flaky Triton shared mem by retrying
Browse files Browse the repository at this point in the history
  • Loading branch information
hcho3 committed Oct 3, 2024
1 parent 6f35b84 commit 187f8a9
Showing 1 changed file with 11 additions and 2 deletions.
13 changes: 11 additions & 2 deletions qa/L0_e2e/test_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -281,6 +281,7 @@ def test_small(client, model_data, hypothesis_data):
model_inputs,
model_data.output_sizes,
shared_mem=shared_mem,
attempts=100,
)
for name, input_ in model_inputs.items():
all_model_inputs[name].append(input_)
Expand Down Expand Up @@ -324,7 +325,11 @@ def test_small(client, model_data, hypothesis_data):
st.one_of(st.just(mode) for mode in valid_shm_modes())
)
all_triton_outputs = client.predict(
model_data.name, all_model_inputs, total_output_sizes, shared_mem=shared_mem
model_data.name,
all_model_inputs,
total_output_sizes,
shared_mem=shared_mem,
attempts=100,
)

for output_name in sorted(ground_truth.keys()):
Expand Down Expand Up @@ -359,7 +364,11 @@ def test_max_batch(client, model_data, shared_mem):
}
shared_mem = valid_shm_modes()[0]
result = client.predict(
model_data.name, max_inputs, model_output_sizes, shared_mem=shared_mem
model_data.name,
max_inputs,
model_output_sizes,
shared_mem=shared_mem,
attempts=100,
)

ground_truth = model_data.ground_truth_model.predict(max_inputs)
Expand Down

0 comments on commit 187f8a9

Please sign in to comment.