Skip to content

Commit

Permalink
Demonstrate passing "max_tokens" param
Browse files Browse the repository at this point in the history
  • Loading branch information
mkhludnev authored Feb 15, 2024
1 parent 6f0afff commit 257a2a9
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion samples/client.py
Original file line number Diff line number Diff line change
Expand Up @@ -90,7 +90,7 @@ async def process_stream(self, prompts, sampling_parameters):
self._results_dict[result.get_response().id].append(i)

async def run(self):
sampling_parameters = {"temperature": "0.1", "top_p": "0.95"}
sampling_parameters = {"temperature": "0.1", "top_p": "0.95", "max_tokens":"100"}
with open(self._flags.input_prompts, "r") as file:
print(f"Loading inputs from `{self._flags.input_prompts}`...")
prompts = file.readlines()
Expand Down

0 comments on commit 257a2a9

Please sign in to comment.