You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
You can repro by starting up the server example and sending requests with different temperatures. Failure should look like
async for i, res in result_generator:
File "/home/cglagovich/vllm/vllm/utils.py", line 506, in merge_async_iterators
item = await d
File "/home/cglagovich/vllm/vllm/engine/multiprocessing/client.py", line 598, in _process_request
raise request_output
File "/home/cglagovich/vllm/vllm/engine/multiprocessing/client.py", line 598, in _process_request
raise request_output
AssertionError: Currently only supporting same temperature for all sequences in batch```
The text was updated successfully, but these errors were encountered:
Anything you want to discuss about vllm.
You can repro by starting up the server example and sending requests with different temperatures. Failure should look like
The text was updated successfully, but these errors were encountered: