Skip to content

Conversation

@rishabh135
Copy link
Contributor

Convert torch.bfloat16, torch.float16, etc. to valid CLI string

Checks dtype params when building the subprocess_command for vllm.

Passing torch.bfloat16 values seems to stop vllm from deploying properly, and result in

vllm serve: error: argument --dtype: invalid choice: 'torch.bfloat16' (choose from 'auto', 'bfloat16', 'float', 'float16', 'float32', 'half')

The update makes it such that vLLM will deploy even when the correct dtype isn't specified isn't specified in the model name

@danielhanchen
Copy link
Contributor

Oh thank you! Could you remove some extra spaces and newlines - thanks :)

Copy link
Contributor Author

@rishabh135 rishabh135 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

removed newline characters as requested. Kindly merge this submission!

@danielhanchen
Copy link
Contributor

Thank you!

@danielhanchen danielhanchen merged commit a15e77f into unslothai:main Jun 30, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants