Skip to content

How to dispatch LLM across GPUs like device_map="auto" of transformers.AutoModelForCausalLM #6372

Closed Answered by SlyJabiru
SlyJabiru asked this question in Q&A
Discussion options

You must be logged in to vote

[Self response]
llm = LLM(model=model_id, tensor_parallel_size=4, gpu_memory_utilization=0.5) solved this issue.
solution from #550

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by SlyJabiru
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
1 participant