Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

[BUG] Does not support multi-GPU in vLLM #156

Copy link
Copy link
Open
@hoaaosnw

Description

@hoaaosnw
Issue body actions

Prerequisites

System Information

2 NV A100 GPUs

Problem Description

When use vLLM and set tensor_parallel_size = 2, failing to load model checkpoints.

Steps to Reproduce

error "Exception in worker VllmWorkerProcess while processing method load_model"

Expected Behavior

No response

Additional Context

No response

Usage Statistics (Optional)

No response

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions

    Morty Proxy This is a proxified and sanitized view of the page, visit original site.