Skip to content

fix: gpu memory report in vllm sleep mode#1101

Closed
hiyouga wants to merge 2 commits intomainfrom
yaowei/perf_log
Closed

fix: gpu memory report in vllm sleep mode#1101
hiyouga wants to merge 2 commits intomainfrom
yaowei/perf_log

Conversation

@hiyouga
Copy link
Copy Markdown
Collaborator

@hiyouga hiyouga commented Apr 15, 2025

What does this PR do?

When we use vLLM's sleep mode, we can no longer obtain the GPU memory statistics from the PyTorch API because vLLM uses cuMem to manage the memory space. We compute the freed memory of vLLM offloading to estimate the memory correctly.

see vllm-project/vllm#11743 (comment)

Before:
image

After:
image

Who can review?

@vermouth1992 @tongyx361 @BearBiscuit05

@hiyouga
Copy link
Copy Markdown
Collaborator Author

hiyouga commented Apr 16, 2025

closed in favor of #1118

@hiyouga hiyouga closed this Apr 16, 2025
@hiyouga hiyouga deleted the yaowei/perf_log branch April 16, 2025 14:15
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants