Skip to content

Conversation

yaochengji
Copy link
Collaborator

No description provided.

@yaochengji yaochengji requested review from vanbasten23 and lsy323 June 3, 2025 23:42
@yaochengji
Copy link
Collaborator Author

The e2e vLLM inference throughput increased from 8.23 req/s to 11.99 req/s in my local benchmark.

Copy link
Collaborator

@vanbasten23 vanbasten23 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks Chengji!

@yaochengji yaochengji enabled auto-merge (squash) June 4, 2025 18:25
@yaochengji yaochengji merged commit e51af25 into master Jun 4, 2025
22 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants