Skip to content

Release/v0.6.1 #117

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 5 commits into from
May 27, 2025
Merged

Release/v0.6.1 #117

merged 5 commits into from
May 27, 2025

Conversation

XkunW
Copy link
Contributor

@XkunW XkunW commented May 26, 2025

PR Type

[Release]

Short Description

  • Added Slurm dependency example
  • Added unit tests for vec-inf client and missing unit tests for vec-inf API
  • Fixed multi-node launch GPU placement group issue: --exclusive option is needed for slurm script and compilation config needs to stay at 0
  • Set environment variables in the generated slurm script instead of in the helper to ensure reusability
  • Replaced python3.10 -m vllm.entrypoints.openai.api_server with vllm serve to support custom chat template usage
  • Added additional launch options: --exclude for excluding certain nodes, --node-list for targeting a specific list of nodes, and --bind for binding additional directories
  • Added remaining vLLM engine arg short-long name mappings for robustness
  • Added notes in documentation to capture some gotchas and added vLLM version info

Tests Added

tests/vec-inf/client/test_api.py:

  • shutdown_model()
  • wait_until_ready()

tests/vec-inf/client/test_helper.py:

  • ModelRegistry
  • PerformanceMetricsCollector
  • ModelStatusMonitor
  • ModelLauncher

@XkunW XkunW requested a review from Copilot May 27, 2025 00:30
Copy link
Contributor

@Copilot Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull Request Overview

This release v0.6.1 bumps the package version and enriches documentation with usage examples, engine-argument references, and gotchas.

  • Bump project version to 0.6.1
  • Expanded user guide with additional notes on engine args, GPU dtype, and compilation config
  • Updated index and README to reference the latest vLLM image version

Reviewed Changes

Copilot reviewed 4 out of 4 changed files in this pull request and generated 2 comments.

File Description
pyproject.toml Bump version from 0.6.0 to 0.6.1
docs/user_guide.md Added detailed NOTE bullets on additional config options, engine args, GPU dtype, and compilation config
docs/index.md Updated Docker recommendation to mention vLLM version 0.8.5.post1
README.md Updated Docker recommendation and note formatting to match index.md

@XkunW XkunW merged commit 7f382ba into main May 27, 2025
7 checks passed
@XkunW XkunW deleted the release/v0.6.1 branch May 27, 2025 01:03
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant