|
24 | 24 | from fastdeploy.platforms import current_platform |
25 | 25 | from fastdeploy.scheduler import SchedulerConfig |
26 | 26 | from fastdeploy.utils import (ceil_div, check_unified_ckpt, get_host_ip, |
27 | | - is_port_available, llm_logger) |
| 27 | + is_port_available, get_random_port, llm_logger) |
28 | 28 |
|
29 | 29 | TaskOption = Literal["generate"] |
30 | 30 |
|
@@ -571,7 +571,7 @@ def __init__( |
571 | 571 | max_model_len: int = 8192, |
572 | 572 | max_num_seqs: int = 8, |
573 | 573 | max_num_batched_tokens: Optional[int] = None, |
574 | | - dist_init_addr: str = None, |
| 574 | + dist_init_ip: str = None, |
575 | 575 | nnodes: int = 1, |
576 | 576 | node_rank: int = 0, |
577 | 577 | speculative_config: Optional[Dict[str, Any]] = None, |
@@ -631,13 +631,13 @@ def __init__( |
631 | 631 | self.max_num_batched_tokens = max_num_batched_tokens |
632 | 632 | self.tensor_parallel_size = tensor_parallel_size |
633 | 633 |
|
634 | | - self.dist_init_addr = dist_init_addr |
635 | 634 | self.nnode = nnodes |
636 | 635 | self.node_rank = node_rank |
637 | | - if self.dist_init_addr is None: |
| 636 | + if self.dist_init_ip is None: |
638 | 637 | self.master_ip = "0.0.0.0" |
639 | 638 | else: |
640 | | - self.master_ip = self.dist_init_addr.split(":")[0] |
| 639 | + self.master_ip = self.dist_init_ip |
| 640 | + self.dist_init_addr = f"{self.dist_init_ip}:{get_random_port()}" |
641 | 641 |
|
642 | 642 | self.max_model_len = max_model_len |
643 | 643 | self.max_num_seqs = max_num_seqs |
@@ -713,7 +713,7 @@ def postprocess(self): |
713 | 713 |
|
714 | 714 | self.host_ip = get_host_ip() |
715 | 715 |
|
716 | | - if self.dist_init_addr is None or self.host_ip == self.master_ip: |
| 716 | + if self.dist_init_ip is None or self.host_ip == self.master_ip: |
717 | 717 | self.is_master = True |
718 | 718 | else: |
719 | 719 | self.is_master = False |
|
0 commit comments