| | |
| | """ |
| | sailing runner is the main front-end to launching multi-worker |
| | training jobs with DeepSpeed. By default this uses pdsh to parallel |
| | ssh into multiple worker nodes and launch all the necessary processes |
| | per rank for training. |
| | """ |
| |
|
| | import os |
| | import sys |
| | import json |
| | import subprocess |
| | import collections |
| | import socket |
| | import signal |
| | import logging |
| |
|
| | import torch.distributed as dist |
| |
|
| |
|
| | def fetch_hostfile(hostfile_path): |
| | if not os.path.isfile(hostfile_path): |
| | print("Unable to find hostfile, will proceed with training " |
| | "with local resources only.") |
| | return None |
| | |
| | with open(hostfile_path, 'r') as fd: |
| | resource_pool = collections.OrderedDict() |
| | for line in fd.readlines(): |
| | line = line.strip() |
| | if line == '': |
| | |
| | continue |
| | try: |
| | hostname, slots = line.split() |
| | _, slot_count = slots.split("=") |
| | slot_count = int(slot_count) |
| | except ValueError as err: |
| | raise err |
| | if hostname in resource_pool: |
| | raise ValueError(f"host {hostname} is already defined") |
| | resource_pool[hostname] = slot_count |
| |
|
| | return resource_pool |
| |
|
| |
|
| | def cmd_load_hyperparam(config_path=None, format="json", encoding="utf-8"): |
| | """ |
| | shell load arguments form argparse and config file |
| | """ |
| | |
| | format = config_path.rsplit('.')[-1] |
| | with open(config_path, 'r', encoding=encoding) as f: |
| | if format == "json": |
| | config_dict = json.load(f) |
| | else: |
| | raise NameError("current format%s for hyperparam file is invalid" % |
| | format) |
| | config_cmd = [] |
| | for key in config_dict: |
| | if len(str(config_dict[key])) == 0: |
| | config_cmd.append('--' + key) |
| | else: |
| | config_cmd.append('--' + key) |
| | config_cmd.append(str(config_dict[key])) |
| | return config_cmd |
| |
|
| |
|
| | def launch_dist( |
| | env_type="DDP", |
| | num_nodes=1, |
| | gpus_per_node=1, |
| | master_addr='localhost', |
| | master_port=17500, |
| | training_script='train.py', |
| | ): |
| |
|
| | if num_nodes != 1: |
| | print("多机多卡待测试。暂不支持。") |
| | os._exit(0) |
| | if env_type == "DDP": |
| | cmd_launch = [] |
| | cmd_launch.extend([ |
| | |
| | |
| | |
| | |
| | |
| | "torchrun" |
| | |
| | ]) |
| | torch_distributed_args = [ |
| | '--nproc_per_node', |
| | str(gpus_per_node), |
| | '--nnodes', |
| | str(num_nodes), |
| | '--node_rank', |
| | str(0), |
| | '--master_addr', |
| | master_addr, |
| | '--master_port', |
| | str(master_port), |
| | ] |
| | cmd_launch.extend(torch_distributed_args) |
| | cmd_launch.append(training_script) |
| | cmd_launch.append('--not_call_launch') |
| | run_cmd = ' '.join(cmd_launch) |
| | p = subprocess.Popen(run_cmd, shell=True, preexec_fn=os.setsid) |
| | def signal_handler(signal, frame): |
| | os.killpg(os.getpgid(p.pid), 9) |
| | signal.signal(signal.SIGINT, signal_handler) |
| | p.wait() |
| | print ('finish') |
| |
|
| | else : |
| | print("不支持的env_type") |
| | os._exit(0) |
| |
|