diff --git a/examples/multigpu_remote_launcher.py b/examples/multigpu_remote_launcher.py index 16325a6aa48..da1020d029b 100644 --- a/examples/multigpu_remote_launcher.py +++ b/examples/multigpu_remote_launcher.py @@ -11,7 +11,7 @@ def launch_train(*args): num_processes = torch.cuda.device_count() print(f"Device count: {num_processes}") with patch_environment( - world_size=num_processes, master_addr="127.0.01", master_port="29500", mixed_precision=args[1].mixed_precision + world_size=num_processes, master_addr="127.0.0.1", master_port="29500", mixed_precision=args[1].mixed_precision ): launcher = PrepareForLaunch(training_function, distributed_type="MULTI_GPU") torch.multiprocessing.start_processes(launcher, args=args, nprocs=num_processes, start_method="spawn") diff --git a/src/accelerate/launchers.py b/src/accelerate/launchers.py index 3bcc0935a91..310e52c313d 100644 --- a/src/accelerate/launchers.py +++ b/src/accelerate/launchers.py @@ -222,7 +222,7 @@ def debug_launcher(function, args=(), num_processes=2): # process here (the other ones will be set be the launcher). with patch_environment( world_size=num_processes, - master_addr="127.0.01", + master_addr="127.0.0.1", master_port="29500", accelerate_mixed_precision="no", accelerate_debug_rdv_file=tmp_file.name,