diff --git a/README.md b/README.md index 969d2a5..88b720d 100644 --- a/README.md +++ b/README.md @@ -70,13 +70,13 @@ We can distribute and run this function (e.g. on 2 machines x 2 GPUs) using **`t ```python import logging -logging.basicConfig(level=logging.INFO) - import torchrunx +logging.basicConfig(level=logging.INFO) + launcher = torchrunx.Launcher( hostnames = ["localhost", "second_machine"], # or IP addresses - workers_per_host = 2 # e.g. number of GPUs per host + workers_per_host = "gpu" # default, or just: 2 ) results = launcher.run( @@ -93,7 +93,7 @@ trained_model: nn.Module = results.rank(0) # or: results.index(hostname="localhost", local_rank=0) # and continue your script -torch.save(trained_model.state_dict(), "output/model.pth") +torch.save(trained_model.state_dict(), "outputs/model.pth") ``` **See more examples where we fine-tune LLMs using:**