diff --git a/README.md b/README.md index 4212153f..db573121 100644 --- a/README.md +++ b/README.md @@ -160,14 +160,14 @@ See `python lerobot/scripts/eval.py --help` for more instructions. Check out [example 3](./examples/3_train_policy.py) that illustrates how to start training a model. -In general, you can use our training script to easily train any policy. To use wandb for logging training and evaluation curves, make sure you ran `wandb login`, and enable it in the configuration. Here is an example of training the ACT policy on trajectories collected by humans on the Aloha simulation environment for the insertion task: +In general, you can use our training script to easily train any policy. Here is an example of training the ACT policy on trajectories collected by humans on the Aloha simulation environment for the insertion task: + ```bash python lerobot/scripts/train.py \ policy=act \ env=aloha \ env.task=AlohaInsertion-v0 \ dataset_repo_id=lerobot/aloha_sim_insertion_human \ - wandb.enable=true ``` The experiment directory is automatically generated and will show up in yellow in your terminal. It looks like `outputs/train/2024-05-05/20-21-12_aloha_act_default`. You can manually specify an experiment directory by adding this argument to the `train.py` python command: @@ -175,12 +175,18 @@ The experiment directory is automatically generated and will show up in yellow i hydra.run.dir=your/new/experiment/dir ``` -A link to the wandb logs for the run will also show up in yellow in your terminal. Here is an example of logs from wandb: +To use wandb for logging training and evaluation curves, make sure you've run `wandb login` as a one-time setup step. Then, when running the training command above, enable WandB in the configuration by adding: + +```bash + wandb.enable=true +``` + +A link to the wandb logs for the run will also show up in yellow in your terminal. Here is an example of what they look like in your browser: + ![](media/wandb.png) Note: For efficiency, during training every checkpoint is evaluated on a low number of episodes. You may use `eval.n_episodes=500` to evaluate on more episodes than the default. Or, after training, you may want to re-evaluate your best checkpoints on more episodes or change the evaluation settings. See `python lerobot/scripts/eval.py --help` for more instructions. - ## Contribute If you would like to contribute to 🤗 LeRobot, please check out our [contribution guide](https://github.com/huggingface/lerobot/blob/main/CONTRIBUTING.md).