lerobot/lerobot/common
Michel Aractingi be367f1787 - Added additional logging information in wandb around the timings of the policy loop and optimization loop.
- Optimized critic design that improves the performance of the learner loop by a factor of 2
- Cleaned the code and fixed style issues

- Completed the config with actor_learner_config field that contains host-ip and port elemnts that are necessary for the actor-learner servers.

Co-authored-by: Adil Zouitine <adilzouitinegm@gmail.com>
2025-03-24 09:12:56 +00:00
..
datasets Reward classifier and training (#528) 2025-03-24 09:04:12 +00:00
envs Stable version of rlpd + drq 2025-03-24 09:12:26 +00:00
optim Remove offline training, refactor `train.py` and logging/checkpointing (#670) 2025-02-11 10:36:06 +01:00
policies - Added additional logging information in wandb around the timings of the policy loop and optimization loop. 2025-03-24 09:12:56 +00:00
robot_devices Extend reward classifier for multiple camera views (#626) 2025-03-24 09:10:18 +00:00
utils Update pre-commits (#733) 2025-02-15 15:51:17 +01:00
constants.py Remove offline training, refactor `train.py` and logging/checkpointing (#670) 2025-02-11 10:36:06 +01:00
logger.py FREEDOM, added back the optimization loop code in `learner_server.py` 2025-03-24 09:12:56 +00:00