I am looking to perform some reinforcement learning training on the real robot. I was wondering:
Is it possible to ‘reset’ the real environment multiple times during a job submission? Based on the example environment it seems like it is not possible (unless we were to implement a reset mechanism ourselves). What mechanism is used to reset the robot and cube to their initial positions between jobs?
Is there a limit to the length of time a single job can run on the real robot?