Hi,
First of all thanks for sharing this work. I was reading the paper related to this and was curious to see how you managed to transfer the learned policy to the actual real-world robot.
I have been trying to do something similar for a Sawyer robot but I am facing some issues. Perhaps if you could share the code you used to transfer the policy to the Stretch RE1, I can take some hints from that.
Thanks in advance!!!