Replies: 10 comments 33 replies
This comment was marked as off-topic.
This comment was marked as off-topic.
-
Hi, we are trying to reproduce the find and avoid in our world. We are not able to transmit the created message from epuck to supervisor. Do you have a quick document on this program and how to use the json scripts for emitter/receiver. |
Beta Was this translation helpful? Give feedback.
-
Hi, I am writing a paper with the goal of implementing the stable_baselines algorithms in a Webots environment (using the IRB4600). The first step for me is now this tutorial. INFO: robotSupervisorController: Starting controller: python.exe -u robotSupervisorController.py Another question: However, I cannot implement this at the same time as the RobotSupervisor. Is there another solution for this? Is there already an example project for this? Many thanks |
Beta Was this translation helpful? Give feedback.
-
Hi, great tutorial! Are there definitions for some of the methods, like |
Beta Was this translation helpful? Give feedback.
-
Hi, thanks for the awesome tutorial! However, when I run the tutorial I got the error, ModuleNotFoundError: No module named 'controller'. Would kindly require your assistance on this. The error is from supervisor_env.py. |
Beta Was this translation helpful? Give feedback.
-
Great, crystal clear tutorial! Many thanks indeed for your work. |
Beta Was this translation helpful? Give feedback.
-
Awesome, thank you so much!
…On Thu, 21 Mar 2024 at 21:25, Kostas Tsampazis ***@***.***> wrote:
Hello @ChrisSim01 <https://github.com/ChrisSim01>, thank your for your
kind words, very happy to hear that the tutorial proved useful for you! 😄
As for your first question, you can find the agent save/load methods here
<https://github.com/aidudezzz/deepbots-tutorials/blob/48f76ecc9791d6b9cfc415623b4ad30d208efc9c/robotSupervisorSchemeTutorial/full_project/controllers/robot_supervisor_controller/PPO_agent.py#L93-L113>,
which are basically saving the actor/critic neural nets. They can be called
whenever you wish, per some number of episodes, every episode or at the end
of training to save the agent.
As long as you saved the models once, you can load them after the agent is
initialized here
<https://github.com/aidudezzz/deepbots-tutorials/blob/48f76ecc9791d6b9cfc415623b4ad30d208efc9c/robotSupervisorSchemeTutorial/full_project/controllers/robot_supervisor_controller/robot_supervisor_controller.py#L95>.
To continue training on a different task with the same agent, you can
modify your environment, webots worlds, etc. and just load the agent as
discussed.
Hope this helps, let me know if you have other questions!
—
Reply to this email directly, view it on GitHub
<#12 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AH4Z5XTCRSLFN6NZYL6KAK3YZNF3ZAVCNFSM4YCUZX52U5DIOJSWCZC7NNSXTOKENFZWG5LTONUW63SDN5WW2ZLOOQ5TQOBXGEYTENY>
.
You are receiving this because you were mentioned.Message ID:
***@***.***
com>
|
Beta Was this translation helpful? Give feedback.
-
Hi again,
You could make an online course out of this tutorial via Udemy.com and make
money from it!
Just an idea,
Best regards,
Chris
…On Thu, 21 Mar 2024 at 21:25, Kostas Tsampazis ***@***.***> wrote:
Hello @ChrisSim01 <https://github.com/ChrisSim01>, thank your for your
kind words, very happy to hear that the tutorial proved useful for you! 😄
As for your first question, you can find the agent save/load methods here
<https://github.com/aidudezzz/deepbots-tutorials/blob/48f76ecc9791d6b9cfc415623b4ad30d208efc9c/robotSupervisorSchemeTutorial/full_project/controllers/robot_supervisor_controller/PPO_agent.py#L93-L113>,
which are basically saving the actor/critic neural nets. They can be called
whenever you wish, per some number of episodes, every episode or at the end
of training to save the agent.
As long as you saved the models once, you can load them after the agent is
initialized here
<https://github.com/aidudezzz/deepbots-tutorials/blob/48f76ecc9791d6b9cfc415623b4ad30d208efc9c/robotSupervisorSchemeTutorial/full_project/controllers/robot_supervisor_controller/robot_supervisor_controller.py#L95>.
To continue training on a different task with the same agent, you can
modify your environment, webots worlds, etc. and just load the agent as
discussed.
Hope this helps, let me know if you have other questions!
—
Reply to this email directly, view it on GitHub
<#12 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AH4Z5XTCRSLFN6NZYL6KAK3YZNF3ZAVCNFSM4YCUZX52U5DIOJSWCZC7NNSXTOKENFZWG5LTONUW63SDN5WW2ZLOOQ5TQOBXGEYTENY>
.
You are receiving this because you were mentioned.Message ID:
***@***.***
com>
|
Beta Was this translation helpful? Give feedback.
-
Hi again Kostas,
I do have one further question please: is there any way to 'lock', or
encapsulate, a neural network module once it is learned, so that it is used
by, but not changed by further neural network training? This would allow
tasks to be broken down into subtasks, and chaining of the subtasks
together I think.
I'm thinking in the style of
https://nn.cs.utexas.edu/downloads/papers/lessin.gecco13.pdf
Best regards,
Chris
…On Thu, 21 Mar 2024 at 21:25, Kostas Tsampazis ***@***.***> wrote:
Hello @ChrisSim01 <https://github.com/ChrisSim01>, thank your for your
kind words, very happy to hear that the tutorial proved useful for you! 😄
As for your first question, you can find the agent save/load methods here
<https://github.com/aidudezzz/deepbots-tutorials/blob/48f76ecc9791d6b9cfc415623b4ad30d208efc9c/robotSupervisorSchemeTutorial/full_project/controllers/robot_supervisor_controller/PPO_agent.py#L93-L113>,
which are basically saving the actor/critic neural nets. They can be called
whenever you wish, per some number of episodes, every episode or at the end
of training to save the agent.
As long as you saved the models once, you can load them after the agent is
initialized here
<https://github.com/aidudezzz/deepbots-tutorials/blob/48f76ecc9791d6b9cfc415623b4ad30d208efc9c/robotSupervisorSchemeTutorial/full_project/controllers/robot_supervisor_controller/robot_supervisor_controller.py#L95>.
To continue training on a different task with the same agent, you can
modify your environment, webots worlds, etc. and just load the agent as
discussed.
Hope this helps, let me know if you have other questions!
—
Reply to this email directly, view it on GitHub
<#12 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AH4Z5XTCRSLFN6NZYL6KAK3YZNF3ZAVCNFSM4YCUZX52U5DIOJSWCZC7NNSXTOKENFZWG5LTONUW63SDN5WW2ZLOOQ5TQOBXGEYTENY>
.
You are receiving this because you were mentioned.Message ID:
***@***.***
com>
|
Beta Was this translation helpful? Give feedback.
-
Thanks again Kostas. I'm on a steep learning curve!
…On Fri, 22 Mar 2024 at 21:50, Kostas Tsampazis ***@***.***> wrote:
Very interesting paper!
I think that what you are describing can be achieved by "freezing"
specific layers of your networks, something along the lines of this
<https://discuss.pytorch.org/t/how-the-pytorch-freeze-network-in-some-layers-only-the-rest-of-the-training/7088>
.
—
Reply to this email directly, view it on GitHub
<#12 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AH4Z5XV3FCKQPHBAY2UESVDYZSRTRAVCNFSM4YCUZX52U5DIOJSWCZC7NNSXTOKENFZWG5LTONUW63SDN5WW2ZLOOQ5TQOBYGI4TANI>
.
You are receiving this because you were mentioned.Message ID:
***@***.***
com>
|
Beta Was this translation helpful? Give feedback.
-
Link to the tutorial.
Beta Was this translation helpful? Give feedback.
All reactions