Using RL, How to train multi-agents such that each agent will navigate from its initial position to goal position avoiding collisions?

4 views (last 30 days)
Let's assume there are a set of agents that are spread into 3d cartesian space. A trajectory should be generated for each agent such that if an agent would follow its trajectory while heading to the goal waypoint, no collision would happen with other agents. Any guidance to solve such a task would be highly appreciated

Answers (1)

Emmanouil Tzorakoleftherakis
Edited: Emmanouil Tzorakoleftherakis on 5 Mar 2021
It's possible that the scenario you described can be solved by training a single agent, and then "deploying" that trained agent to all uavs/uuvs in your fleet. That would make the problem easier and less expensive to train. For a 2D example, take a look at this.
  3 Comments
Emmanouil Tzorakoleftherakis
I think it's a matter of what inputs you provide to the policy and the coordinate system you use (although I was thinking the scenario where each agent has its own sensors). If you only use odometry data from all agents, I guess you could transform it to distance from each nearby agent (include heading/bearing probably) and feed all this info into the policy.
Steve Jessy
Steve Jessy on 6 Mar 2021
The coordinate system in which the agents are acting is 2D cartesian coordinate system. Yes I can access the distance from an agent to all the other agents in the space. I'd like to kindly ask you if you can provide an example/code in which the multi-agent system is trained based on odom data

Sign in to comment.

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!