Search Unity

Navigating a dynamic environment

Discussion in 'ML-Agents' started by WardPeeters, Apr 6, 2020.

  1. WardPeeters

    WardPeeters

    Joined:
    Mar 19, 2015
    Posts:
    25
    Hello, I am creating an experiment in which the agent has to learn to navigate (move) and avoid obstacles. The environment is generated randomly so that the objects to avoid are always in a different location and with a different size. I will give the agent raycast observations that collide with the obstacles and a goal that it has to reach.

    Goal:

    When a user creates an environment the trained agent should be able to navigate in it and avoid obstacles. But the environment that the user creates could be different from the procedural environment in which the agent was trained. For example bigger or smaller, more or less obstacles, ... The user would also be able to change the environment and obstacles as the trained agent is navigating it.

    Now my questions:

    Will the agent learn to actually avoid obstacles as he moves around the environment to find the goals? Or will it start to remember different generated environment setups? (which are not unlimited)

    As far as I observed all the examples have static environments only the wall jump one changes the ramp position but since this environment is so small I could imagine the agent would remember all the different locations.