Good work! Liked the moving obstacles scenario, specially.
@Roser, if I’m not mistaken, the simulation has been (at least partially) inspired by https://github.com/erlerobot/gym-gazebo environments. Is that right @Gilbert?
A very cool project would be to extend this and try a few policy gradient methods. You could even go ahead an compare it with the value iteration one you just tried (DQN). I did a while ago a tutorial comparing different methods for a simple environment but yours is indeed much cooler.