I’m no Generative AI expert, but I think it’s a bit premature to say the software architecture designs in ROS are “radically challenged” by things like ChatGPT. Most robotics systems follow (at a very very high level) a sense->perceive->plan->act paradigm. I haven’t seen much evidence that modern generative AI (like chat GPT) can change the sense or act portions in a meaningful way as these are usually low level interfaces to sensors, actuators, and controllers possibly with some sophisticated data filtering on inputs and outputs. The perceive and plan portions will certainly be affected but the level of impact to architecture might not be that great. At the moment, both these pieces are already AI focused with perception being handled by various “AI” techniques like CNNs, Particle Filters, etc. The planning portion is also leveraging a wide range of AI techniques like A* path planning, behavior trees, q-learning etc.
A big portion of robotics is the need to experiment with different strategies to find the one most suited to the current domain. So most architectures already allow for a fair bit of flexibility in the underlying AI algorithms. Nav2 is a great example of this letting users select different path planners and controllers based on their needs.
I think things like chat gpt would probably be wrapped in ROS nodes just like any other AI algorithm. You could either convert inputs and outputs as needed to provide structure or maybe even get away with passing raw message content directly. (that would certainly be interesting to experiment with)
However, I think this ignores the two biggest architectural hurdle to using these tools, that is safety assurance and execution time. Chat GPT is quite slow and uses a large amount of resources to run so it’s not something you could run on a robot and maintain any type of hard realtime constraint. Similarly, the more unpredictable nature of it means using it in a safety critical system would be difficult.
1 Like