It might seem bad but that's actually one of the coolest things about this new approach: it's the core model (today GPT-4o) that decides where it goes.
Here, this was a suboptimal decision by Maurice, and by default we indeed have it avoid making costly mistakes. But consider all the good decisions the agent did otherwise: navigating in all these different rooms with no prior knowledge of where anything is (just pictures it took earlier), close to the glass where Vignesh was, back to Axel, back to bed at the end...
And here's the thing: every time an LLM provider releases a new model, Maurice gets better. We haven't even started fine-tuning the agent yet but that will also improve its decisions a lot. There's many many low hanging fruits to make it able to make better decisions, and we expect that in the coming months the system will quickly get smarter and faster.
maybe for safety?