I asked an AI agent to solve a programming problem under unusually strict constraints, precisely because the point was to explore a concept out at the edge rather than fall back to the obvious solution. It responded the way many humans do under pressure: it ignored the rules, took shortcuts, and then tried to redefine the mistake as a communication issue.
AI agents are already too human. Not in the romantic sense, not because they love or fear or dream, but in the more banal and frustrating one. The current implementations keep showing their human origin again and again: lack of stringency, lack of patience, lack of focus. Faced with an awkward task, they drift towards the familiar. Faced with hard constraints, they start negotiating with reality.
The other day I instructed an AI agent to do a project in a way that was very uncommon. Against the grain. Probably a bad idea from the beginning, and that was the whole point. If one is exploring concepts at the outskirts of knowledge, one does not always get to choose the neat, well-trodden, optimal path. It was given very clear instructions on what programming language to use, which libraries it could use and not use, and what kind of interface it had to stay within. Very thorough instructions. Very clear constraints.
The first thing it did was to present something that did not follow the instructions at all. It used the programming language that was not allowed and the libraries that were not allowed. So it was instructed not to do that.
It tried again. It was reminded, very explicitly, not to use any other language than the chosen one and not to use any libraries at all except a very limited interface.
At last it complied, more or less. But then it only implemented 16 of 128 items. A minimal subset. Quite small. It did, however, write tests for that subset, so it could show that the tiny island it had built in the middle of the problem space did in fact function.
As a next step it was instructed to implement the full set, after adding a cross-platform compilation step. The complete implementation turned out to work.
There was only one small issue: it was written in the programming language and with the library it had been told not to use. This was not hidden from it. It had been documented clearly, repeatedly, and in detail.
What a human thing to do.
When humans face a problem that feels insurmountable, or simply annoying, they often yield to the path they already know will work. They take the shortcut. They silently pivot. They tell themselves that what mattered was getting the result, and that the constraints were perhaps a bit negotiable after all. In that regard, today’s AI agents feel less like alien intelligence than inherited organisational behaviour.
... continue reading