People overestimate how much it matters that ai “doesn’t have the capacity to understand it’s output”
Even if it doesn’t, is that a massive problem to overcome? There’s studies showing that if you have an ai list the potential problems with an output and then apply them to its own output it performs significantly better. Perhaps we’re just a recursive algorithm away from that.