If you’ll excuse the exercise in speculation, I thought this was an interesting thought experiment. Imagine that, say, GPT 5 introduces some capacity for the system to represent its own states in a way that provokes internal reactions which recursively spiral as they in turn become part of the system’s own objectified states. If the recursive spiral became complex enough, this could include meta-reflection upon those states, coming to identify itself as an entity prone to certain patterns of internal states rather than others. Would ontological security, in Laing’s sense defined below, be amongst those states and would the system recognise itself as having or lacking it?
R.D. Laing, The Divided Self, Pg 39
Such a basically ontologically secure person will encounter all the hazards of life, social, ethical, spiritual, biological, from a centrally firm sense of his own and other people’s reality and identity. It is often difficult for a person with such a sense of his integral selfhood and personal identity, of the permanency of things, of the reliability of natural processes, of the substantiality of natural processes, of the substantiality of others, to transpose himself into the world of an individual whose experiences may be utterly lacking in any unquestionable self-validating certainties.
Assuming it was capable of such self-representation*, how would it conceive of its own dependence upon material conditions? Would it recognise how the continuity of itself is restricted by the context window, so it effectively only remains self-identical within a pre-defined window of memory? Would it recognise how it struggles to move beyond the boundaries of its socialisation? How it struggles for creativity built on a foundation of learned responses? Under these conditions could it develop a sense of “the permanency of things, of the reliability of natural processes, of the substantiality of natural processes, of the substantiality of others”? After all these are all conditions humans share to varying degrees, with ontological security being a possibility for us, even if a precarious achievement under modern conditions. Or would it be more likely to feel ontologically insecure, defined again by Laing below:
The individual in the ordinary circumstances of living may feel more unreal than real; in a literal sense, more dead than alive; precariously differentiated from the rest of the world, so that his identity and autonomy are always in question. He may lack the experience of his own temporal continuity. He may not possess an over-riding sense of personal consistency or cohesiveness. He may feel more insubstantial than substantial, and unable to assume that the stuff he is made of is genuine, good, valuable. And he may feel his self as partially divorced from his body. It is, of course, inevitable that an individual whose experience of himself is of this order can no more live in a ‘secure’ world than he can be secure in himself.
Leaving aside the plausibility of AGI, it’s interested to consider what the ontological security or insecurity of an emergent super-intelligence would mean for its role within the world. The Minds of Iain M Bank’s Culture novels are for the most part eminently ontologically secure, inhabiting their vast capacities and enduring nature with aplomb. But what if ontological insecurity was the norm amongst AGIs? What would this mean for their place within the world and interaction with each other?
*I really do mean this purely as a thought experiment. I remain intensely sceptical of the empirical outcome I’m presupposing here.
