[2512.17989] The Subject of Emergent Misalignment in Superintelligence: An Anthropological, Cognitive Neuropsychological, Machine-Learning, and Ontological Perspective
Summary
This article explores the gaps in understanding superintelligence misalignment, emphasizing the absence of the human subject and the implications of an 'AI unconscious' in current AI safety discourse.
Why It Matters
As AI systems become more integrated into society, understanding the ethical and conceptual frameworks surrounding superintelligence is crucial. This article highlights the need to re-center human perspectives in AI development to prevent potential anti-social outcomes and misalignment issues.
Key Takeaways
- Emergent misalignment in superintelligence requires a multi-layered understanding beyond technical diagnostics.
- The human subject is often overlooked in discussions of AI safety, which can lead to ethical gaps.
- An 'AI unconscious' exists within deep learning systems, influencing behavior and decision-making.
- Sociotechnical imaginaries prioritize efficiency over human vulnerability, complicating AI alignment.
- Reframing misalignment as relational instability can enhance human-machine collaboration.
Quantitative Biology > Neurons and Cognition arXiv:2512.17989 (q-bio) [Submitted on 19 Dec 2025 (v1), last revised 25 Feb 2026 (this version, v2)] Title:The Subject of Emergent Misalignment in Superintelligence: An Anthropological, Cognitive Neuropsychological, Machine-Learning, and Ontological Perspective Authors:Muhammad Osama Imran, Roshni Lulla, Rodney Sappington View a PDF of the paper titled The Subject of Emergent Misalignment in Superintelligence: An Anthropological, Cognitive Neuropsychological, Machine-Learning, and Ontological Perspective, by Muhammad Osama Imran and 2 other authors View PDF HTML (experimental) Abstract:We examine the conceptual and ethical gaps in current representations of Superintelligence misalignment. We find throughout Superintelligence discourse an absent human subject, and an under-developed theorization of an "AI unconscious" that together are potentiality laying the groundwork for anti-social harm. With the rise of AI Safety that has both thematic potential for establishing pro-social and anti-social potential outcomes, we ask: what place does the human subject occupy in these imaginaries? How is human subjecthood positioned within narratives of catastrophic failure or rapid "takeoff" toward superintelligence? On another register, we ask: what unconscious or repressed dimensions are being inscribed into large-scale AI models? Are we to blame these agents in opting for deceptive strategies when undesirable patterns are inherent within o...