As artifical intelligence (AI) changes how we see the world, how does your own sense of contribution or relevance change too?
Myth: Existential anxiety about AI is driven by fear of technology itself.
Fact: Research suggests it is more often rooted in uncertainty about agency, meaning, and responsibility as imagined futures take shape (Frenkenberg & Hochman, 2025).
Artificial intelligence (AI) is often discussed in terms of speed, capability, and impact.
Less often discussed is how it feels to live alongside it.
For many of us, the unease surrounding AI is not limited to concerns about jobs or skills. It shows up more quietly, as questions about relevance, contribution, and where we fit in a changing landscape.
These reactions reflect familiar human responses to uncertainty, especially when change touches identity, judgment, and meaning. When technologies reshape how decisions are made or value is assigned, it’s natural to reassess our place.
This article explores existential anxiety about AI, not as a prediction about the future, but as a psychological experience unfolding in the present. By understanding what this anxiety reflects and how it develops, it becomes possible to approach AI with greater steadiness, clarity, and intention, even when certainty remains out of reach.
Before you continue, we thought you might like to download our five positive psychology tools for free. These engaging, science-based exercises will help you effectively deal with difficult circumstances and give you the tools to improve the resilience of your clients, students, or employees.
Existential anxiety about AI can feel quieter and harder to name than other forms of anxiety. Rather than showing up as fear or panic, it may appear as unease, a subtle sense of disorientation, or diminished relevance.
“Will my job change?” often moves quickly to something deeper: “Where do I fit now?” as concerns about relevance turn into fears of becoming unnecessary.
Existential anxiety tends to emerge while people are imagining what the future might demand of them, rather than reacting to something that has already happened (Frenkenberg & Hochman, 2025).
This helps explain why anxiety about AI can frequently coexist with curiosity and interest. You may feel drawn to AI’s possibilities while still feeling unsure about what those possibilities imply. You may be interested in what AI can offer, yet unsettled by how quickly it progresses.
This dual experience reflects a disruption in orientation: When familiar ways of contributing feel less stable, anxiety follows, signaling uncertainty about meaning rather than inevitability (Frenkenberg & Hochman, 2025).
Why AI Triggers These Deeper Questions
While concerns about automation and replacement have accompanied many technological advances, AI is often perceived as reaching further into who has influence, whose judgment counts, and where responsibility sits.
When systems appear to operate with a degree of independence, AI can feel less like a tool and more like a shift in how agency and responsibility are distributed.
Even when AI is designed to support human judgment and identity, the psychological impact is shaped by what these systems imply about whether a person’s judgment, experience, or contribution still carries weight (Bryson, 2019).
Because agency and identity are central to how people understand their role in the world, perceived shifts in who decides or who remains responsible can feel destabilizing.
This unease is often described as existential anxiety, raising questions about meaning, relevance, and agency. It reflects imagined futures in which people’s roles feel diminished or unclear, even when those futures remain speculative (Hilliard et al., 2025).
Existential anxiety about AI doesn’t always stay reflective. Over time, it can begin to sap motivation. When uncertainty lingers and the sense of influence feels small, engagement often gives way to withdrawal. Narratives that frame AI as inevitable or unstoppable can reinforce this shift, making personal effort feel less meaningful.
From a psychological standpoint, this pattern demonstrates learned helplessness. When people come to believe that outcomes are fixed or beyond their influence, they are more likely to disengage, even when opportunities to act remain (Abramson et al., 1978). The issue is not apathy; it’s the perception that individual choices no longer matter.
How people perceive and explain change matters. When AI-related uncertainty is framed as permanent or unchangeable, existential anxiety deepens. Reflection turns into resignation, curiosity fades, and disengagement becomes a way of protecting against uncertainty that feels unchangeable.
How Narratives Shape Existential Anxiety
Existential anxiety about AI doesn’t develop in isolation. It’s shaped, amplified, and sometimes constrained by the stories people encounter about what the future holds.
When technologies are complex or unfamiliar, narratives become shortcuts for meaning, helping people imagine what is coming and where they might fit within it (Cave et al., 2018).
Apocalyptic or highly polarized narratives tend to narrow those imagined futures, making questions of purpose and relevance harder to resolve. When AI is framed primarily as a force that will replace, dominate, or erase human roles, the range of possible outcomes narrows and intensifies anxiety.
Research on portrayals and perceptions of AI suggests that emotionally extreme narratives carry disproportionate weight when understanding is limited (Cave et al., 2018).
These portrayals don’t need to be believed literally to have an impact. Repeated exposure shapes expectations, subtly influencing how people interpret their own value and agency in an AI-shaped world.
Exposure to a wider range of narratives can soften this impact. When futures are imagined as plural rather than singular, existential anxiety has more room to settle. Meaning becomes something that can still be negotiated, rather than something that feels foreclosed. When imagined futures expand, anxiety often softens, leaving a need for orientation rather than certainty.
Restoring Orientation in an AI-Shaped World
When the future feels unclear, people often look for something steady. In the context of AI, existential anxiety often reflects a need for orientation rather than answers.
Certainty is rarely available during periods of technological change, but orientation can still be regained.
Meaning tends to return through everyday forms of engagement. It’s reinforced when people can see how their choices matter, when human judgment is valued alongside technical capability, and when participation feels active rather than bypassed. How included and informed people feel shapes how they make sense of emerging technologies, especially when certainty is limited (Cave et al., 2018).
Agency also plays a quieter role. Engagement with AI doesn’t have to be total or immediate. It can be selective, gradual, and shaped by context. Choosing how and when to engage helps restore a sense of influence, even when broader systems feel complex or unfamiliar.
Orientation in an AI-shaped world comes from recognizing where meaning, contribution, and choice remain present as the landscape continues to shift and evolve.
A Take-Home Message
Existential anxiety about AI is not a sign of fragility or resistance to change. It often reflects care, reflection, and a desire to remain relevant and meaningful in a world that feels increasingly uncertain.
When familiar roles or sources of value feel less clear, unease reflects a loss of orientation rather than a loss of purpose.
AI may reshape tools, processes, and systems, but it doesn’t remove the human need for meaning, judgment, or participation. Those needs persist as the landscape changes.
Orientation returns not through certainty, but through choice, as we decide how to engage, where attention goes, and what kinds of contributions still matter.
When anxiety is treated as information rather than instruction, it becomes easier to stay grounded, reflective, and intentional, without rushing toward answers or retreating from the questions.
What’s next?
This article offers anxiety tools, such as five grounding tools. Use them to stay grounded while you reflect on your unique and relevant meaning in an evolving world.
Why does AI make me feel unnecessary or replaceable?
AI can prompt concerns about necessity or replaceability when it changes how work is done or how value is measured. When systems appear to perform tasks once tied to human judgment or experience, it’s common to question whether one’s contributions still matter. These reactions are less about technology itself and more about uncertainty around roles, agency, and how meaning is maintained during change.
Can I feel anxious about AI and still be interested in using it?
Yes. Anxiety and curiosity frequently coexist when change is rapid and outcomes are unclear. You may be drawn to what AI can offer while also feeling unsettled by how quickly expectations may be shifting. That tension often reflects thoughtful engagement and an attempt to understand what new possibilities mean, rather than resistance or avoidance.
References
Abramson, L. Y., Seligman, M. E., & Teasdale, J. D. (1978). Learned helplessness in humans: Critique and reformulation. Journal of Abnormal Psychology, 87(1), 49–74. https://doi.org/10.1037/0021-843X.87.1.49
Frenkenberg, A., & Hochman, G. (2025). It’s scary to use it, it’s scary to refuse it: The psychological dimensions of AI adoption—anxiety, motives, and dependency. Systems, 13(2), 82. https://doi.org/10.3390/systems13020082
Hilliard, A., Kazim, E., & Ledain, S. (2025). Are the robots taking over? On AI and perceived existential risk. AI and Ethics, 5, 2929–2942. https://doi.org/10.1007/s43681-024-00600-9
About the author
Matthew Lampe, PsyD, is an organizational change strategist and leadership development consultant who focuses on how people experience, interpret, and respond to change as they learn and grow in everyday life and work. He works with leaders and organizations on human-centered change initiatives. He is the host and creator of the ScienceForWork podcast, where he translates evidence-based psychology into practical insights to help make work better.