UCAHub

Understanding Cyber-Aggression through AI Use, Trust, and Personality Factors

Chapter 2 — Literature Review Mind Map

Scroll to zoom · Drag to pan · Hover a node for detail

Classic view →
Click any branch node to scroll to that section below. Hover child nodes for detail.

Chapter 2 — Literature Review

This chapter traces the theoretical foundations of cyber-aggression from classical frustration-aggression theory and the General Aggression Model through to the specific psychological and structural conditions of computer-mediated communication, moral disengagement, and the emerging role of AI in human communicative behaviour.

The central argument is that the psychology underlying online hostile behaviour is not new — it is the same frustration, threat attribution, and moral disengagement that Dollard, Berkowitz, and Bandura documented in pre-digital contexts — but that successive reductions in the cost of communication have progressively removed the structural friction that previously contained its expression.

Against this theoretical background, the chapter identifies three specific gaps that the present study addresses: the underrepresentation of adult populations in cyber-aggression research; the theoretically anomalous but empirically consistent negative relationship between anonymity and hostile response likelihood; and the absence of AI-specific constructs from existing predictive models. A further and largely unanticipated finding — the AI catharsis effect, driven specifically by the venting function of AI interaction — suggests that the relationship between AI use and online aggression may be more complex, and more theoretically interesting, than either the disinhibition or the null-effect hypothesis alone would predict.

2.1 Theoretical Foundations of Aggression

From the frustration-aggression hypothesis through Berkowitz's cognitive neoassociationistic model to the General Aggression Model — the theoretical lineage that underpins the present study's predictors.

2.1.1 The Frustration-Aggression Hypothesis

The systematic psychological study of aggression begins with Dollard et al.'s (1939) frustration-aggression hypothesis (FAH), which proposed a strict causal relationship: frustration — defined as the blocking of goal-directed behaviour — always leads to aggression, and aggression is always the product of frustration. The theoretical economy of this formulation was also its weakness. Empirical challenges accumulated quickly. Miller (1941), one of the original authors, acknowledged within two years that frustration need not produce overt aggression but may instead generate a drive toward aggression that is inhibited or displaced. Berkowitz (1962) extended this revision, demonstrating that frustration was not a sufficient condition for aggression but a necessary precursor that elevated the probability of aggressive responding under conditions that facilitated its expression. The FAH, in its revised form, bequeathed to subsequent theorising two enduring propositions: that aggression is motivationally grounded in aversive affective states, and that situational variables mediate between those states and behavioural outcomes.

2.1.2 Berkowitz's Cognitive Neoassociationistic Model

Berkowitz's (1989; 1990) cognitive neoassociationistic (CNA) model represented a fundamental reconceptualisation of the frustration-aggression relationship. The model draws on associative network theory (Anderson, 1983) to propose that any aversive stimulus — frustration, pain, heat, provocation, or even ambient discomfort — generates negative affect, which activates a network of semantically and affectively linked cognitive associations. These associations include memories of prior aversive experiences, appraisals of threat, and behavioural scripts for responding, among them both fight and flight tendencies. The specific output — aggression, avoidance, or suppression — is determined by the cognitive appraisal that follows: the individual's attribution of intent to the provocateur, their assessment of the legitimacy and likely consequences of aggressive responding, and their evaluation of available alternatives.

Two aspects of the CNA model are particularly consequential for understanding cyber-aggression. First, the model places the locus of explanation at the intersection of affect and cognition rather than in either alone. This accounts for the well-documented phenomenon of displaced aggression — hostile behaviour directed at targets other than the original source of frustration — which is theoretically problematic for models that require a direct frustration-target link. In online environments, where the original provocateur may be absent, pseudonymous, or diffuse, displaced aggression is structurally facilitated in ways that offline contexts typically are not. Second, the CNA model's emphasis on higher-order cognitive appraisal identifies the intervention points at which individual differences — including the moral disengagement mechanisms examined later in this chapter — can amplify or attenuate the translation of aversive affect into aggressive behaviour.

2.1.3 The General Aggression Model

Anderson and Bushman (2002) synthesised the preceding theoretical lineage into the General Aggression Model (GAM), the most widely applied integrative framework in contemporary aggression research. GAM organises the determinants of aggressive behaviour across three analytical levels. At the input level, personal variables (trait hostility, aggressive attitudes, prior reinforcement history) and situational variables (provocation, frustration, presence of weapons, temperature) interact to shape the internal state of the individual. At the level of internal state, the model specifies three interdependent pathways: the person's current affect, their active cognitive schemas and scripts, and their level of arousal. At the appraisal and decision level, a dual-process mechanism produces either an immediate, largely automatic response consistent with activated scripts, or a more deliberative reappraisal if the situation is sufficiently novel or the costs of automatic responding sufficiently salient.

Several theoretical extensions of GAM are directly relevant to the online context. DeWall et al. (2011) demonstrated that GAM accommodates indirect and displaced forms of aggression — including cyber-aggression, where physical distance from the target and absence of real-time feedback remove the corrective social signals that ordinarily constrain aggressive responding in GAM's appraisal stage. Allen et al. (2018), in a comprehensive narrative review, confirmed GAM's generalisability across contexts and modalities, including digital communication, while noting that the model's appraisal mechanisms were developed around assumptions of social co-presence that online environments do not satisfy.

This last point is theoretically significant for the present study. GAM's appraisal stage assumes access to immediate social feedback — visible reactions, reciprocal adjustment, real-time consequence — as inputs to the reappraisal process. In computer-mediated communication, these inputs are structurally attenuated or absent. Asynchronous delivery removes the temporal coupling between act and consequence; anonymity removes the identity-level accountability that constrains reappraisal in face-to-face contexts; physical separation from the target removes the empathic feedback that ordinarily moderates dehumanisation (Bandura, 1999).

Key references: Dollard et al. (1939); Miller (1941); Berkowitz (1962; 1989; 1990); Anderson (1983); Anderson & Bushman (2002); DeWall et al. (2011); Allen et al. (2018)

2.2 Communication Cost and Structural Friction

For most of human history, the cost of communication served as an inadvertent brake on aggressive expression. CMC removed that brake simultaneously across reach, anonymity, and cost.

The psychology underlying aggression is not new. What changes across history is not the capacity for hostile behaviour but the structural conditions that contain or release it. Hostile messages required significant effort to produce, could reach only a limited audience, and exposed the sender to social accountability. These friction-bearing constraints were not designed to suppress aggression — they were simply properties of the medium — but their effect was to contain it within manageable bounds (Shannon & Weaver, 1949; Innis, 1950).

The emergence of computer-mediated communication (CMC) altered this structure fundamentally. For the first time, a message could be composed and delivered to a global audience, anonymously, at zero marginal cost, with no requirement for physical co-presence (Postmes et al., 1998). CMC did not create new aggressive tendencies; it removed the structural friction that had previously constrained the expression of pre-existing ones.

Wiener (1948), whose foundational work on cybernetics established feedback as the central regulatory mechanism of any communication system, anticipated the consequences of its removal: without the corrective loops that social co-presence provides — immediate reaction, visible consequence, reciprocal adjustment — a communication system loses its capacity for self-regulation. What CMC stripped away was not merely social cue richness but the cybernetic feedback that ordinarily bounds aggressive expression within tolerable limits.

The current moment is arguably unprecedented in that two such revolutions — the internet and AI — are running simultaneously, and neither has reached settlement. When an individual composes a hostile message with AI assistance, vents frustration to an AI before deciding how to respond to a human provocation, or operates within a social environment where AI-mediated interaction has become the norm, the psychological dynamics that existing models were built to explain may no longer fully apply.

Key references: Shannon & Weaver (1949); Innis (1950); Wiener (1948); Postmes et al. (1998)

2.3 Cyber-Aggression: Definition, Taxonomy, and Measurement

Cyber-aggression is defined broadly and distinguished from cyberbullying. Hostile response likelihood (HRL) is adopted as the primary outcome measure following Tennakoon et al. (2024).

Cyber-aggression is broadly defined as any aggressive behaviour enacted through digital technology, including hostile messaging, harassment, public humiliation, and the deliberate spread of harmful content (Kowalski et al., 2014). It is frequently conflated with cyberbullying in both popular and academic discourse, but the distinction is theoretically consequential. Cyberbullying, as originally defined, requires three criteria: intent to harm, repetition, and a power imbalance between perpetrator and target (Olweus, 1993; Kowalski et al., 2014). Cyber-aggression carries no such requirements. A single hostile act, directed at a peer, a stranger, or a public figure, qualifies — and in the online environment, single acts can propagate at a scale that renders the repetition criterion largely redundant as a meaningful boundary condition.

Runions (2013) proposed a motivational taxonomy of cyber-aggression organised around four drivers: rage, revenge, reward, and recreation. This four-R model is useful precisely because it decouples cyber-aggression from the grievance-based framing that cyberbullying research tends to assume. Not all hostile online behaviour is retaliatory; some is instrumental, some recreational, and some is better understood as opportunistic — the product of a low-cost, low-accountability environment rather than any specific interpersonal grievance.

Measuring cyber-aggression empirically presents its own difficulties. Behavioural self-report is subject to social desirability bias, recall error, and the definitional ambiguities noted above. An alternative approach, adopted by Tennakoon et al. (2024) and followed in the present study, is to measure hostile response likelihood (HRL) — the probability that an individual would respond aggressively to a standardised online provocation — rather than actual aggressive behaviour. HRL captures behavioural intent under controlled conditions, circumventing the retrospective distortions of self-report while remaining grounded in the psychological mechanisms that produce aggressive outcomes.

Key references: Kowalski et al. (2014); Olweus (1993); Runions (2013); Tennakoon et al. (2024)

2.4 SIDE and Online Disinhibition

SIDE explains the group-level effects of anonymity; Suler explains the individual-level loosening of constraint. Together they account for why CMC amplifies pre-existing aggressive tendencies.

The Social Identity model of Deindividuation Effects (SIDE; Postmes et al., 1998) offers the first systematic account of how anonymity in CMC produces its social effects. Where earlier deindividuation theory (Zimbardo, 1969) proposed that anonymity weakens individual identity and releases impulsive, antisocial behaviour, SIDE reframes the mechanism: anonymity does not dissolve identity but shifts its salience from personal to social. When individual identity cues are absent, group identity becomes more salient, and behaviour becomes more strongly governed by group norms — whatever those norms happen to be. In a hostile online community, anonymity intensifies aggression. In a prosocial one, it intensifies cooperation. The medium does not determine the valence; it amplifies the norm.

Suler (2004) approached the same phenomenon from a different direction, identifying six psychological factors that together produce the online disinhibition effect: dissociative anonymity, invisibility, asynchronicity, solipsistic introjection, dissociative imagination, and minimisation of authority. Suler distinguished between benign disinhibition — the loosening of social constraints that allows authentic self-expression — and toxic disinhibition, in which the same mechanisms produce hostility, harassment, and cruelty directed at others.

Lapidot-Lefler and Barak (2012) subsequently isolated anonymity and invisibility as the most potent drivers of toxic disinhibition specifically, with the absence of eye contact emerging as a particularly powerful inhibition-releasing condition. Lowry et al. (2016) integrated the online disinhibition framework with SIDE and social learning theory, demonstrating that the effects operate across multiple levels simultaneously — individual psychological, group normative, and structural.

What SIDE and the online disinhibition effect provide, taken together, is a dual-level account of the same underlying phenomenon: SIDE explains why the group context under anonymity amplifies whatever norms are already operating; Suler explains why the individual feels psychologically freed to act on them.

Key references: Postmes et al. (1998); Zimbardo (1969); Suler (2004); Lapidot-Lefler & Barak (2012); Lowry et al. (2016)

2.5 Moral Disengagement

The dominant predictor in the UCA model (β = .478, p < .001). Bandura's eight mechanisms explain why ordinary people suspend the moral standards that would otherwise inhibit harmful behaviour.

UCA finding: β = .478, p < .001 — strongest predictor of HRL across all model specifications

Bandura's (1999; 2002) theory of moral disengagement addresses a question that neither SIDE nor Suler fully answers: given that the online environment creates conditions permissive of aggression, why do some individuals act on those conditions and others do not? The answer lies not in personality or pathology but in cognitive mechanism — specifically, in the set of self-regulatory processes by which ordinary people suspend the moral standards that would otherwise inhibit harmful behaviour.

Bandura identified eight such mechanisms, operating at three levels. At the level of the behaviour itself: moral justification (recasting harmful acts as serving a higher purpose), euphemistic labelling (sanitising language to obscure the nature of the act), and advantageous comparison (minimising the harm by contrasting it with worse behaviour). At the level of the agency relationship: displacement of responsibility (attributing one's actions to authority) and diffusion of responsibility (distributing accountability across a group until no individual feels culpable). At the level of the victim: dehumanisation (stripping the target of human qualities) and attribution of blame (recasting the victim as the cause of their own mistreatment). A further mechanism — disregard or distortion of consequences — operates across all three levels.

The online environment is structurally well-suited to trigger several of these mechanisms simultaneously: physical distance from the target facilitates dehumanisation; anonymity enables diffusion of responsibility; the group dynamics described by SIDE normalise moral justification within hostile communities; and the asynchronous, consequence-absent character of CMC makes disregard of consequences the path of least resistance.

The application of moral disengagement theory to the online context has been well supported empirically. Pornari and Wood (2010) demonstrated that moral disengagement predicted both peer and cyber-aggression in adolescents, with cyber-aggression showing a stronger relationship. Gaffney et al. (2019) confirmed these relationships in a meta-analytic review, establishing moral disengagement as one of the most robust predictors of cyberbullying perpetration in the literature.

Key references: Bandura (1999; 2002); Pornari & Wood (2010); Runions & Bak (2015); Gaffney et al. (2019)

2.6 AI and Human Communicative Behaviour

Habitual AI use, trust in AI systems, and AI-specific disinhibition — the three constructs absent from prior predictive models, including Tennakoon et al. (2024).

The theoretical frameworks established thus far were developed in the context of human-to-human CMC. The rapid integration of artificial intelligence into everyday communication introduces a qualitatively distinct set of conditions that existing frameworks were not designed to address.

Habitual AI use — the routine, unreflective incorporation of AI tools into daily communicative and cognitive tasks — represents a new form of technology engagement that differs from prior CMC in one structurally significant respect: the interlocutor is not human. The social feedback mechanisms that Wiener (1948) identified as the regulatory core of communication systems are, in AI-mediated interaction, simulated rather than genuine. The consequence is that habitual AI use may progressively recalibrate the user's expectations of communicative interaction, normalising a model of exchange in which the social cost of hostile expression is structurally absent.

Trust in AI systems introduces a further complication. Automation bias — the tendency to over-rely on automated outputs and defer to system recommendations — is well established in human factors research (Parasuraman & Riley, 1997). Applied to AI communication tools, elevated AI trust may reduce the individual's critical engagement with their own communicative behaviour: if the AI validates, reformulates, or simply processes hostile content without resistance, the absence of pushback may function as implicit permission. This is a form of diffusion of responsibility — one of Bandura's core moral disengagement mechanisms — operating through the human-machine rather than the human-human relationship.

AI disinhibition, as a construct, extends Suler's (2004) online disinhibition framework to the specific conditions of human-AI interaction. The dissociative imagination mechanism — Suler's observation that online actors often frame interaction as consequence-free — is potentially amplified when the other party is known to be a machine: the moral weight of the exchange is reduced by the certainty that no human is being harmed in the immediate interaction, even when the content being generated is intended for human targets.

These constructs are theoretically compelling but empirically underdeveloped. The literature on AI-specific effects on aggressive behaviour remains sparse, and the relationship between AI use, AI trust, and hostile behavioural intent has not been systematically examined prior to this study.

Key references: Wiener (1948); Parasuraman & Riley (1997); Suler (2004); Tennakoon et al. (2024)

2.7 Personality Factors

Big Five personality traits as distal predictors. Theoretically plausible, but empirically modest once situational and cognitive mechanism variables are controlled.

The relationship between personality and aggressive behaviour has a substantial empirical history, with trait-level individual differences implicated as distal predictors that shape the likelihood of aggressive responding across situational contexts. Within the Big Five framework, agreeableness has the most consistent negative relationship with aggression, with low agreeableness associated with heightened hostile attribution and reduced inhibition of aggressive response (Bettencourt et al., 2006). Neuroticism, reflecting emotional instability and negative affectivity, has also been theorised as a facilitating factor, operating through its relationship with threat sensitivity and rumination. Conscientiousness, conversely, has been proposed as a protective factor, associated with impulse regulation and normative compliance.

The empirical picture, however, is considerably less tidy than the theoretical one. Personality traits account for modest variance in aggressive behaviour in general, and their predictive utility is substantially attenuated when situational variables — provocation, anonymity, normative context — are included in the model (Bettencourt et al., 2006). In the cyber-aggression literature specifically, personality factors have rarely emerged as robust independent predictors once psychological and situational mechanisms are controlled for, suggesting that their influence is primarily mediated through the same cognitive and moral processes that more proximal predictors capture directly.

Key references: Bettencourt et al. (2006)

2.8 Emotional Regulation and the AI Catharsis Hypothesis

A novel, unanticipated finding. The venting item from the AI Disinhibition scale was the only AI construct to show a significant (negative) relationship with HRL — suggesting AI may function as a displacement channel rather than a disinhibitor.

UCA finding: Venting item r = −.201, p = .017; strengthens to r = −.223, p = .008 when moral disengagement is partialled out

A recurring assumption in both popular and academic discourse is that the expression of frustration — venting — serves a pressure-release function, reducing the likelihood of subsequent harmful behaviour by displacing affective arousal into relatively harmless expressive output. This catharsis hypothesis has a long theoretical pedigree, traceable to Freudian hydraulic models of drive and discharge, but its empirical status in the aggression literature has historically been contested. Laboratory studies of aggressive catharsis have generally found that venting increases rather than decreases subsequent aggression, particularly when the target of venting is the source of the original frustration (Bushman, 2002).

The emergence of AI as a communicative interlocutor introduces a theoretically distinct variant of this question. When the target of expressive venting is a non-human system — one that does not react, retaliate, or bear social consequences — the dynamics that typically undermine cathartic effects may not apply. The absence of a human target removes the reciprocal escalation mechanism that laboratory catharsis research has consistently identified as the driver of post-venting aggression. What remains is the affective displacement function — the relief of expressive output — without the social amplification that ordinarily follows.

Of the five items comprising the AI disinhibition scale, only one — "it is easier to vent or let off steam with AI" — showed a significant negative relationship with hostile response likelihood at the bivariate level (r = −.201, p = .017). The remaining AI disinhibition items — expressing true thoughts, saying things one would not say to a person, feeling less judged, being more direct — showed no significant relationship with HRL. This item-level specificity is theoretically important: it is not disinhibition per se that predicts reduced hostile intent, but specifically the venting function.

This interpretation is necessarily tentative. The finding is item-level, cross-sectional, and based on a single student sample. It requires replication and experimental follow-up before strong claims can be sustained. But it represents a genuinely novel empirical observation — one that opens a theoretically productive line of inquiry at the intersection of emotion regulation, human-AI interaction, and cyber-aggression research.

Key references: Bushman (2002); Suler (2004)

2.9 Literature Gaps and Research Rationale

Three specific gaps in the existing literature motivate the present study's design. Each is addressed directly by the UCA research framework.

The literature reviewed above establishes a coherent theoretical lineage from the frustration-aggression hypothesis through GAM, SIDE, and online disinhibition theory to the empirical study of cyber-aggression. Several gaps, however, remain insufficiently addressed.

First, the majority of cyber-aggression research has been conducted in the context of adolescent cyberbullying, with adult populations and non-school settings remaining comparatively understudied (Kowalski et al., 2014). The mechanisms that govern hostile online behaviour in adult populations — where power imbalances are less structurally fixed, institutional oversight is absent, and AI tool use is substantially higher — cannot be assumed to replicate those found in adolescent samples.

Second, while anonymity has been consistently theorised as a risk factor for cyber-aggression (Suler, 2004; Lapidot-Lefler & Barak, 2012; Tennakoon et al., 2024), the SIDE framework's prediction that anonymity effects are normatively conditional has rarely been tested in the context of hostile response likelihood specifically. The possibility that anonymity operates differently in AI-mediated communicative contexts — where the normative reference group may be less clearly defined and social identity cues are further attenuated — has not been examined. The present study's negative anonymity finding (β = −.232, p < .05) directly engages this gap.

Third, and most significantly, the constructs most directly relevant to the contemporary online environment — habitual AI use, trust in AI systems, and AI-specific disinhibition — are absent from existing predictive models of cyber-aggression. Tennakoon et al. (2024) provided the most methodologically proximate precedent, employing PLS-SEM to model personal and situational predictors of cyber-aggression in an adult sample, but their framework predates the mass adoption of generative AI and includes no AI-specific constructs. The theoretical case for their inclusion, established in the preceding sections, has not yet been matched by empirical investigation.

The present study addresses these gaps directly. Drawing on Tennakoon et al.'s (2024) framework as its methodological foundation, it extends the predictive model to include habitual AI use, AI trust, and AI disinhibition alongside the established predictors of moral disengagement and anonymity, in a sample of Irish university students — an adult population for whom AI tool use is both routine and consequential.

Key references: Kowalski et al. (2014); Suler (2004); Lapidot-Lefler & Barak (2012); Tennakoon et al. (2024)