From moral panic to pragmatic governance: reframing AI’s societal impacts in employment, education, and ethics
- Open Access
- 16-02-2026
- Open Forum
Activate our intelligent search to find suitable subject content or patents.
Select sections of text to find matching patents with Artificial Intelligence. powered by
Select sections of text to find additional relevant content using AI-assisted search. powered by (Link opens in a new window)
Abstract
1 Introduction
Public debates about artificial intelligence (AI) oscillate between alarm and reassurance. This pattern aligns with classic analyses of moral panic—episodes of amplified public anxiety in which new technologies are cast as threats to social order through cycles of claims-making, media escalation, and calls for control (Cohen 2011; Critcher 2003; Hier 2008; Goode and Ben-Yehuda 1994). Digital technologies have repeatedly been processed through this repertoire, from videogames to social media (Dubèl et al. 2025; Marwick and Lewis 2017). Yet contemporary AI discourse reflects a distinctive combination of long-horizon uncertainties and evidence-based concerns about bias, opacity, and governance failures. These dual dynamics—panic signals and tractable problems—form the starting point for this paper.
Recent media coverage intensifies this oscillation between alarm and assurance. Headlines warning that “AI Is Taking Over the World” (The Guardian 2020), predicting that machines “Could Destroy Jobs” (The Times 2021), or framing AI as an “extinction-level threat” (CNN 2024) help constitute a discursive environment saturated with existential and socio-economic anxiety. The sociology of expectations shows how such promissory and catastrophic narratives co-produce hype cycles that shape investment priorities and policy agendas well before technical capacities are settled (Borup et al. 2006; Brown and Michael 2003). Communication research further demonstrates how “strong AI” imaginaries—sentience, superintelligence—amplify utopian and apocalyptic projections, while “weak AI” imaginaries sustain more grounded debates about capabilities, limits, and social embedding (Bory et al. 2025). These framings matter: positioning AI as an existential agent pulls attention toward speculative futures, whereas framing it as an infrastructural tool redirects scrutiny toward the actionable, present-day mechanisms through which risks and responsibilities are produced (Chubb et al. 2024).
Advertisement
Against this backdrop, we adopt a pragmatist risk-governance orientation. Rather than casting AI as either existential threat or technological salvation, this approach centers on observable failure surfaces—statistical bias, data lineage and provenance gaps, non-reproducible evaluation, unclear responsibility chains, and assurance dependence—and links each to concrete mechanisms of remediation. This stance aligns with risk-society accounts that highlight institutional reflexivity and “organised irresponsibility” in technologically complex systems (Beck 1992), with regulatory-science work showing how evidence infrastructures render contested technologies governable (Jasanoff 2003), and with analyses of the audit society, where checks lose meaning unless tied to performance evidence (Power 1997). In AI, a substantial toolset now translates principles into verifiable controls: model cards and datasheets (Gebru et al. 2021; Mitchell et al. 2019), algorithmic impact assessments and third-party audits (Raji et al. 2020), risk-management frameworks (NIST 2023), management-system standards (ISO/IEC 42001:2023), and risk-based regulation (EU AI Act 2024). While imperfect, these instruments anchor concerns in testable claims, measurable thresholds, and documentary trails. A pragmatic orientation, therefore, treats AI’s risks as identifiable, diagnosable, and governable, providing the foundation for the governance architecture developed later in the paper.
Although some fears—such as AI triggering human extinction—are clearly exaggerated, they draw attention to the wider societal and ethical stakes of rapid, poorly governed technological change (Bentley and Evans 2025; Hirsch-Kreinsen 2024; Tegmark 2017). Such anxieties echo long-standing patterns in which innovation is framed as both threat and opportunity. The crucial distinction, however, lies in whether concerns are channeled into actionable, evidence-based analysis or amplified into speculative, apocalyptic narratives. A pragmatic orientation focuses on the former: by articulating risks in tractable, empirically grounded terms rather than through moral-panic framings, it becomes possible to design targeted governance measures and policies capable of addressing real failure points.
In this article, moral panic refers to a discourse pattern characterized by disproportionate threat construction and volatile amplification, whereas pragmatism denotes a program for measurable risk remediation (see Table 1). Building on this distinction, we map widely voiced worries in employment, education, and ethics—three domains where AI anxieties most often crystallize—onto two elements: (a) their relevant empirical evidence base and (b) the governance controls capable of materially improving outcomes. This move responds to the tendency for recent practice-oriented literatures to remain under-recognized in broader public debate. We then develop a collaborative governance architecture grounded in the Neo-Triple Helix, coordinating universities, industry, government, standards bodies, and civil society to translate ethical principles into auditable practice while addressing power asymmetries and risks of assurance capture (Cai 2022; DeNardis 2014; Raji et al. 2020; Srnicek 2017).
Table 1
Differentiating moral panic and pragmatism in AI discourse
Aspect | Framing AI as a moral panic | Framing AI as a challenge |
|---|---|---|
Definition | An exaggerated media and societal reaction characterized by fear and speculative narratives about the implications of AI | Identified, evidence-based issues arising from the implementation and integration of AI systems |
Focus | Speculative fears and uncertainties, often amplified by sensational media discourse | Observable shortcomings in AI systems and their societal impacts |
Examples of concerns | Problem orientated only: Employment: Automation and job displacement (e.g., autonomous vehicles replacing drivers, automated checkout systems replacing cashiers) Unequal opportunities as AI creates high-skill jobs but leaves behind low-skill workers without reskilling opportunities Education: Inequitable access to AI-powered educational tools (e.g., Coursera, Duolingo) widening the digital divide - Bias in predictive analytics disadvantaging students from underrepresented backgrounds due to biased training data Ethics: Algorithmic bias replicating existing inequalities (e.g., facial recognition misidentifying individuals based on race) - Lack of transparency and accountability in decision-making (e.g., black-box systems in autonomous vehicle errors) - Privacy and surveillance risks (e.g., facial recognition used for mass surveillance affecting marginalized communities) | Recognizing problems but with a solution-orientated approach: Employment: Reskilling workers to adapt to AI-driven job demands (e.g., training programs for transitioning from routine jobs to tech roles) Ensuring equitable access to high-quality job opportunities created by AI advancements Education: Closing the digital divide to ensure underserved regions benefit from AI-powered tools (e.g., adaptive learning) Developing fairness-aware machine learning to prevent biases in educational analytics Ethics: Implementing fairness-aware algorithms to mitigate bias during model development Enhancing explainability in AI systems to ensure transparency and accountability Strengthening regulations to prevent data misuse and uphold privacy rights |
Framing | Dramatic and often lacking grounding in evidence-based risks | Rooted in evidence and framed to guide actionable, practical solutions |
Outcomes | Heightened fear, resistance to AI adoption, and misinformed public debates | Informed discussions, targeted policy interventions, and balanced AI adoption |
Approach to solutions | Focus on halting or heavily regulating AI to prevent perceived risks | Focus on addressing specific, identified issues through regulation, innovation, and collaboration |
Role of media | Amplifies fears and uncertainties through sensational headlines and speculative reporting | Highlights real issues while promoting informed and balanced discourse |
Governance and policy | Distrust in different stakeholders including government, academia, industry, etc., and their ability to keep pace with AI innovation | Proactive regulatory measures, such as UNESCO’s AI Ethics Guidelines, foster responsible innovation (Floridi and Cowls 2019) The Quadruple Helix Model strengthens governance by integrating academic research (academia), policy creation (government), technological advances (industry), and public needs (society) |
Impact on public perception | May lead to fear-driven skepticism, hindering constructive engagement with AI | Encourages constructive dialog and proactive solutions to maximize AI’s benefits |
The article proceeds in four steps. First, it traces AI’s evolution into an upstream infrastructure of models, data, and compute, showing why accountability now hinges on lineage, reproducible evaluation, and provenance. Second, it examines social consequences across employment, education, and ethics, mapping widely voiced concerns onto the empirical evidence and the governance levers that can address them. Third, it develops a Neo-Triple Helix coordination framework—linking universities, industry, government, standards bodies, and civil society through participation mechanisms, assurance guarantees, and procurement levers—while specifying the scope conditions under which it can operate. Finally, it outlines a measurement agenda to support cross-site evaluation and concludes by reframing moral panic as an impetus for actionable, portable governance capable of supporting accountable AI development.
Advertisement
2 From moral panic to pragmatic uptake: a brief sociological history of AI
Public debate often treats AI’s sudden visibility—first with deep learning breakthroughs after 2012 and then the rapid diffusion of large-scale generative models from 2022 onward—as if it were an entirely new phenomenon. However, AI is a long-running project with distinct technical phases and institutional arrangements, each with different loci of authority, characteristic failure surfaces, and governance needs (McCorduck 2004; Nilsson 2010; Russell and Norvig 2021). Reconstructing this history is central to a pragmatic stance: it anchors claims about risk and benefit in how AI systems have actually been built and embedded over time, rather than in moral panic headlines.
The mid-twentieth-century imaginary of “thinking machines” mixed public fascination with apprehension about autonomy and control. Turing’s operational proposal for studying machine intelligence (Turing 1950) prompted both enthusiasm and skepticism; critics questioned whether symbolic manipulation could scale to robust intelligence (Dreyfus 1972). Pragmatically, early systems showed narrow but real value: ELIZA’s pattern matching (Weizenbaum 1966), SHRDLU’s language-guided manipulation (Winograd 1972), and Shakey’s integrated perception and action (Nilsson 2010). The moral-panic repertoire (fear of dehumanization, loss of control) traveled alongside practical progress in constrained domains—an early instance of what the sociology of expectations calls the co-production of hype and skepticism (Borup et al. 2006; Brown and Michael 2003). The 1970s–1980s expert-systems wave vested epistemic authority in hand-crafted rules and knowledge engineering. MYCIN and XCON demonstrated concrete benefits in well-bounded settings (McDermott 1982; Shortliffe 1976), even as brittleness and maintenance costs limited generalization (Russell and Norvig 2021). Panics about machines replacing human judgment were countered pragmatically by domain scoping, human-in-the-loop protocols, and liability rules tailored to decision support rather than autonomy—a pattern that still informs safety–critical deployments.
From the 1990s, statistical learning and increased compute/data yielded milestones—Deep Blue’s chess victory (Campbell et al. 2002), and later deep learning leaps in vision and games (Krizhevsky, Sutskever and Hinton 2012; Silver et al. 2016)—and underwrote routine applications such as spam filtering, recommendation, and navigation (Jordan and Mitchell 2015). Here, moral-panic narratives shifted toward surveillance, opacity, and manipulation, as data-driven services scaled (Crawford 2021; Zuboff 2019). Pragmatic countermeasures emerged: privacy and data-protection law, fairness and accountability research, and the first templates for auditing algorithmic systems (Barocas, Hardt and Narayanan 2019; Pasquale 2015).
Transformer architectures and massive pre-training turned AI from point solutions into an inference infrastructure whose defaults (training data, architectures, alignment layers, API exposure) travel across tasks and sectors (Bommasani et al. 2021; Vaswani et al. 2017). Generative models extend this to high-fidelity text, image, and audio synthesis (Goodfellow et al. 2014; Ho et al. 2020; Rombach, Blattmann, Lorenz, Esser and Ommer 2022). Sociologically, authority migrates upstream—from local application teams to model/compute providers and benchmark communities—so that lineage, reproducible evaluation, and content provenance become central to accountability (Bender et al. 2021; Plantin et al. 2018). Moral-panic cycles amplify fears of sentience or extinction; a pragmatic response focuses on observable failure surfaces: statistical bias and distribution shift, opaque data provenance, non-deterministic outputs, mis/disinformation risks, and unclear liability chains (Burrell 2016; Mitchell et al. 2019; Raji et al. 2020; Weidinger et al. 2022).
Seen through this history, AI’s current controversies are less about an unprecedented rupture than about governing an evolving infrastructure that pre-formats downstream practice. In workplaces, this means model defaults and API policies shape task allocation and oversight long before managers encounter specific tools; pragmatism thus asks for human–AI teaming protocols (escalation thresholds, override rights), targeted upskilling, and performance evidence linking productivity claims to wages and job quality (Acemoglu and Restrepo 2020; Brynjolfsson et al. 2023; Noy and Zhang 2023). In education, the same infrastructure can scaffold formative feedback yet threatens assessment integrity unless guardrails, disclosure norms, and provenance are built into learning management systems (Holmes et al. 2019; Kasneci et al. 2023; UNESCO 2023). In ethics and governance, pragmatism translates principles into verifiable controls: datasheets and model cards, open benchmarks with reproducible evaluations, incident databases with time-to-disclosure metrics, and risk-based obligations codified in frameworks such as NIST AI RMF 1.0, ISO/IEC 42001, and the EU AI Act (EU 2024; ISO/IEC 2023; Mitchell et al. 2019; NIST 2023). Throughout, moral-panic frames (from folk-devil narratives to extinction talk) remain sociologically important because they steer attention and investment, but their analytic value improves when paired with tractable mechanisms and measurable outcomes (Brown and Michael 2003; Cohen 2011; Goode and Ben-Yehuda 1994).
This historical reconstruction embeds the panic ↔ pragmatism contrast within concrete sociotechnical change (see Table —From Fear to Action: Differentiating Between AI's Moral Panic and Real Challenges). It sets the stage for the sections that follow by showing why today’s debates about employment, education, and ethics must grapple with AI as an upstream inference infrastructure: one that lowers the cost of analysis and synthesis while redistributing decision rights and responsibility, and therefore demands governance that prioritizes lineage, evaluation reproducibility, provenance, and meaningful avenues for contestation and redress (Amann et al. 2020; Floridi and Cowls 2019; Raji et al. 2020).
3 Employment: from automation alarm to task-level evidence
Public discourse about AI and employment often shifts from discussion of potential benefits to a mood of moral panic, with headlines that emphasize mass displacement and existential threats. Early task-based projections, such as Frey and Osborne’s estimates of automatable employment and related scenario planning (Manyika et al. 2017; Spencer 2025), helped focus attention on risk but also fostered a simplified narrative that “robots will take all jobs.” A more measured reading of the accumulating evidence suggests that impacts are nuanced and actionable: effects vary across tasks, sectors, and workers; short-run productivity gains are evident for certain activities; and the net effects on employment and wages depend on complementary investments, institutional settings, and how algorithmic management is governed.
Historical perspectives show how past technological revolutions have caused disruption and job upheaval, while also giving rise to new industries, new roles, and new opportunities. The late twentieth century technological shift—driven by automation, computing, and digitalization—illustrated how gains in efficiency could accompany a decline in traditional labor but also catalyze the growth of knowledge-based sectors and digital economies. Industrial automation reduced the demand for manual labor in fields like shipbuilding, mining, and heavy manufacturing (Harvey 1989), while computer-aided design and manufacturing increased production capabilities and precision. The social costs—long-term unemployment, job insecurity, and weakened career pathways—were real, but so too were the opportunities for new work and new skill sets.
It can be claimed that AI presents a similar duality. It can pose genuine risks to existing job structures, yet it also holds the potential to drive new opportunities, contingent on policy choices, workforce adaptation, and accessible reskilling. Randomized and field studies consistently show that generative and assistive AI can raise measured productivity for routine, well-specified cognitive work, with the largest gains accruing to less experienced workers and thus reducing some gaps in performance (Brynjolfsson et al. 2023; Noy and Zhang 2023). In software engineering, deployments of code assistants can shorten completion times and improve success rates on clearly defined problems, though they tend to shift effort toward review and verification when tasks are open-ended (Peng et al. 2023). Analyses that examine job exposure through task content and model capabilities suggest that many professional and administrative roles are augmented rather than replaced, with changes concentrated in task mixes and required skills (Eloundou et al. 2023; Felten et al. 2021).
On a macro level, caution is warranted against equating productivity gains with net employment growth or higher wages. Historical analyses of automation indicate that whether technology substitutes for or complements labor depends on task reallocation, demand elasticities, and institutional responses (Acemoglu and Restrepo 2020). Cross-country assessments find that generative AI is more likely to reconfigure jobs than to eliminate them outright, with higher exposure in clerical functions and uneven risks by gender and income groups (ILO 2023; OECD 2023; World Bank 2023). In short, augmentation appears most reliable when tasks are templated and evaluable; in open-ended or safety–critical contexts, error risk and oversight costs tend to rise. These employment impacts do not arise from individual AI tools alone. They reflect AI’s role as an inference infrastructure, where upstream choices in data and model design shape how tasks are assigned, monitored, and valued. As a result, job-quality risks appear long before organizations deploy a specific system.
Beyond headcounts, AI influences how work is coordinated and evaluated. The literature on algorithmic management shows that it can intensify monitoring, compress discretion, and shift some risk onto workers in sectors such as logistics, ride hailing, and content moderation, with these dynamics gradually expanding into “white collar” settings through dashboards, automated performance targets, and AI-mediated scheduling. The central policy question, therefore, is not only about displacement but about the governance of task delegation, accountability, and job quality in AI-enabled organizations.
To place these concerns in context, apocalyptic claims about “the end of work” tend to generalize from worst-case scenarios and overlook the levers that shape outcomes. A more pragmatic view emphasizes contingency: technologies can lead to different employment trajectories depending on training systems, bargaining institutions, procurement rules, and competition in the model/compute layer (Bessen 2019; Kenney and Zysman 2016). When organizations invest in complementary skills and redesign workflows to support human–AI teaming, productivity gains are more likely to translate into better jobs; conversely, pursuing cost-only automation with weak guardrails increases the likelihood of displacement and erosion of job quality (Acemoglu and Restrepo 2020).
The evidence base, thus, points toward an actionable agenda rather than a narrative of inevitability. Concrete steps include designing human–AI teaming with explicit escalation thresholds, override rights, and error budgets tied to task criticality, while measuring verification workload and quality drift throughput (Brynjolfsson et al. 2023; Peng et al. 2023). It also suggests targeted upskilling and clear pathways for mid-career workers, aligning curricula with AI-complementary skills such as data literacy, prompt-driven workflow design, and critical review (ILO 2023; OECD 2023). Safeguards for job quality in algorithmic management are warranted, including transparency about data inputs and performance metrics, audit trails for consequential decisions, and ensuring worker voice in deployment and evaluation (Kellogg et al. 2020; Rosenblat and Stark 2016). Attention to competition and rent sharing is also important, with monitoring of model/compute-layer concentration to ensure productivity gains do not decouple from wage growth and downstream investment (Bessen 2019; Kenney and Zysman 2016).
Ultimately, the best available evidence does not support a narrative of inevitable widespread redundancy, nor does it justify complacency. The likely trajectory is conditional: outcomes depend on how tasks are recomposed, who controls capability exposure and performance metrics, and whether institutions translate short-run productivity gains into enduring, broad-based improvements.
4 Education: from cheating fears to measured learning gains
Public debate around AI in education often swings between concerns about cheating, surveillance, and deskilling, and optimistic claims of frictionless personalization. This moral panic register can obscure a substantial body of empirical and practice-oriented work that has begun to specify where, how, and under what conditions AI can support learning—and where it can pose risks (Reich 2020; Selwyn 2019; Wang 2025). A pragmatic reading of the evidence suggests that there are identifiable conditions under which benefits may accrue, alongside clear risks, and concrete institutional levers that can influence outcomes. In education, many of these issues emerge from the upstream structure of AI’s inference infrastructure. The way models are trained, aligned, and sourced shapes how they generate feedback, handle uncertainty, and interact with assessment systems, which in turn affects integrity, accuracy, and equity in learning environments.
What we know works emerges from decades of research in Artificial Intelligence in Education (AIED), which indicates that well-designed systems can improve learning, particularly for routine practice and targeted feedback. Meta-analyses of intelligent tutoring systems report moderate to large effects on achievement relative to business-as-usual instruction (George 2025; Ma et al. 2014; VanLehn 2011), and field studies of “personalised learning” models show gains when curricula, teacher development, and data use are implemented as an integrated package rather than as standalone tools (Pane et al. 2015). Recent reviews of AI in higher education similarly find positive, though heterogeneous, effects for feedback, writing support, and study guidance, while cautioning that outcomes depend on pedagogy and context (Holmes et al. 2019; Zawacki Richter et al. 2019). Early work on generative AI aligns with this picture: large language models (LLMs) can scaffold idea generation, formative feedback, and language support, with the largest benefits observed for less-prepared learners when activities are structured and teacher mediated (Kasneci et al. 2023). Where the risks are real, those same capacities that enable scalable feedback can also challenge assessment validity and potentially widen inequalities if guardrails are not in place. Empirical studies indicate that performance gains from educational technology often track pre-existing advantages unless access, language fit, and digital literacies are explicitly addressed (Reich 2020; Zawacki Richter et al. 2019). Learning analytics research documents ethical considerations around opacity, consent, and secondary data use in schools (Slade and Prinsloo 2013; Williamson 2017). For LLMs in particular, concerns include hallucination, hidden training data biases, and the risk of task over-automation that could deskill learners if systems replace rather than augment reasoning (Holmes et al. 2019; Kasneci et al. 2023). From a machine learning perspective, fairness-aware methods are necessary but not sufficient: bias can re-enter through target labels, domain shift, or deployment practices (Mehrabi et al. 2021).
From panic to practice: international guidance has moved beyond generic principles toward actionable controls. UNESCO’s Guidance for Generative AI in Education and Research (2023) recommends limiting high-stakes decision automation, keeping teachers “in the loop,” documenting data sources, and building institutional capacity for prompt and rubric design (UNESCO 2023). OECD analyses similarly emphasize human oversight, the measurement of learning—not only tool usage—and governance for procurement and data protection (OECD 2023). Teacher-facing AIED research shows that co-designed orchestration tools are more likely to be used and to improve instruction than student-only automation (Holstein et al. 2019). A pragmatic causal story emerges: when AI supports learning, it tends to do so through three mediators—first, more time on task and deliberate practice enabled by instant feedback (Ma et al. 2014; VanLehn 2011); second, cognitive offloading of lower level tasks so teacher and student attention can shift to higher order reasoning (Holmes et al. 2019); and third, responsiveness—adaptive hints and examples calibrated to current performance (Pane et al. 2015). Conversely, harms may arise when automation targets ill-defined tasks (open-ended judgments), when incentives emphasize output quantity over understanding, or when data pipelines lack transparency and contestability (Mehrabi et al. 2021; Slade and Prinsloo 2013). What institutions can implement now points toward practice-oriented levers that translate principles into verifiable routines. SOPs for classroom use can require disclosure of AI assistance on graded work; the logging of prompts and outputs (provenance) within the LMS; and the separation of practice spaces (where AI is allowed) from assessment spaces (where it is restricted) (OECD 2023; UNESCO 2023).
Teacher-in-the-loop design should prioritize tools that expose model rationales or exemplars teachers can edit, and provide dashboards that surface uncertainty and suggest next pedagogical actions (Holstein et al. 2019). Assessment integrity could be supported by shifting high-stakes evaluation toward authentic, process-revealing tasks (draft trails, orals, in-class builds) while using AI for low-stakes formative support (Reich 2020; Selwyn 2019). Equity by design calls for budgeting for access (devices, connectivity), language support, and digital literacy instruction, and for evaluating tools against subgroup outcomes, not just averages (Mehrabi et al. 2021; Zawacki Richter et al. 2019). Measurement should track learning gains relative to baselines, not solely tool engagement, and should include error audits for AI-generated feedback and for student misconceptions corrected versus those introduced (Pane et al. 2015; UNESCO 2023).
The rationale behind these recommendations is not to sensationalize AI as an external threat to academic integrity or the teacher’s role, but to recognize that AI can support learning when it is embedded in pedagogy, oversight, and equity infrastructure, and that it may have diminishing value if used as a shortcut or if governance treats documentation and accountability as optional. The practical agenda—SOPs, teacher-in-the-loop design, equity budgeting, and outcome-focused evaluation—addresses these conditions directly (Holstein et al. 2019; Ma et al. 2014; OECD 2023; UNESCO 2023; VanLehn 2011).
5 Ethics: from abstract principles to auditable practice
The ethical stakes of AI center on accountability and explainability, fairness and bias, human autonomy, and privacy and surveillance; these dimensions interact and therefore require integrated responses rather than piecemeal fixes. In the realm of bias and fairness, empirical audits show that AI systems can reproduce and amplify existing social inequalities when trained on skewed data or deployed without domain-specific guardrails (Metcalf 2025). Benchmark studies have documented disparate error rates in facial analysis for darker skinned women (Buolamwini and Gebru 2018) and systematic disparities in algorithmic hiring and screening (Cowgill 2020; Raghavan et al. 2020). Survey syntheses map the technical sources of bias—labeling, sampling, and measurement—and the trade-offs among formal fairness criteria (Corbett Davies and Goel 2018; Mehrabi et al. 2021). Social science work further demonstrates how data infrastructures and institutional arrangements embed inequality upstream, so purely technical remedies are insufficient without organizational change (Benjamin 2019; Eubanks 2018). The implication is that fairness must be treated as a sociotechnical property, combining dataset governance, context-appropriate metrics, and procedures for appeal and redress. Echoing the dynamics observed in employment and education, also the ethical issues surrounding AI arise from its nature as an inference infrastructure. Opacity, authority, and data governance reside upstream in the training pipeline, conditioning how fairness, accountability, and explainability appear downstream. Ethical failures, thus, reflect systemic infrastructural factors rather than individual model mistakes.
On accountability and explainability, many high-performing models remain opaque in ways that hinder error tracing and responsibility assignment (Binns 2018; Kroll et al. 2017). While technical work on explainability and interpretability provides useful tools, there are limits: post hoc explanations can mislead, and in safety–critical domains fully interpretable models may be preferable (Adadi and Berrada 2018; Doshi-Velez and Kim 2017; Rudin 2019). Recent risk taxonomies emphasize that accountability must extend across the model lifecycle—data, training, evaluation, and deployment—and be evidenced through logs, documentation, and incident reporting (Raji et al. 2020; Weidinger et al. 2022). The implication is to move from “explainability in principle” to auditable evidence of performance and process.
Regarding human autonomy and manipulation, algorithmic decision support can improve consistency but may also displace judgment or steer behavior in ways that are difficult to detect (Bruneault and Laflamme 2021; Buccella 2025; Yeung 2018; Zuboff 2019). Empirical studies of automated eligibility and scoring illustrate how decision authority can drift from professionals to defaults unless escalation thresholds and override rights are explicit (Eubanks 2018). The implication is to design for human-in-the-loop control where humans have genuine intervention points, supported by training, documentation of model limits, and downstream accountability.
Privacy and surveillance concerns add another layer, as large-scale data collection, model inversion, and membership inference attacks show that personal data (or close proxies) can leak from models even after training (Carlini et al. 2021; Shokri et al. 2017). These risks are unevenly distributed, with marginalized groups bearing disproportionate harms (O’Neil 2016). Regulatory baselines—the GDPR and related jurisprudence—set duties of purpose limitation, data minimization, and rights to contest automated decisions (European Union 2016; Veale and Zuiderveen Borgesius 2021). The implication is that privacy assurance must be tested empirically (for example, through leakage audits and documentation of retention, deletion, and access controls), not merely asserted.
From principles to practice, the field has moved toward operational instruments rather than abstract value lists: model cards and datasheets document intended use, data lineage, and evaluation (Gebru et al. 2021; Mitchell et al. 2019); the NIST AI Risk Management Framework offers process controls for mapping, measuring, and managing risk (NIST 2023); ISO/IEC 42001 establishes an AI management-system standard; and the EU AI Act institutionalizes risk-based obligations—conformity assessment, logging, transparency, and human oversight—for high-risk uses (EU 2024). Persistent challenges include opacity by design (limited access to training data and alignment procedures), normative indeterminacy (no shared thresholds for “acceptable error” across tasks), and assurance capture (auditors reliant on vendor access) (Mittelstadt 2019; Morley et al. 2021; Raji et al. 2020; Weidinger et al. 2022). The implication is that credible governance requires verifiable controls—publicly inspectable documentation of data and model lineage, reproducible evaluations using registered test sets, incident databases with time-to-remedy metrics, and independence safeguards for audits.
The agenda suggested by these considerations is concrete. It begins with documenting and testing: requiring dataset provenance, model cards, and registered evaluation reports with replication artifacts (Gebru et al. 2021; Mitchell et al. 2019). It continues with measuring fairness in context: selecting domain-appropriate metrics, publishing error budgets, and reporting appeal and reversal rates (Corbett Davies and Goel 2018; Mehrabi et al. 2021). It then emphasizes designing for human control: specifying override rights, escalation thresholds, and training for end users (Eubanks 2018; Yeung 2018). It also calls for hardening privacy: testing for leakage and membership inference and aligning retention and deletion with regulatory duties (Carlini et al. 2021; European Union 2016; Shokri et al. 2017). Strengthening assurance involves adopting NIST and ISO processes and implementing independent audits with access guarantees to avoid capture (EU 2024; ISO/IEC 42001 2023; NIST 2023; Raji et al. 2020).
Taken together, the literature suggests that AI governance needs to be evidence-based, measurable, and auditable across the AI supply chain, ensuring that fairness, accountability, autonomy, and privacy are verifiable in practice.
For a summary of the key points, see Table 1. Differentiating moral panic and pragmatism in AI discourse showing contrasting interpretive frames, underlying values, and implications for ethical and regulatory debates.
6 Collaborative governance of AI: a Neo-Triple Helix perspective
Collaborative governance of AI through a Neo-Triple Helix: from moral panic to pragmatic oversight presents a way to understand how to manage AI in a practical, evidence-based manner. Public discourse often moves between existential alarm and optimistic forecasts of frictionless progress, a pattern that reflects moral panic in which diffuse fears, hostile elements, and media amplification can outpace policy and evidence (Cia 2022; Cohen 2011; Goode and Ben Yehuda 1994; Sioumalas-Christodoulou and Tympas 2025). A more constructive approach treats risks as identifiable and manageable through concrete controls—documentation, evaluation, redress, and oversight—rather than as abstract threats (Beck 1992; Mitchell et al. 2019; Raji et al. 2020). The Neo-Triple Helix (NTH) provides a sociologically grounded way to organize that pragmatism. It reframes the traditional university–industry–government relationship as a co-evolving ecosystem with fluid roles shaped by local institutions, platform orchestrators, standards bodies, and assurance intermediaries, while treating civil society actors—professional groups, unions, NGOs, and communities—as constitutive participants entering through standards processes, procurement, consultation, and oversight forums (Cai 2022; DeNardis 2014; Etzkowitz and Leydesdorff 2000; Ostrom 2010). In this lens, AI governance becomes a policy mix—regulation, supervision, mission-oriented programs, public procurement, standards, audits, and shared data/model infrastructures—that can be iteratively adjusted as new evidence accumulates (Kuhlmann and Rip 2018; Mazzucato 2018; Sabel and Zeitlin 2012).
Operationally, the state steers through risk-based obligations and public-interest infrastructures such as open evaluation suites and incident databases, as exemplified by the EU AI Act’s tiered duties for high-risk systems, the NIST AI Risk Management Framework, and the ISO/IEC 42001 management-system standard (EU 2024; ISO/IEC 2023; NIST 2023). Industry—model vendors, integrators, and deployers in health, finance, and education—governs capability exposure, guardrails, and post-deployment monitoring through documentation, testing, red teaming, and incident response (Raji et al. 2020). Universities and public research bodies contribute public value by advancing fairness, interpretability, and evaluation through datasheets, model cards, benchmark design, and incident taxonomies (Gebru et al. 2021; Mitchell et al. 2019). Intermediaries such as standards organizations, professional associations, auditors, and certification bodies translate broad norms into auditable practice and indicators, linking ethics to evidence (Gorwa 2019; Power 1997).
This ecosystemic view aligns with analyses of employment, education, and ethics and helps move the debate beyond panic. In employment, randomized and field studies show that generative AI can raise productivity on routine cognitive tasks, often with the greatest gains for less experienced workers, while creating new verification work in open-ended tasks (Brynjolfsson et al. 2023; Noy and Zhang 2023; Peng et al. 2023). Macro-level evidence cautions that productivity gains do not automatically lead to higher employment or wages without complementary investment and task reallocation (Acemoglu and Restrepo 2020; Eloundou et al. 2023; Webb 2020). The NTH guides concrete responses: social dialog institutions and skills agencies co-design human–AI teaming with escalation thresholds, override rights, and error budgets, while competition and procurement policies address rent concentration at model/compute layers (Kenney and Zysman 2016; Srnicek 2017). In education, guidance from UNESCO and the OECD specifies guardrails—limit high-stakes automation, keep teachers involved, protect data, and require provenance—and institutions report learning and writing gains from scaffolded AI feedback, particularly for less-prepared learners, contingent on pedagogy and integrity measures (Holmes et al. 2019; Kasneci et al. 2023; OECD 2023; UNESCO 2023). The NTH helps convert these principles into practical operations: disclosure norms for AI assistance, provenance and logging in learning management systems, recognition of teacher workload, and shared evaluation resources that enable cross-vendor comparison. In ethics and assurance, a practical toolkit exists—datasheets and model cards, risk-management frameworks, management-system standards, and end-to-end audit protocols—and these can be tied to verifiable evidence such as error budgets, appeal and reversal rates, time to remedy, and adoption of content provenance (Floridi and Cowls 2019; ISO/IEC 2023; Mitchell et al. 2019; Mittelstadt 2019; NIST 2023; Raji et al. 2020). Persistent opacity in training data, alignment procedures, and evaluation sets requires independent scrutiny and clear disclosure incentives (Weidinger et al. 2022).
Three families of instruments make the NTH actionable across sectors. First, participation mechanisms involve standing, well-resourced forums where affected publics and professional communities co-define evidence thresholds and review deployments—such as citizens’ assemblies, worker councils, and panels for patients and parents—consistent with civic epistemologies and experimentalist governance (Davis et al. 2012; Jasanoff 2003; Sabel and Zeitlin 2012). Second, assurance guarantees call for documented data and model lineage, reproducible evaluations on registered test suites, continuous logging for audits and incident investigations, independent red-teaming, and post-deployment monitoring with time-to-disclosure service-level agreements, complemented by third-party audits to reduce “compliance theater” and auditor dependence (EU 2024; ISO/IEC 2023; Mitchell et al. 2019; NIST 2023; Power 1997; Raji et al. 2020). Third, procurement levers use public and large-buyer procurement to embed accountability upstream, requiring complete documentation, externally verifiable evaluations, incident reporting, content provenance for synthetic material, energy-use reporting, data portability, and audit access, with tenders scored on accuracy, bias mitigation, and time to remedy alongside price (Kuhlmann and Rip 2018; Mazzucato 2018; Raji et al. 2020).
Progress can be measured with sector-specific indicators that connect micro-practices to macro-outcomes: in employment, metrics include task delegation maps, override and escalation rates, participation in upskilling, and wage and job-quality trends; in education, indicators cover disclosure and provenance rates, integrity incidents and reversals, and learning gains by baseline attainment; in ethics and assurance, measures track the use of datasheets and model cards, reproducibility of evaluations, time to remedy, and provenance adoption. Important limitations and scope conditions temper expectations. Concentrated model and compute markets can give some actors agenda-setting power and increase the risk of assurance capture when auditors rely on vendor access (Power 1997; Raji et al. 2020; Srnicek 2017). The evidence base is still skewed toward the Global North, which limits generalizability to data-poor contexts and different civic cultures (Birhane 2021; Mohamed et al. 2020). Sectoral differences mean that acceptable error budgets and escalation rights vary across hiring, clinical decision support, and education; a single checklist may not fit all cases (Amann et al. 2020; Selbst et al. 2019). Opacity by design remains a structural challenge despite formal audits (Weidinger et al. 2022). Administrative burdens can also overwhelm small firms and public agencies unless there is shared assurance infrastructure and common test resources. Framed against moral panic, the Neo-Triple Helix does not deny public concern; it channels it into participatory spaces, verifiable assurance, and procurement-backed incentives that make claims testable and responsibilities clear, while retaining democratic legitimacy as AI systems increasingly organize work, learning, and decision-making (DeNardis 2014; Gorwa 2019; Scott 2014).
7 Conclusion
This article has argued that much of today’s AI discourse oscillates between moral panic and pragmatic governance. Moral-panic framings—extinction talk, claims of inevitable mass redundancy, or “end of the university” narratives—help explain the volatility of public attention but do little to guide decision-makers (Brown and Michael 2003; Cohen 2011; Goode and Ben-Yehuda 1994). A pragmatic stance, by contrast, anchors judgments in mechanisms that can be observed, measured, and governed. Our core contribution has been threefold: (i) to reframe contemporary AI as an inference infrastructure—a stack of data, models, compute, and assurance processes that pre-formats downstream practice—thereby relocating explanation upstream and away from headline-driven fears (Bender et al. 2021; Plantin et al. 2018); (ii) to operationalize this stance through an indicator-ready program (lineage, reproducible evaluation, provenance, appeal and redress), turning ethics from principle to evidence (Gebru et al. 2021; Mitchell et al. 2019; Raji et al. 2020); and (iii) to specify a Neo-Triple Helix (NTH) architecture that coordinates roles for government, industry, and universities together with standards bodies and organized publics so that assurance is credible and democratic, not theatrical (Cai 2022; DeNardis 2014; Power 1997).
Viewed historically, AI’s trajectory—from expert systems to statistical learning to foundation models—confirms why panic-based generalities are analytically weak. Each phase relocated authority (from local rules to upstream model providers), altered failure surfaces (from brittleness to distribution shift and data opacity), and demanded new forms of oversight (from domain scoping to documentation and third-party audit) (Bommasani et al. 2021; Jordan and Mitchell 2015; Shortliffe 1976; Weidinger et al. 2022). The inference infrastructure lens, therefore, explains both the renewed visibility of AI and the kinds of controls that matter: lineage and access to training data and alignment procedures, documented evaluations that can be reproduced, incident logging with time-to-remedy metrics, and clear accountability chains across vendors and deployers (EU AI Act 2024; Mitchell et al. 2019; NIST 2023; Raji et al. 2020).
Applied to employment, the evidence does not support apocalyptic redundancy claims nor uncritical optimism. Randomized and field studies show sizable productivity gains for routinized cognitive work—often largest for less-experienced workers—coupled with new verification workloads and uneven effects when tasks are open-ended (Brynjolfsson et al. 2023; Noy and Zhang 2023; Peng et al. 2023). Macroeconomic work cautions that productivity gains do not automatically raise employment or wages; outcomes hinge on task reallocation, demand elasticities, and complementary investment (Acemoglu and Restrepo 2020; Autor 2015; ILO 2023). A pragmatic program, therefore, centers on human–AI teaming (escalation thresholds, override rights, error budgets), job-quality safeguards in algorithmic management (transparency, audit trails, worker voice), and skills pathways targeted to mid-career workers (Kellogg et al. 2020; Kenney and Zysman 2016; OECD 2023). Moral-panic narratives can be repurposed as motivation for this institutional work, but they cannot substitute for it.
In education, the panic repertoire focuses on cheating and deskilling; the empirical record is more differentiated. Decades of AIED research and recent studies of large language models indicate that well-scaffolded systems improve practice and writing—especially for lower prepared learners—when teachers remain in the loop and integrity rules are clear (Holmes et al. 2019; Kasneci et al. 2023; Ma et al. 2014; VanLehn 2011). Risks—assessment invalidity, inequities from uneven access and digital literacies, opacity in learning analytics—are real but governable with disclosure norms, provenance and logging in learning management systems, equity budgeting, and a shift toward process-revealing assessment (OECD 2023; UNESCO 2023; Williamson 2017). Here again, a pragmatic stance converts anxiety into auditable routines that track learning gains, integrity incidents and reversals, and subgroup outcomes, rather than tool usage alone (Pane et al. 2015).
In ethics, widely cited problems—bias, opacity, manipulation, and privacy—are best treated as sociotechnical and life-cycle properties. Technical mitigations must be combined with organizational procedures for contestation and redress. Empirical audits document disparate error for identity-relevant systems (Buolamwini and Gebru 2018) and hiring pipelines (Raghavan et al. 2020), while accountability research shows the limits of post-hoc explainability and the need for interpretable models in safety–critical settings (Doshi-Velez and Kim 2017; Rudin 2019). Privacy risks extend to model leakage and membership inference, which require testing, not assertion (Carlini et al. 2021; Shokri et al. 2017). The regulatory stack—NIST AI RMF, ISO/IEC 42001, and the EU AI Act—already points toward evidence-based oversight; the task is to prevent assurance capture by ensuring auditor independence and access to evaluation artifacts (EU AI Act 2024; ISO/IEC 42001 2023; NIST 2023; Power 1997; Raji et al. 2020).
The NTH lens integrates these domain findings. Governments steer through risk-based obligations and public-interest infrastructures (open benchmarks, incident databases), industry governs capability exposure and post-deployment monitoring, and universities build the methods and shared test resources that make claims comparable across vendors and sectors (Cai 2022; DeNardis 2014; Mitchell et al. 2019). Participation mechanisms—citizens’ assemblies, worker councils, patient and parent panels—translate diverse values into evidence thresholds for acceptable error and remedy (Jasanoff 2003; Sabel and Zeitlin 2012). Procurement levers and competition policy help align private incentives with public value by conditioning market access on documentation, reproducibility, provenance, and energy-use reporting, while guarding against rent concentration at the model/compute layer (Kuhlmann and Rip 2018; Srnicek 2017).
Several scope conditions temper these conclusions. The evidence base is skewed toward the Global North, limiting generalizability to data-poor contexts and different civic epistemologies (Birhane 2021; Mohamed, Png, and Isaac 2020). Sectoral heterogeneity means that error budgets and escalation rights must be domain-specific—hiring is not medicine, and schooling is not credit scoring (Amann et al. 2020; Selbst et al. 2019). Opacity by design—confidential training corpora, alignment procedures, and evaluation sets—remains a structural obstacle even under formal audits (Weidinger et al. 2022). Finally, assurance can devolve into ritual unless indicators track performance (time to remedy, appeal and reversal rates, distributional impacts), not just process (Power 1997).
For research, the agenda is cumulative and comparative: within-firm before/after designs, cross-firm matched comparisons, and sectoral fieldwork that stitch micro-case handling to meso-procedures and macro-settlements, reporting a common indicator set for agency, opacity, normativity, and automation across employment, education, and high-risk ethics domains (Mitchell et al. 2019; Morley et al. 2021; Scott 2014). For policy, the recommendation is to fund public evidence infrastructures (benchmark registries, incident databases), protect auditor and researcher access, and institutionalize participatory fora that make assurance responsive to those most affected (Davis et al. 2012; DeNardis 2014).
In sum, the paper’s contribution is to reposition the debate: moral panic names a recurring discourse pattern; pragmatism names a program of measurable remediation. Treating AI as an inference infrastructure clarifies where authority now sits and which levers work. The NTH provides an organizational blueprint for translating ethics into auditable practice while maintaining social legitimacy. If adopted, this orientation allows institutions to move past headline-driven swings and toward accountable gains in productivity, learning, and rights protection—outcomes that can be demonstrated rather than declared (Beck 1992; EU AI Act 2024; Jasanoff 2004; NIST 2023).
Acknowledgements
We are grateful to the anonymous reviewers for their careful reading and insightful comments. Their suggestions have strengthened the arguments and sharpened the focus of this work.
Declarations
Conflict of interest
The authors declare no competing interests.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.