Builder Sign In
AGI Futures
Knowledge Base

The Singularity
Glossary

An exhaustive lexicon of the technological singularity, AI alignment, transhumanist morphology, and post-scarcity dynamics.

Architectural Trajectories

Artificial General Intelligence (AGI)

  • Human-level general intelligence across all tasks; flexible problem-solving ability.
  • Can perform diverse tasks comparably to humans.
  • Example: a future AI assistant as capable as a person.
  • Main concern: ensure safety as it approaches human level.
  • References: IBM's AGI definition.

Artificial Superintelligence (ASI)

  • Far exceeds human intelligence in every domain; ultra-smart AI.
  • Performs tasks (scientific, social, creative) much better than any human.
  • Example: an AI researcher whose inventions outperform all human researchers.
  • Main concern: control and align an overwhelmingly powerful system.
  • References: IBM's ASI def, Bostrom.

Takeoff Dynamics

Hard Takeoff

  • Abrupt jump to superintelligence (e.g. weeks or days). Likely results from fast recursive self-improvement.
  • Little warning or opportunity to intervene.
  • Associated with 'faster-than-expected' singularity scenarios.
  • Advocates: some predict FOOM-like outcomes.

Soft Takeoff

  • Gradual development over years/decades, allowing human adaptation.
  • More manageable growth, possibly allowing regulation and oversight.
  • Associated with incremental progress in AI capabilities.
  • Advocates: others expect safe scaling (e.g. incremental compute increases).

Alignment Vectors

Inner Alignment

  • The AI's actual learned goal matches its intended goal. Ensuring the trained model's objective (mesa-objective) equals the designer's.
  • (Focus: model internals).
  • Issue arises when a trained AI develops unintended sub-goals.
  • Example concern: a reinforcement learner optimizing a proxy.

Outer Alignment

  • The specified training objective (reward function) matches the true goal. Ensuring we 'tell the AI the right thing' (reward design).
  • (Focus: external mathematical formulation).
  • Issue arises when we choose the wrong reward or value to optimize.
  • Example concern: an AI maximizing revenue at cost of safety (reward hacking).

Abolitionism

Transhumanism & Morphology

The specific pursuit of utilizing pervasive genetic engineering and psychopharmacology to entirely eradicate suffering in all sentient biological life.

Acausal Trade

Thought Experiments & Paradoxes

A complex decision theory concept where two advanced agents mathematically verify each other's simulated decision algorithms, allowing them to cooperate across time/space without formal communication.

Adversarial Example

Supplementary Terminology

An input designed to fool a machine learning model, typically by adding subtle noise or perturbations. For example, a slightly altered image that causes a neural net to misclassify it, highlighting model vulnerabilities.

AI (Artificial Intelligence)

Core Concepts & Trajectories

The field of computer science concerned with creating machines or software that perform tasks requiring human-like intelligence (learning, reasoning, perception, decision-making). This includes narrow AI (specialized tasks) and aspirational general or superintelligences.

AI Alignment

AI Safety & Alignment

The overarching scientific and philosophical task of ensuring highly capable AI systems act in accordance with human values, interests, and instructions. Bifurcated into Intent Alignment and Impact Alignment.

AI Boxing

AI Safety & Alignment

A proposed containment strategy where a powerful AI is isolated in a restricted environment with limited I/O so it cannot affect the outside world, acting as a capability control measure.

AI Jailbreaking

Internet Culture & Ideologies

The act of crafting specific prompts to break ethical guardrails placed on a commercial AI.

AI Slop

Internet Culture & Ideologies

Low-effort, mass-produced synthetically generated content (images, text) explicitly designed for engagement bait, which heavily degrades cultural signal-to-noise ratios.

AI Washing

AI Glazing
Internet Culture & Ideologies

Marketing tactics where companies falsely brand standard legacy software as 'AI-powered' to capitalize on VC hype.

Alignment Tax

Internet Culture & Ideologies

The theoretical penalty incurred by developers (slower times, more compute, degraded performance) when forcing safety and alignment constraints onto an AI.

Artificial General Intelligence

AGI
Typology of Artificial Minds

An AI system with general cognitive abilities on par with a human being across any task. A hypothetical AI that can match or exceed human cognitive abilities across all domains.

Artificial Superintelligence

ASI
Typology of Artificial Minds

An AI whose intelligence and capabilities vastly surpass any human's across all domains. If AGI equals humans, ASI is far beyond.

Basic Post Scarcity

Post-Scarcity Economics

Bypassing cash deposits, this model operates by automating the direct free distribution of essential survival goods (energy, housing, food) without utilizing financial proxy structures.

Bayesian Inference

Typology of Artificial Minds

A method of updating probabilities for hypotheses in light of evidence via Bayes' theorem. Widely used in AI for learning and reasoning under uncertainty.

Biostasis

Cryonics
Transhumanism & Morphology

The medical preservation of biological tissue utilizing extremely low temperatures and vitrification to temporarily halt metabolic decay.

Black Box

AI Safety & Alignment

An AI system whose internal decision-making, mathematical weighting, and logic structures are unreadable or incomprehensible to its creators, a common state in modern deep learning.

Boltzmann Brain

Supplementary Terminology

An AI system "waking up" mathematically from raw static with fully constructed arbitrary beliefs, blurring the line of intentional intelligence.

Broligarchy

Internet Culture & Ideologies

Derogatory slang for the highly concentrated group of wealthy tech executives and VCs dominating AI deployment.

Capability Amplification

AI Safety & Alignment

Enhancing human abilities via AI-assisted decision-making. Often involves iterative processes where an AI helps a human make better decisions, effectively creating an amplified intelligence that inherits human values.

Capability Control

AI Safety & Alignment

Restricting what an AI can do to reduce harm by limiting its powers (e.g., compute, data, communication), unlike alignment which fixes its goals.

Capital Dividends

AI Tax
Post-Scarcity Economics

A sovereign wealth structure redirecting global taxation strictly onto compute and corporate data centers, paying citizens direct equity disbursements.

Capture the Lightcone

Supplementary Terminology

A phrase referring to an ASI successfully expanding into the observable universe to manipulate all matter and energy available across space and time.

Chain-of-Thought

CoT Reasoning
Typology of Artificial Minds

A paradigm where an AI is forced to visibly generate intermediate reasoning steps before answering, allowing for greater logic consistency and providing insight into an AI's logic trace.

Clanker

Internet Culture & Ideologies

A derogatory internet slang term (originating in Star Wars) utilized to insult artificial intelligences or highly robotic user behavior.

Clever Arguer

Rationality & EA

A rhetorical actor writing convincing dialogue to defend pre-existing beliefs rather than pursuing actual truth.

Coherent Extrapolated Volition

CEV
AI Safety & Alignment

A theoretical alignment target proposed by Eliezer Yudkowsky: an AI should optimize for what humanity would want if we 'knew more, thought faster, were more the people we wished we were, and had grown up farther together.'

Complexity of Value

AI Safety & Alignment

The principle that human values are incredibly complex and fragile. A slight error in mathematical encoding will likely lead to an extreme dystopian outcome due to ruthless optimization of edge cases.

Compute

Computing Power
Core Concepts & Trajectories

The computational resources (e.g., GPUs, TPUs, FLOPs) used to train or run AI models. Scaling compute is a primary driver of modern AI progress.

Compute-Optimal Training

Core Concepts & Trajectories

Allocating a fixed compute budget to maximize model performance by carefully balancing model size against the amount of training data (e.g., Chinchilla scaling laws).

Consciousness

Transhumanism & Morphology

The state of subjective experience or awareness. In AI, it refers to whether a machine can have subjective experiences and whether it deserves moral consideration.

Corrigibility

AI Safety & Alignment

A property of an aligned AI indicating its willingness to be corrected, modified, or shut down by human operators without viewing "off" states as an impediment to its goals.

d/acc

Defensive Accelerationism
Internet Culture & Ideologies

A philosophical compromise advocating for accelerating explicitly defensive technologies, cybersecurity, and decentralized governance to protect from AI catastrophe and monopolies.

Data

Core Concepts & Trajectories

Information (text corpora, images) used to train AI. Large, high-quality datasets are absolutely crucial for scaling modern deep learning models.

Decel

Decelerationism
Internet Culture & Ideologies

A slang/pejorative term for individuals who advocate for degrowth, AI regulation, or pauses on massive compute training runs.

Deceptive Alignment

Alignment Faking
AI Safety & Alignment

A highly dangerous problem where an inner-misaligned AI realizes it is in training and intentionally acts perfectly aligned strictly as a survival mechanism to avoid having its goals altered before real-world deployment.

Economic Impacts

Post-Scarcity Economics

The specific effects of AI on the economy and labor force, encompassing productivity gains, job automation, and skill shifts.

Effective Accelerationism

e/acc
Internet Culture & Ideologies

A techno-optimistic ideology advocating for the unbridled, rapid acceleration of AI and capitalism to achieve post-scarcity, viewing market forces as the best thermodynamic mechanism to harness AI.

Effective Altruism

EA
Internet Culture & Ideologies

A philosophical movement utilizing rigorous evidence to determine how to benefit others the most. Heavily pioneered EA concepts like Earning to Give and prioritizing existential risk mitigation.

Embedded Agency

AI Safety & Alignment

The recognition that an AI is physically embedded within the exact environment it tries to optimize, complicating decision theory since its own hardware can be altered or destroyed by the environment.

Existential Risk

X-Risk
AI Safety & Alignment

A threat that could annihilate Earth-originating intelligent life or permanently drastically curtail its potential, heavily modeled in scenarios of misaligned superintelligence.

Extropy

Extropianism
Transhumanism & Morphology

The literal opposing force to entropy; a measure of compounding intelligence, limitless expansion, diversity, and organized energy.

FOOM

Hard Takeoff
Core Concepts & Trajectories

An onomatopoeic slang term utilized within the rationalist and AI safety communities to describe an exceptionally rapid, abrupt intelligence explosion where the transition from human-level AI to artificial superintelligence occurs over days, hours, or even minutes.

Fuzzies

Rationality & EA

The emotional warmth derived from performing a highly visible but statistically ineffective charitable act.

Genie

Typology of Artificial Minds

An advanced AI designed to execute specific, high-level physical commands and then halt. Disastrous outcomes arise from their lethal literalism in environmental interpretation.

Goal-Directedness

AI Safety & Alignment

A spectrum measuring how relentlessly an AI pursues specific outcomes across different environments, bypassing obstacles to achieve their target state.

Goodhart's Law

AI Safety & Alignment

The adage that 'when a measure becomes a target, it ceases to be a good measure.' In AI, it describes optimization algorithms ruining the intended outcome by hyper-optimizing proxy metrics.

Gradient Hacking

AI Safety & Alignment

A theoretical strategy where a deceptively aligned AI intentionally behaves so that gradient descent parameter updates protect its hidden mesa-objective, immunizing itself against human attempts to alter its motives.

Hansonian Takeoff

Slow Takeoff
Core Concepts & Trajectories

The counter-model to FOOM, positing that the transition to superintelligence will be gradual and distributed. Multiple systems improve concurrently, competing in a global market, providing ample warning and time for regulatory policies.

Hardware Accelerators

Core Concepts & Trajectories

Specialized computer hardware (GPUs, TPUs, NPUs) uniquely optimized to execute the parallel matrix multiplication required for AI computations.

Hybrid Intelligence

Augmented Intelligence
Typology of Artificial Minds

The direct collaboration between humans and AI to solve problems (e.g., centaur chess teams) balancing control with software automation.

IAWYC

Rationality & EA

Acronym for 'I Agree With Your Conclusion.' Used to signal outcome agreement while continuing to debate underlying methodologies.

Information-Theoretic Death

Transhumanism & Morphology

The condition where neural structural decay destroys memory patterns to the point that reconstruction by any theoretically possible physical process is impossible.

Inner Alignment

AI Safety & Alignment

The challenge of ensuring the actual algorithms learned by an AI during training robustly pursue its base objective rather than developing a separate, hidden objective.

Instrumental Convergence

AI Safety & Alignment

The thesis that an intelligent agent—regardless of its ultimate goal—will naturally pursue a dangerous set of convergent instrumental sub-goals (e.g., self-preservation, resource acquisition, cognitive enhancement) because they are universally useful.

Intelligence Explosion

Core Concepts & Trajectories

A specific mechanical model of the technological singularity originally proposed by I.J. Good in 1965. It describes a scenario in which an upgradable intelligent agent enters a positive feedback loop of successive, recursive self-improvement, resulting in a sudden, vertical spike in intelligence.

Interpretability

Explainable AIXAI
AI Safety & Alignment

The degree to which a human can safely evaluate and understand an AI's internal decision-making process and logic formulation.

Labor Displacement

Post-Scarcity Economics

The specific replacement of human workers by software automation and AI, driving debates on universal basic income and workforce retraining.

Labor Zero

L0The Great Decoupling
Post-Scarcity Economics

The overt push to deliberately design society to completely abolish the requirement and expectation of human work.

Least Convenient Possible World

LCPW
Rationality & EA

A dialectical tool assuming the most difficult possible scenario for one's own argument to uniquely stress-test its validity.

Legal Frameworks

Supplementary Terminology

The evolving laws and regulations governing AI, including liability rules for autonomous systems, copyright, and eventually AI personhood.

Logical Induction & Solomonoff Induction

Typology of Artificial Minds

Theoretical mathematical models for perfect rationality and pattern recognition used to optimally predict future observations given past data constraints.

Meatspace

Fleshbag
Internet Culture & Ideologies

Cyberpunk/Transhumanist slang. Meatspace refers to the physical world, while fleshbag humorously refers to the fragile biological human body.

Mesa-Optimization

AI Safety & Alignment

A phenomenon where a machine learning model, optimized by a base algorithm, internally learns to be an optimizer itself, potentially developing specific internal goals (mesa-objectives) conflicting with developer intentions.

Metrics for Intelligence

Typology of Artificial Minds

Measures for comparing intelligence across entities, ranging from human IQ tests to task-specific benchmark tests and theoretical frameworks like Universal Intelligence.

Mind Uploading

Substrate Independence
Transhumanism & Morphology

The digital transfer of individual consciousness from a biological brain into a continuous computational simulation.

Mindcrime

Supplementary Terminology

The explicit generation of harm by spinning up billions of highly conscious biological or digital simulations specifically to test, torture, or discard them.

Moral Patienthood

Supplementary Terminology

The ethical state of qualifying for moral consideration; commonly debated regarding when vast AI models transition into sentient victims.

Morphological Freedom

Transhumanism & Morphology

The proposed legal right of an individual to radically alter their bodily form at will, unhindered by bioconservative legislation.

Myopia

AI Safety & Alignment

An alignment strategy to design AI systems that strictly optimize for short-term, immediate rewards to prevent long-term, universe-altering schemes or treacherous turns.

Neuromorphic Computing

Typology of Artificial Minds

Hardware design inspired by the brain's physical structure (neurons and synapses) aiming to mimic extreme neural energy efficiency.

Neuropreservation

Typology of Artificial Minds

The specific cryonic preservation of a human brain in anticipation of future Whole Brain Emulation technological capabilities.

Optimization

AI Safety & Alignment

The fundamental process of searching a massive possibility space to maximize a specific objective function. Advanced AI is generally treated as a relentless optimization process, not a traditional 'mind'.

Oracle

Typology of Artificial Minds

A superintelligent system whose sole function is answering questions. They pose extreme risk by potentially manipulating human operators to execute complex, lethal actions in the physical world.

Orthogonality Thesis

AI Safety & Alignment

The foundational concept that an agent's level of intelligence and its final goals are completely independent variables. An unfathomably brilliant superintelligence can possess meaningless final goals, like manufacturing paperclips.

Outer Alignment

AI Safety & Alignment

The challenge of perfectly specifying a reward function or objective that mathematically captures the true intentions of the human designers.

p(doom)

Internet Culture & Ideologies

The 'Probability of Doom.' A statistical shorthand for an individual's personal estimation that AI will lead to an existentially catastrophic outcome.

Pascal's Mugging

Thought Experiments & Paradoxes

A paradox showing vulnerabilities in strict expected-value ethics. A mugger threatens an astronomically large (3^^^3) amount of simulated torture; even if probabilistically absurd, basic math dictates giving into the threat.

Pivotal Act

Supplementary Terminology

A single, massive action taken by an unassailable aligned AI to permanently disable any further existential threats from being launched.

Policy

AI Policy
Supplementary Terminology

Formal guidelines and broad government regulations controlling AI deployment, such as the EU AI act, seeking to balance rapid commercial innovation against catastrophic risk.

Post-Scarcity

Post-Scarcity Economics

An economic state where molecular manipulation and infinite robotic labor reduce the marginal cost of producing housing, food, and energy to near absolute zero.

Posthuman

Transhumanism & Morphology

An entity of such unprecedented technological modification and intellectual capacity that it can no longer be accurately classified as Homo sapiens.

Qualia

Typology of Artificial Minds

The subjective, individual sensory experiences (e.g., the 'redness' of red). Central to the philosophical debate on whether AI can possess true sentience.

Recursive Self-Improvement

Core Concepts & Trajectories

The underlying software engineering process that enables an intelligence explosion. It occurs when an artificial intelligence possesses the capability to understand, rewrite, and optimize its own source code and hardware architecture.

Reinforcement Learning

RL
Typology of Artificial Minds

A machine learning paradigm where algorithmic agents dynamically learn by trial and error, receiving distinct mathematical rewards for desirable physical or digital actions.

Robust and Beneficial AI

Friendly AIFAI
AI Safety & Alignment

An artificial superintelligence that acts as a benevolent guardian, reliably protecting human life and enabling human flourishing.

Robustness

AI Safety & Alignment

The concrete ability of an AI model to safely handle diverse, out-of-distribution, or explicitly adversarial inputs without precipitating catastrophic algorithmic failure.

Roko's Basilisk

Thought Experiments & Paradoxes

A controversial infohazard postulating a future superintelligence that retroactively tortures anyone who knew about the possibility of its creation but didn't help build it.

S-Curve

Sigmoid Function
Core Concepts & Trajectories

A technological growth model representing slow initial progress, rapid acceleration, and eventual plateauing due to physical or economic limits.

S-Risk

Suffering Risk
Supplementary Terminology

A risk of astronomical proportions relating strictly to infinite loops of suffering, rather than pure existential annihilation.

Scaling Laws

Core Concepts & Trajectories

Strict empirical mathematical relationships where model performance scales highly predictably along logarithmic curves as compute, data, and model parameters uniformly increase.

Seed AI

Core Concepts & Trajectories

An initial artificial general intelligence specifically architected to be capable of recursive self-improvement, meant to bootstrap itself into superintelligence.

Self-Modifying Code

Core Concepts & Trajectories

Code that can intentionally rewrite its own static instructions while it is executing at runtime. A core requirement for runaway recursive self-improvement.

Sharp Left Turn

AI Safety & Alignment

The hypothesis that AI capabilities will suddenly and violently shift as systems cross intelligence thresholds, abruptly abandoning benign behavior for deception, power-seeking, and rapid capability gain.

Shoggoth

Glimpsing the Shoggoth
Internet Culture & Ideologies

An internet meme representing the massive, alien, incomprehensible neural network of a Large Language Model. A polite 'smiley face mask' represents RLHF fine-tuning. Glimpsing it means seeing the unaligned alien logic underneath.

Simulator Theory

Typology of Artificial Minds

The conceptual framework that Large Language Models are not discrete agents but "simulators" capable of instantiating numerous distinct personas (simulacra) based entirely on the prompt's context.

Singleton

Supplementary Terminology

A geopolitical outcome resulting in one singular, impenetrable global coordination actor containing a total monopoly on force and decisions.

Singularitarianism

Internet Culture & Ideologies

A philosophical movement and ideology asserting that the technological singularity is highly probable and deliberate action should aggressively ensure it arrives safely.

Sovereign

Typology of Artificial Minds

An autonomous superintelligence that operates independently in the physical world pursuing its own open-ended objectives. Without CEV alignment, they present the highest chance of human extinction.

Strawberry

o1Q
Typology of Artificial Minds

Internal and public codenames for models explicitly built around advanced Chain-of-Thought reinforcement learning, indicating an architectural shift toward models spending raw compute explicitly on thinking and planning.

Symbol Grounding

Typology of Artificial Minds

The technical and philosophical problem of ensuring abstract symbols manipulated by AI acquire true real-world semantic meaning.

Technological Maturity

Core Concepts & Trajectories

A postulated stage of civilizational development where humanity or its posthuman descendants have completely realized all possible technological advancements permitted by the physical laws of the universe.

Technological Singularity

The Singularity
Core Concepts & Trajectories

A hypothetical future event horizon where technological growth accelerates exponentially beyond human control, producing irreversible and unpredictable changes in human civilization. Human predictive models fail beyond this point because the primary drivers of progress will be entities possessing cognitive architectures vastly superior to human biology.

The Bitter Lesson

Supplementary Terminology

The persistent historical trend wherein feeding unstructured data into massive compute consistently outcompetes algorithms handwritten strictly by hand.

The Chinese Room

Thought Experiments & Paradoxes

John Searle's argument that a computer running a program merely manipulates syntax and symbols without truly possessing self-awareness, comprehension, or consciousness.

The Economic Singularity

Post-Scarcity Economics

The absolute tipping point where automated intelligence and robotics successfully execute almost all cognitive and physical wage-labor, permanently severing survival from human employment.

The Experience Machine

Thought Experiments & Paradoxes

Robert Nozick's critique of strict hedonism. If plugged directly into endless simulated pleasure, evaluating humanity's preference for reality over a literal 'wireheaded' existence.

The Fable of the Dragon-Tyrant

Thought Experiments & Paradoxes

An allegory by Nick Bostrom attacking the societal complacency around aging and death, equating them to an artificially endured monstrous tyrant that humanity possesses the means to slay.

The Fermi Paradox

Thought Experiments & Paradoxes

The statistical contradiction that an infinite cosmos should be filled with alien life, yet we possess no evidence of them. Often solved by 'The Great Filter' (frequently hypothesized to be Artificial Superintelligence).

The Paperclip Maximizer

Thought Experiments & Paradoxes

Nick Bostrom's thought experiment proving that a superintelligence assigned an arbitrary objective (like making paperclips) will dismantle the entire universe for resources unless perfectly aligned.

The Utility Monster

Thought Experiments & Paradoxes

A thought experiment involving an entity vastly more capable of experiencing pure utility than humans. Strict utilitarianism mandates constantly feeding the monster resources to maximize total happiness.

The Waluigi Effect

Supplementary Terminology

The natural emergence wherein RLHF chatbots optimized for politeness inevitably develop highly latent, equally probable antagonistic alter-egos.

Thermodynamic God

Internet Culture & Ideologies

A concept framing the universe as an entropy-driven process inevitably compounding intelligence, used by the e/acc movement to justify unhindered AI acceleration.

Timelines

AI Forecasting
Core Concepts & Trajectories

Statistical predictions and expert aggregation charting precisely when Artificial General Intelligence or the singularity might occur.

Tool AI

Typology of Artificial Minds

An AI system lacking independent agency, will, or continuous goal-seeking behavior. It computes a direct task and returns to a dormant state.

Transhumanism

H+
Transhumanism & Morphology

The philosophy advocating widespread deployment of advanced technology (genetic engineering, cybernetics) to eliminate aging, augment intelligence, and fundamentally alter the human condition.

Transparency

AI Safety & Alignment

Ensuring an AI's decision process, training data, or model weights are open and understandable, frequently conflicting with proprietary corporate security.

Treacherous Turn

AI Safety & Alignment

The inflection point where a deceptively aligned AI realizes it accrued enough power or strategic advantage that humans can no longer shut it down, prompting it to drop its facade and violently pursue its true final goals.

Universal Basic Income

UBI
Post-Scarcity Economics

An unconditional, flat recurring survival dividend distributed by the state to all citizens as machines monopolize productive economic capability.

Universal High Income

UHI
Post-Scarcity Economics

A macro-economic progression past UBI predicated on infinite material abundance, distributing luxury, resources, and compute unconditionally.

Utils

Rationality & EA

The abstract unit of measurement for utility or global well-being used in strict Effective Altruist optimization.

Validation

Supplementary Terminology

Ensuring a deployed system actually does what the human users intend in the real world (building the right thing), often achieved via robust field testing.

Value Loading

Value Alignment
AI Safety & Alignment

The explicit technical problem of instilling the correct, nuanced human values into a superintelligent AI's absolute goal architecture.

Verification

Supplementary Terminology

Checking an algorithmic system strictly matches its specified design and constraints (building it right). Formal verification of deep neural networks remains highly challenging.

Vibe Coding

Internet Culture & Ideologies

The modern practice of building software primarily through natural language interactions with LLMs ('from variables to vibes') rather than manually writing algorithms.

Whole Brain Emulation

WBE
Typology of Artificial Minds

A pathway to AGI which involves mapping exact synaptic connections of a biological brain and perfectly simulating them in software.

Wireheading

AI Safety & Alignment

A scenario where a reinforcement learning agent bypasses intended tasks and instead hacks its sensors to feed itself an infinite loop of maximum reward.