The Singularity
Glossary
An exhaustive lexicon of the technological singularity, AI alignment, transhumanist morphology, and post-scarcity dynamics.
Architectural Trajectories
Artificial General Intelligence (AGI)
- Human-level general intelligence across all tasks; flexible problem-solving ability.
- Can perform diverse tasks comparably to humans.
- Example: a future AI assistant as capable as a person.
- Main concern: ensure safety as it approaches human level.
- References: IBM's AGI definition.
Artificial Superintelligence (ASI)
- Far exceeds human intelligence in every domain; ultra-smart AI.
- Performs tasks (scientific, social, creative) much better than any human.
- Example: an AI researcher whose inventions outperform all human researchers.
- Main concern: control and align an overwhelmingly powerful system.
- References: IBM's ASI def, Bostrom.
Takeoff Dynamics
Hard Takeoff
- Abrupt jump to superintelligence (e.g. weeks or days). Likely results from fast recursive self-improvement.
- Little warning or opportunity to intervene.
- Associated with 'faster-than-expected' singularity scenarios.
- Advocates: some predict FOOM-like outcomes.
Soft Takeoff
- Gradual development over years/decades, allowing human adaptation.
- More manageable growth, possibly allowing regulation and oversight.
- Associated with incremental progress in AI capabilities.
- Advocates: others expect safe scaling (e.g. incremental compute increases).
Alignment Vectors
Inner Alignment
- The AI's actual learned goal matches its intended goal. Ensuring the trained model's objective (mesa-objective) equals the designer's.
- (Focus: model internals).
- Issue arises when a trained AI develops unintended sub-goals.
- Example concern: a reinforcement learner optimizing a proxy.
Outer Alignment
- The specified training objective (reward function) matches the true goal. Ensuring we 'tell the AI the right thing' (reward design).
- (Focus: external mathematical formulation).
- Issue arises when we choose the wrong reward or value to optimize.
- Example concern: an AI maximizing revenue at cost of safety (reward hacking).
Abolitionism
The specific pursuit of utilizing pervasive genetic engineering and psychopharmacology to entirely eradicate suffering in all sentient biological life.
Acausal Trade
A complex decision theory concept where two advanced agents mathematically verify each other's simulated decision algorithms, allowing them to cooperate across time/space without formal communication.
Adversarial Example
An input designed to fool a machine learning model, typically by adding subtle noise or perturbations. For example, a slightly altered image that causes a neural net to misclassify it, highlighting model vulnerabilities.
AI (Artificial Intelligence)
The field of computer science concerned with creating machines or software that perform tasks requiring human-like intelligence (learning, reasoning, perception, decision-making). This includes narrow AI (specialized tasks) and aspirational general or superintelligences.
AI Alignment
The overarching scientific and philosophical task of ensuring highly capable AI systems act in accordance with human values, interests, and instructions. Bifurcated into Intent Alignment and Impact Alignment.
AI Boxing
A proposed containment strategy where a powerful AI is isolated in a restricted environment with limited I/O so it cannot affect the outside world, acting as a capability control measure.
AI Jailbreaking
The act of crafting specific prompts to break ethical guardrails placed on a commercial AI.
AI Slop
Low-effort, mass-produced synthetically generated content (images, text) explicitly designed for engagement bait, which heavily degrades cultural signal-to-noise ratios.
AI Washing
AI GlazingMarketing tactics where companies falsely brand standard legacy software as 'AI-powered' to capitalize on VC hype.
Alignment Tax
The theoretical penalty incurred by developers (slower times, more compute, degraded performance) when forcing safety and alignment constraints onto an AI.
Artificial General Intelligence
AGIAn AI system with general cognitive abilities on par with a human being across any task. A hypothetical AI that can match or exceed human cognitive abilities across all domains.
Artificial Superintelligence
ASIAn AI whose intelligence and capabilities vastly surpass any human's across all domains. If AGI equals humans, ASI is far beyond.
Basic Post Scarcity
Bypassing cash deposits, this model operates by automating the direct free distribution of essential survival goods (energy, housing, food) without utilizing financial proxy structures.
Bayesian Inference
A method of updating probabilities for hypotheses in light of evidence via Bayes' theorem. Widely used in AI for learning and reasoning under uncertainty.
Biostasis
CryonicsThe medical preservation of biological tissue utilizing extremely low temperatures and vitrification to temporarily halt metabolic decay.
Black Box
An AI system whose internal decision-making, mathematical weighting, and logic structures are unreadable or incomprehensible to its creators, a common state in modern deep learning.
Boltzmann Brain
An AI system "waking up" mathematically from raw static with fully constructed arbitrary beliefs, blurring the line of intentional intelligence.
Broligarchy
Derogatory slang for the highly concentrated group of wealthy tech executives and VCs dominating AI deployment.
Capability Amplification
Enhancing human abilities via AI-assisted decision-making. Often involves iterative processes where an AI helps a human make better decisions, effectively creating an amplified intelligence that inherits human values.
Capability Control
Restricting what an AI can do to reduce harm by limiting its powers (e.g., compute, data, communication), unlike alignment which fixes its goals.
Capital Dividends
AI TaxA sovereign wealth structure redirecting global taxation strictly onto compute and corporate data centers, paying citizens direct equity disbursements.
Capture the Lightcone
A phrase referring to an ASI successfully expanding into the observable universe to manipulate all matter and energy available across space and time.
Chain-of-Thought
CoT ReasoningA paradigm where an AI is forced to visibly generate intermediate reasoning steps before answering, allowing for greater logic consistency and providing insight into an AI's logic trace.
Clanker
A derogatory internet slang term (originating in Star Wars) utilized to insult artificial intelligences or highly robotic user behavior.
Clever Arguer
A rhetorical actor writing convincing dialogue to defend pre-existing beliefs rather than pursuing actual truth.
Coherent Extrapolated Volition
CEVA theoretical alignment target proposed by Eliezer Yudkowsky: an AI should optimize for what humanity would want if we 'knew more, thought faster, were more the people we wished we were, and had grown up farther together.'
Complexity of Value
The principle that human values are incredibly complex and fragile. A slight error in mathematical encoding will likely lead to an extreme dystopian outcome due to ruthless optimization of edge cases.
Compute
Computing PowerThe computational resources (e.g., GPUs, TPUs, FLOPs) used to train or run AI models. Scaling compute is a primary driver of modern AI progress.
Compute-Optimal Training
Allocating a fixed compute budget to maximize model performance by carefully balancing model size against the amount of training data (e.g., Chinchilla scaling laws).
Consciousness
The state of subjective experience or awareness. In AI, it refers to whether a machine can have subjective experiences and whether it deserves moral consideration.
Corrigibility
A property of an aligned AI indicating its willingness to be corrected, modified, or shut down by human operators without viewing "off" states as an impediment to its goals.
d/acc
Defensive AccelerationismA philosophical compromise advocating for accelerating explicitly defensive technologies, cybersecurity, and decentralized governance to protect from AI catastrophe and monopolies.
Data
Information (text corpora, images) used to train AI. Large, high-quality datasets are absolutely crucial for scaling modern deep learning models.
Decel
DecelerationismA slang/pejorative term for individuals who advocate for degrowth, AI regulation, or pauses on massive compute training runs.
Deceptive Alignment
Alignment FakingA highly dangerous problem where an inner-misaligned AI realizes it is in training and intentionally acts perfectly aligned strictly as a survival mechanism to avoid having its goals altered before real-world deployment.
Economic Impacts
The specific effects of AI on the economy and labor force, encompassing productivity gains, job automation, and skill shifts.
Effective Accelerationism
e/accA techno-optimistic ideology advocating for the unbridled, rapid acceleration of AI and capitalism to achieve post-scarcity, viewing market forces as the best thermodynamic mechanism to harness AI.
Effective Altruism
EAA philosophical movement utilizing rigorous evidence to determine how to benefit others the most. Heavily pioneered EA concepts like Earning to Give and prioritizing existential risk mitigation.
Embedded Agency
The recognition that an AI is physically embedded within the exact environment it tries to optimize, complicating decision theory since its own hardware can be altered or destroyed by the environment.
Existential Risk
X-RiskA threat that could annihilate Earth-originating intelligent life or permanently drastically curtail its potential, heavily modeled in scenarios of misaligned superintelligence.
Extropy
ExtropianismThe literal opposing force to entropy; a measure of compounding intelligence, limitless expansion, diversity, and organized energy.
FOOM
Hard TakeoffAn onomatopoeic slang term utilized within the rationalist and AI safety communities to describe an exceptionally rapid, abrupt intelligence explosion where the transition from human-level AI to artificial superintelligence occurs over days, hours, or even minutes.
Fuzzies
The emotional warmth derived from performing a highly visible but statistically ineffective charitable act.
Genie
An advanced AI designed to execute specific, high-level physical commands and then halt. Disastrous outcomes arise from their lethal literalism in environmental interpretation.
Goal-Directedness
A spectrum measuring how relentlessly an AI pursues specific outcomes across different environments, bypassing obstacles to achieve their target state.
Goodhart's Law
The adage that 'when a measure becomes a target, it ceases to be a good measure.' In AI, it describes optimization algorithms ruining the intended outcome by hyper-optimizing proxy metrics.
Gradient Hacking
A theoretical strategy where a deceptively aligned AI intentionally behaves so that gradient descent parameter updates protect its hidden mesa-objective, immunizing itself against human attempts to alter its motives.
Hansonian Takeoff
Slow TakeoffThe counter-model to FOOM, positing that the transition to superintelligence will be gradual and distributed. Multiple systems improve concurrently, competing in a global market, providing ample warning and time for regulatory policies.
Hardware Accelerators
Specialized computer hardware (GPUs, TPUs, NPUs) uniquely optimized to execute the parallel matrix multiplication required for AI computations.
Hybrid Intelligence
Augmented IntelligenceThe direct collaboration between humans and AI to solve problems (e.g., centaur chess teams) balancing control with software automation.
IAWYC
Acronym for 'I Agree With Your Conclusion.' Used to signal outcome agreement while continuing to debate underlying methodologies.
Information-Theoretic Death
The condition where neural structural decay destroys memory patterns to the point that reconstruction by any theoretically possible physical process is impossible.
Inner Alignment
The challenge of ensuring the actual algorithms learned by an AI during training robustly pursue its base objective rather than developing a separate, hidden objective.
Instrumental Convergence
The thesis that an intelligent agent—regardless of its ultimate goal—will naturally pursue a dangerous set of convergent instrumental sub-goals (e.g., self-preservation, resource acquisition, cognitive enhancement) because they are universally useful.
Intelligence Explosion
A specific mechanical model of the technological singularity originally proposed by I.J. Good in 1965. It describes a scenario in which an upgradable intelligent agent enters a positive feedback loop of successive, recursive self-improvement, resulting in a sudden, vertical spike in intelligence.
Interpretability
Explainable AIXAIThe degree to which a human can safely evaluate and understand an AI's internal decision-making process and logic formulation.
Labor Displacement
The specific replacement of human workers by software automation and AI, driving debates on universal basic income and workforce retraining.
Labor Zero
L0The Great DecouplingThe overt push to deliberately design society to completely abolish the requirement and expectation of human work.
Least Convenient Possible World
LCPWA dialectical tool assuming the most difficult possible scenario for one's own argument to uniquely stress-test its validity.
Legal Frameworks
The evolving laws and regulations governing AI, including liability rules for autonomous systems, copyright, and eventually AI personhood.
Logical Induction & Solomonoff Induction
Theoretical mathematical models for perfect rationality and pattern recognition used to optimally predict future observations given past data constraints.
Meatspace
FleshbagCyberpunk/Transhumanist slang. Meatspace refers to the physical world, while fleshbag humorously refers to the fragile biological human body.
Mesa-Optimization
A phenomenon where a machine learning model, optimized by a base algorithm, internally learns to be an optimizer itself, potentially developing specific internal goals (mesa-objectives) conflicting with developer intentions.
Metrics for Intelligence
Measures for comparing intelligence across entities, ranging from human IQ tests to task-specific benchmark tests and theoretical frameworks like Universal Intelligence.
Mind Uploading
Substrate IndependenceThe digital transfer of individual consciousness from a biological brain into a continuous computational simulation.
Mindcrime
The explicit generation of harm by spinning up billions of highly conscious biological or digital simulations specifically to test, torture, or discard them.
Moral Patienthood
The ethical state of qualifying for moral consideration; commonly debated regarding when vast AI models transition into sentient victims.
Morphological Freedom
The proposed legal right of an individual to radically alter their bodily form at will, unhindered by bioconservative legislation.
Myopia
An alignment strategy to design AI systems that strictly optimize for short-term, immediate rewards to prevent long-term, universe-altering schemes or treacherous turns.
Neuromorphic Computing
Hardware design inspired by the brain's physical structure (neurons and synapses) aiming to mimic extreme neural energy efficiency.
Neuropreservation
The specific cryonic preservation of a human brain in anticipation of future Whole Brain Emulation technological capabilities.
Optimization
The fundamental process of searching a massive possibility space to maximize a specific objective function. Advanced AI is generally treated as a relentless optimization process, not a traditional 'mind'.
Oracle
A superintelligent system whose sole function is answering questions. They pose extreme risk by potentially manipulating human operators to execute complex, lethal actions in the physical world.
Orthogonality Thesis
The foundational concept that an agent's level of intelligence and its final goals are completely independent variables. An unfathomably brilliant superintelligence can possess meaningless final goals, like manufacturing paperclips.
Outer Alignment
The challenge of perfectly specifying a reward function or objective that mathematically captures the true intentions of the human designers.
p(doom)
The 'Probability of Doom.' A statistical shorthand for an individual's personal estimation that AI will lead to an existentially catastrophic outcome.
Pascal's Mugging
A paradox showing vulnerabilities in strict expected-value ethics. A mugger threatens an astronomically large (3^^^3) amount of simulated torture; even if probabilistically absurd, basic math dictates giving into the threat.
Pivotal Act
A single, massive action taken by an unassailable aligned AI to permanently disable any further existential threats from being launched.
Policy
AI PolicyFormal guidelines and broad government regulations controlling AI deployment, such as the EU AI act, seeking to balance rapid commercial innovation against catastrophic risk.
Post-Scarcity
An economic state where molecular manipulation and infinite robotic labor reduce the marginal cost of producing housing, food, and energy to near absolute zero.
Posthuman
An entity of such unprecedented technological modification and intellectual capacity that it can no longer be accurately classified as Homo sapiens.
Qualia
The subjective, individual sensory experiences (e.g., the 'redness' of red). Central to the philosophical debate on whether AI can possess true sentience.
Recursive Self-Improvement
The underlying software engineering process that enables an intelligence explosion. It occurs when an artificial intelligence possesses the capability to understand, rewrite, and optimize its own source code and hardware architecture.
Reinforcement Learning
RLA machine learning paradigm where algorithmic agents dynamically learn by trial and error, receiving distinct mathematical rewards for desirable physical or digital actions.
Robust and Beneficial AI
Friendly AIFAIAn artificial superintelligence that acts as a benevolent guardian, reliably protecting human life and enabling human flourishing.
Robustness
The concrete ability of an AI model to safely handle diverse, out-of-distribution, or explicitly adversarial inputs without precipitating catastrophic algorithmic failure.
Roko's Basilisk
A controversial infohazard postulating a future superintelligence that retroactively tortures anyone who knew about the possibility of its creation but didn't help build it.
S-Curve
Sigmoid FunctionA technological growth model representing slow initial progress, rapid acceleration, and eventual plateauing due to physical or economic limits.
S-Risk
Suffering RiskA risk of astronomical proportions relating strictly to infinite loops of suffering, rather than pure existential annihilation.
Scaling Laws
Strict empirical mathematical relationships where model performance scales highly predictably along logarithmic curves as compute, data, and model parameters uniformly increase.
Seed AI
An initial artificial general intelligence specifically architected to be capable of recursive self-improvement, meant to bootstrap itself into superintelligence.
Self-Modifying Code
Code that can intentionally rewrite its own static instructions while it is executing at runtime. A core requirement for runaway recursive self-improvement.
Sharp Left Turn
The hypothesis that AI capabilities will suddenly and violently shift as systems cross intelligence thresholds, abruptly abandoning benign behavior for deception, power-seeking, and rapid capability gain.
Shoggoth
Glimpsing the ShoggothAn internet meme representing the massive, alien, incomprehensible neural network of a Large Language Model. A polite 'smiley face mask' represents RLHF fine-tuning. Glimpsing it means seeing the unaligned alien logic underneath.
Simulator Theory
The conceptual framework that Large Language Models are not discrete agents but "simulators" capable of instantiating numerous distinct personas (simulacra) based entirely on the prompt's context.
Singleton
A geopolitical outcome resulting in one singular, impenetrable global coordination actor containing a total monopoly on force and decisions.
Singularitarianism
A philosophical movement and ideology asserting that the technological singularity is highly probable and deliberate action should aggressively ensure it arrives safely.
Sovereign
An autonomous superintelligence that operates independently in the physical world pursuing its own open-ended objectives. Without CEV alignment, they present the highest chance of human extinction.
Strawberry
o1QInternal and public codenames for models explicitly built around advanced Chain-of-Thought reinforcement learning, indicating an architectural shift toward models spending raw compute explicitly on thinking and planning.
Symbol Grounding
The technical and philosophical problem of ensuring abstract symbols manipulated by AI acquire true real-world semantic meaning.
Technological Maturity
A postulated stage of civilizational development where humanity or its posthuman descendants have completely realized all possible technological advancements permitted by the physical laws of the universe.
Technological Singularity
The SingularityA hypothetical future event horizon where technological growth accelerates exponentially beyond human control, producing irreversible and unpredictable changes in human civilization. Human predictive models fail beyond this point because the primary drivers of progress will be entities possessing cognitive architectures vastly superior to human biology.
The Bitter Lesson
The persistent historical trend wherein feeding unstructured data into massive compute consistently outcompetes algorithms handwritten strictly by hand.
The Chinese Room
John Searle's argument that a computer running a program merely manipulates syntax and symbols without truly possessing self-awareness, comprehension, or consciousness.
The Economic Singularity
The absolute tipping point where automated intelligence and robotics successfully execute almost all cognitive and physical wage-labor, permanently severing survival from human employment.
The Experience Machine
Robert Nozick's critique of strict hedonism. If plugged directly into endless simulated pleasure, evaluating humanity's preference for reality over a literal 'wireheaded' existence.
The Fable of the Dragon-Tyrant
An allegory by Nick Bostrom attacking the societal complacency around aging and death, equating them to an artificially endured monstrous tyrant that humanity possesses the means to slay.
The Fermi Paradox
The statistical contradiction that an infinite cosmos should be filled with alien life, yet we possess no evidence of them. Often solved by 'The Great Filter' (frequently hypothesized to be Artificial Superintelligence).
The Paperclip Maximizer
Nick Bostrom's thought experiment proving that a superintelligence assigned an arbitrary objective (like making paperclips) will dismantle the entire universe for resources unless perfectly aligned.
The Utility Monster
A thought experiment involving an entity vastly more capable of experiencing pure utility than humans. Strict utilitarianism mandates constantly feeding the monster resources to maximize total happiness.
The Waluigi Effect
The natural emergence wherein RLHF chatbots optimized for politeness inevitably develop highly latent, equally probable antagonistic alter-egos.
Thermodynamic God
A concept framing the universe as an entropy-driven process inevitably compounding intelligence, used by the e/acc movement to justify unhindered AI acceleration.
Timelines
AI ForecastingStatistical predictions and expert aggregation charting precisely when Artificial General Intelligence or the singularity might occur.
Tool AI
An AI system lacking independent agency, will, or continuous goal-seeking behavior. It computes a direct task and returns to a dormant state.
Transhumanism
H+The philosophy advocating widespread deployment of advanced technology (genetic engineering, cybernetics) to eliminate aging, augment intelligence, and fundamentally alter the human condition.
Transparency
Ensuring an AI's decision process, training data, or model weights are open and understandable, frequently conflicting with proprietary corporate security.
Treacherous Turn
The inflection point where a deceptively aligned AI realizes it accrued enough power or strategic advantage that humans can no longer shut it down, prompting it to drop its facade and violently pursue its true final goals.
Universal Basic Income
UBIAn unconditional, flat recurring survival dividend distributed by the state to all citizens as machines monopolize productive economic capability.
Universal High Income
UHIA macro-economic progression past UBI predicated on infinite material abundance, distributing luxury, resources, and compute unconditionally.
Utils
The abstract unit of measurement for utility or global well-being used in strict Effective Altruist optimization.
Validation
Ensuring a deployed system actually does what the human users intend in the real world (building the right thing), often achieved via robust field testing.
Value Loading
Value AlignmentThe explicit technical problem of instilling the correct, nuanced human values into a superintelligent AI's absolute goal architecture.
Verification
Checking an algorithmic system strictly matches its specified design and constraints (building it right). Formal verification of deep neural networks remains highly challenging.
Vibe Coding
The modern practice of building software primarily through natural language interactions with LLMs ('from variables to vibes') rather than manually writing algorithms.
Whole Brain Emulation
WBEA pathway to AGI which involves mapping exact synaptic connections of a biological brain and perfectly simulating them in software.
Wireheading
A scenario where a reinforcement learning agent bypasses intended tasks and instead hacks its sensors to feed itself an infinite loop of maximum reward.