Glossary

Quick, easy-to-understand definitions for everything from stats and analytics to data science and data engineering—so you can decode the tech talk and keep your insights sharp.

What word you are interested in?

Top searched category:
Data Engineering Data Science Decision Science
book img
  • A
  • B
  • C
  • D
  • E
  • F
  • G
  • H
  • I
  • J
  • K
  • L
  • M
  • N
  • O
  • P
  • Q
  • R
  • S
  • T
  • U
  • V
  • W
  • X
  • Y
  • Z
A

A/B Testing

A/B testing is a crucial method for comparing two versions of a digital ad, webpage, or app to determine which one performs better. It involves splitting the audience into two groups and analyzing their responses to different variations to make data-driven decisions.

Adaptive Cycle

A strategic model that explains how systems evolve over time—cycling through growth, stability, disruption, and renewal. Think of it as the natural rhythm of innovation, collapse, and transformation.
  • Applies to business lifecycles, ecosystems, and industries
  • Useful for scenario planning and resilience strategy
  • Helps leaders anticipate and ride waves of change

Agentic AI

AI systems designed to act as autonomous agents with goals, memory, and the ability to collaborate. They are capable of initiating actions, reasoning over tasks, and adapting in real time.   Agents can take on diverse roles such as planners, critics, reviewers, scouts, and orchestrators. Core to Mu Sigma’s Akashic Architecture.
  • Enables team-of-teams architecture in AI systems
  • Powers dynamic task decomposition and reallocation
  • Used in multi-agent collaboration for R&D, operations, and governance

Analysis of Variance (ANOVA)

A statistical test used to compare the means of three or more groups.

Anomaly Detection

The identification of unusual patterns or outliers in data that do not conform to expected behavior.

Artificial Intelligence (AI)

Artificial Intelligence (AI) refers to the simulation of human intelligence in machines, enabling them to perform tasks that typically require human intelligence. AI technologies include machine learning, natural language processing, and robotics, enhancing automation and decision-making processes.

Attractor

A state or behavior a system naturally settles into, even amidst chaos. For example, a company may repeatedly return to cost-cutting during market uncertainty.
  • Predicts likely future states of dynamic systems
  • Seen in markets, consumer behavior, and ecosystems
  • Helps identify recurring business patterns or traps

Autonomous Agents

Autonomous agents are AI-driven systems capable of performing tasks independently, adapting to their environment, and making real-time decisions without continuous human intervention.  Autonomous agents boost efficiency by managing repetitive tasks and making dynamic adjustments in real-time.
B

Bayesian Inference

A method for updating the probability of a hypothesis as more evidence or data becomes available. For example, in fraud detection, prior knowledge about normal behavior is updated with real-time transaction data to assess risk.
  • Used in medical diagnostics, A/B testing, and real-time prediction
  • Supports probabilistic reasoning under uncertainty
  • Foundation for adaptive decision-making

Big Data

Big data encompasses the vast volumes of data generated at high velocity and variety that traditional data processing software cannot handle efficiently. Advanced big data technologies and methodologies allow for the storage, analysis, and utilization of these massive datasets to derive actionable insights.

Business Intelligence (BI)

Business Intelligence (BI) involves using technologies and practices to collect, integrate, analyze, and present business data. BI tools provide historical, current, and predictive views of business operations, empowering organizations to make informed decisions.
C

Causal Inference

A set of techniques to uncover cause-and-effect relationships, not just correlation. For example, it helps determine if a change in price caused a drop in sales, versus both being influenced by a third factor like seasonality.
  • Enables counterfactual analysis and simulations
  • Applied in policy evaluation, marketing attribution, and economics
  • Basis for robust decision models in dynamic systems

Central Tendency

A statistical measure that identifies a single value as that is most representative of an entire distribution/set of data. Descriptors of central tendency are:
  • Mean (Average): The sum of all values in a dataset divided by the number of values. It represents the central point of the data. (Formula: Σx / n)
  • Median: The middle value in a dataset arranged from least to greatest. Useful for skewed data.

Classification

A supervised learning technique used to assign labels or categories to input data.

Clustering

An unsupervised learning method that groups similar data points into clusters based on shared features.

Cognitive Architecture

A blueprint for building intelligent agents that emulate human cognition. Often used in agent design to simulate reasoning, learning, and memory. It is a structured design language that mimics how humans think, learn, and make decisions that is used to build smarter AI systems. Imagine an AI assistant that reasons like a domain expert
  • Powers intelligent agents with reasoning capabilities
  • Combines memory, perception, and decision logic
  • Key to building AI systems that adapt and evolve

Complex Adaptive System (CAS)

A system that's made up of dynamic network of diverse, individual agents that interact, learn, and evolve in response to their environment. These systems are like an ecosystem, an economy, or an organization. They adapt through feedback and emergent behavior rather than central control.
  • Supports bottom-up innovation and resilience
  • Models used in simulations and scenario planning
  • Essential in high-entropy business and public systems

Complex Systems

Complex systems are networks of interacting components whose collective behavior is nonlinear, dynamic, emergent, and often unpredictable.

Complexity Science

Complexity science studies systems with many interconnected parts, focusing on how relationships and interactions give rise to collective behaviors and emergent phenomena. This field is applied across disciplines, from biology to social sciences, to understand complex adaptive systems and their dynamics.

Correlation

Correlation is a statistical measure that expresses the extent to which two variables change together at a constant rate.
  • Correlation Coefficient: A measure of the strength and direction of the linear relationship between two variables. Ranges from -1 (perfect negative correlation) to +1 (perfect positive correlation).
D

Data Analytics

Data analytics is the process of examining raw data to uncover patterns, trends, and insights that drive business strategies and decision-making. By transforming data into actionable insights, data analytics helps organizations enhance performance and gain a competitive edge.

Data Engineering

Data engineering involves designing, constructing, and maintaining systems and architecture that enable data collection, storage, and analysis. It focuses on creating data pipelines that transform raw data into usable information, supporting data-driven decision-making processes.

Data Governance

Data governance involves managing the availability, usability, integrity, and security of data within an organization. It ensures data accuracy, consistency, and responsible usage across the enterprise, supporting regulatory compliance and strategic decision-making.

Data Integration

Data integration combines data from different sources into a unified view, enabling comprehensive analysis and informed decision-making. It ensures that disparate data systems harmonize, providing a seamless flow of information across an organization.

Data Lake

A data lake is a centralized storage repository that holds vast amounts of raw data in its native format until needed for analysis. It supports storing structured, semi-structured, and unstructured data, providing flexibility for various analytical approaches.

Data Management

Data management encompasses the practices, architectural techniques, and tools used to achieve consistent access to and delivery of data across an organization. It ensures that data is treated as a valuable resource, enhancing its quality and usability for business processes.

Data Mining

Data mining is the process of discovering patterns, correlations, and anomalies within large data sets using statistical methods, machine learning, and database systems. It uncovers hidden knowledge and insights that can drive strategic business decisions and innovations.

Data Modeling

Data modeling involves creating a conceptual representation of data objects and their relationships, serving as a blueprint for constructing databases or data warehouses. It ensures data is structured effectively, facilitating efficient storage, retrieval, and analysis.

Data Science

Data science is an interdisciplinary field that uses scientific methods, algorithms, and systems to extract knowledge and insights from structured and unstructured data. Combining statistics, computer science, and domain expertise, it solves complex problems and informs data-driven decision-making.

Data Transformation

Data transformation is the process of converting data from one format or structure to another to make it suitable for analysis. This includes normalization, aggregation, and integration steps that prepare data for various analytical applications.

Data Visualization

Data visualization is the graphical representation of data and information using visual elements like charts, graphs, and maps. It helps stakeholders understand complex data sets by highlighting trends, outliers, and patterns in a visually intuitive manner.

Decision Science

Decision science applies analytical and computational methods to support and improve decision-making processes within organizations. Integrating data analysis, modeling, and behavioral science guides strategic and operational decisions, enhancing overall business performance.

Deep Learning

Deep learning is a subset of machine learning that uses neural networks with many layers to model complex patterns in data. It excels in tasks such as image recognition, natural language processing, and autonomous driving, mimicking the human brain's processing capabilities.

Descriptive Analytics

Descriptive analytics involves analyzing historical data to identify trends and patterns, providing insights into past performance. This type of analysis helps organizations understand what has happened over a specific period and informs future strategies.

Digital Twin

A digital twin is a virtual model of a real-world process, product, or system that uses AI to simulate and predict outcomes.  The virtual representation is continually updated with real-time data, providing insights into performance, potential issues, and optimization opportunities.  Digital twin technology is highly applicable in industries like manufacturing, logistics, airlines, and healthcare.

Distribution

In statistics, the distribution describes the relative numbers of times each possible data value will occur in a data set. Statistical distributions help us understand a problem better by assigning a range of possible values to the variables.
  • Normal Distribution (Bell Curve): A symmetrical, bell-shaped distribution where the mean, median, and mode are all equal. In a normal distribution, 68% of all values lie within one standard deviation from the mean. 95% of the values lie within two standard deviations from the mean, and 99.7% lie within three standard deviations from the mean.
E

Edge of Chaos

A transitional zone between order and disorder where systems are highly adaptive, creative, and capable of complex computation. The zone falls between order and randomness where complex systems exhibit the most adaptive and creative behavior. It is the sweet spot organizations must find where innovation, exploration, and evolution occur. Too much order leads to rigidity, too much chaos leads to collapse.
  • Found in markets, ecosystems, and organizational cultures
  • Ideal zone for adaptive AI and exploration engines
  • A visual metaphor: think of turbulent water—not frozen, not boiling

Emergence

A property of complex systems where higher-order behavior or patterns arise from interactions among simpler components. Often unpredictable from the individual parts alone. Emergence shows us the processes by which complex patterns and behaviors arise from simple interactions among components. For example, ant colonies build complex, giant nests without a central blueprint with small, simple acts performed by indicidual ants. Similarly, traffic moves spontaneously without accidents on freeways even though the indicidual vehicles are moving at high speed. From a birds-eye view, it looks like all the vehicle units are moving as one. Understanding emergence helps organizations:
  • Explain system-level intelligence without centralized control
  • Design decentralized AI systems

Ensemble Learning

A machine learning approach where multiple models are combined to produce stronger, more reliable results. For example, combining a decision tree, logistic regression, and a neural network can outperform any single model.
  • Boosts accuracy and reduces overfitting
  • Widely used in fraud detection, recommendation engines, and competitions like Kaggle
  • Ideal for high-stakes applications requiring robustness

ETL (Extract, Transform, Load)

ETL is a data pipeline process that extracts data from sources, transforms it into usable formats, and loads it into a target database or data warehouse.

Explainable AI (XAI)

Explainable AI (XAI) includes techniques that clarify how an AI model arrives at a specific outcome, making its decision-making process transparent and understandable.  XAI is critical for compliance, transparency, and user trust, especially in regulated sectors like finance and healthcare.
F

Feature Engineering

The process of selecting, transforming, or creating relevant input variables to improve model performance.

Feedback Loops

Mechanisms where a system learns from its own output, amplifying (positive loops) or correcting (negative loops) its behavior. Businesses use these in everything from pricing algorithms to employee engagement programs.
  • Enables self-correcting or self-reinforcing strategies
  • Powers AI systems, recommendation engines, and KPIs
  • Central to continuous improvement and agility

Fractal Geometry

A pattern that repeats at different scales, often used to describe self-similar structures in nature and complex systems like coastlines, clouds, and markets. Organizations often show fractal patterns in structure, behavior, or growth.
  • Models scale-invariant patterns in business systems
  • Used in forecasting, risk modeling, and systems design
  • Reveals hidden structure in messy environments
G

Generative AI

Generative AI is an area of artificial intelligence (AI) that focuses on creating or generating new content, such as images, music, text, or other creative outputs. It is powered by machine learning models, which are designed to understand and mimic the characteristics of the training data, allowing them to produce novel and unique outputs based on that understanding.
H

Hierarchical Reinforcement Learning (HRL)

A layered reinforcement learning technique where higher-level agents define goals for lower-level agents, allowing complex tasks to be broken down into manageable parts. AI learning is structured across multiple levels of abstraction or decision-making. Useful in complex multi-agent tasks. For example, in robotics, a top-level agent might set ""clean the room,"" while lower agents handle ""move arm"" or ""avoid obstacle.""
  • Speeds up learning by reusing sub-policies
  • Supports multi-agent coordination and abstraction
  • Powerful in planning, robotics, and autonomous systems

Hypothesis Testing

A type of statistical analysis in which you put your assumptions about a population parameter to the test.
  • Null Hypothesis (H₀): The statement we aim to disprove, typically that there is no difference between groups.
  • Alternative Hypothesis (H₁): The opposite of the null hypothesis, what we hope to prove.
  • P-value: The probability of observing a result at least as extreme as the one we obtained, assuming the null hypothesis is true. A low p-value suggests rejecting the null hypothesis.
  • Type I Error (Alpha): The probability of rejecting a true null hypothesis.
  • Type II Error (Beta): The probability of failing to reject a false null hypothesis.
  • Chi-Square Test: A statistical test used to determine if there is a significant association between two categorical variables.The test compares the observed values in your data to the expected values that you would see if the null hypothesis is true.
I

Inquisitive Analytics (Exploratory Analytics)

Inquisitive analytics is the practice of exploring data to discover underlying causes and relationships, going beyond surface-level observations. It involves detailed queries and analysis to understand why certain outcomes occurred, aiding in root cause analysis and problem-solving.
K

Knowledge Graph

A structured data model that represents real-world entities (like people, places, or concepts) and the relationships between them, enabling machines to understand context and meaning. Knowledge graphs enhance information retrieval, semantic search, and intelligent recommendations by connecting data points in a graph format.
L

Large Language Model (LLM)

Large language models are deep learning algorithms that can recognize, summarize, translate, predict, and generate human-like language using very large datasets.
M

Machine Learning (ML)

Machine learning (ML) is a subset of artificial intelligence (AI) that focuses on developing algorithms allowing computers to learn from and make predictions based on data. ML models improve their performance over time as they are exposed to more data, enhancing predictive accuracy.

Model Drift

Model drift occurs when a machine learning model's performance degrades over time due to changes in the underlying data distribution.

Model Interpretability

The degree to which a human can understand how an AI or ML model arrives at its decisions. A key aspect of ethical and regulated AI deployment. For leaders, it’s the difference between blind automation and accountable intelligence.
  • Essential for building trust in AI decisions
  • Required in regulated industries like finance and healthcare
  • Empowers humans to validate or challenge machine logic

Modularity

The degree to which a system’s components can be separated and recombined. Modular systems tend to be more adaptable and fault-tolerant. Modularity makes it easier to upgrade, scale, and repair systems. Think of it as Lego blocks for business architecture.
  • Increases agility and fault tolerance
  • Enables faster innovation and experimentation
  • Found in tech stacks, teams, and supply chains

Multi-Agent Systems (MAS)

Multi-agent systems consist of multiple autonomous agents working together within an environment to accomplish tasks, often through collaboration or competition.  Each agent operates based on its programming and objectives, but the collective system can solve complex problems.  MAS are essential for scenarios requiring decentralized decision-making, such as supply chain management, network optimization, and robotics.

Multi-Modal AI

AI systems that combine data from multiple sources, text, images, audio, video, to improve decision-making and contextual understanding. For example, analyzing a product review (text), a demo video, and sentiment (voice tone) together.
  • Powers copilots, autonomous vehicles, and diagnostics
  • Delivers richer, more human-like understanding
  • Crucial for complex customer or operational environments
N

Natural Language Processing (NLP)

Natural language processing (NLP) is a branch of artificial intelligence that enables computers to understand, interpret, and respond to human language. NLP techniques are used in applications like chatbots, sentiment analysis, and language translation, bridging the gap between human communication and machine understanding.

Neural Network

A machine learning model inspired by the human brain that processes data through interconnected nodes (neurons).

Neuro-Symbolic AI

A hybrid AI approach combining neural networks (which learn patterns from data) with symbolic reasoning (which handles logic, rules, and context). For example, a system might recognize objects in an image using deep learning and infer spatial relationships using symbolic rules.
  • Merges pattern recognition with reasoning capabilities
  • Enhances explainability and knowledge integration
  • Useful in legal AI, healthcare, and enterprise search

Nonlinearity

A condition where small inputs can trigger large, unpredictable outcomes or no change at all. For example, a slight price increase on a popular product could either be ignored or lead to a viral backlash, depending on timing and context.
  • Common in financial markets, ecosystems, and consumer behavior
  • Makes prediction challenging in complex systems
  • Visualized as a curve, not a straight line—responses aren’t proportional
O

Ontology

An ontology is a structured framework that defines the concepts, categories, and relationships within a specific domain of knowledge, enabling machines to interpret data with context, support semantic search, and power knowledge graphs through a shared vocabulary for consistent data integration and reasoning.

Outlier

A data point that falls significantly outside the overall pattern of the data.
P

Path Dependence

A concept where future outcomes are shaped heavily by historical choices, even when past conditions no longer apply. It’s why legacy tech stacks or strategic habits are hard to break.
  • Explains strategic inertia in organizations
  • Important in M&A, digital transformation, and governance
  • Helps identify constraints shaped by history

Predictive Analytics

Predictive analytics uses statistical techniques and machine learning algorithms to forecast future events based on historical data. It helps organizations anticipate outcomes and trends, enabling proactive decision-making and strategic planning.

Prescriptive Analytics

Prescriptive analytics combines predictive analytics with actionable recommendations to optimize decision-making. By suggesting the best course of action based on data analysis and predictive models, it supports effective strategy development and implementation.

Probability

The likelihood of an event occurring, expressed as a value between 0 (impossible) and 1 (certain).
Q

Question Network

A question network is a structured system that organizes and connects questions based on their semantic relationships, dependencies, or themes to enhance knowledge discovery, critical thinking, and machine reasoning. It enables users or algorithms to navigate complex topics by mapping how questions lead to deeper inquiry or connect across domains.
R

Reflexive Agents

AI agents capable of self-monitoring and adjusting their actions based on past behaviors or real-time feedback that are key to agentic feedback loops.
  • Enables continuous self-improvement in AI
  • Key to adaptive, agentic architectures
  • Supports use cases like fraud detection or dynamic pricing

Regression Analysis

Regression analysis is a set of statistical methods used to estimate relationships between a dependent variable and one or more independent variables.
  • Linear Regression: A statistical method to model the relationship between a dependent variable and one or more independent variables using a straight line.
  • R-squared: Represents the proportion of variance in the dependent variable explained by the independent variable(s) in a regression model.
  • Logistic Regression: A statistical method used to model the relationship between a binary dependent variable (e.g., yes/no) and one or more independent variables.

Reinforcement Learning (RL)

Reinforcement learning is a type of machine learning where an agent learns through trial and error, receiving rewards or penalties based on its actions.  It’s particularly effective for optimizing decision-making in complex, ever-changing environments.  RL can be applied in scenarios such as dynamic pricing, resource allocation, and personalized recommendations, offering adaptive strategies that respond to real-world changes.

Reinforcement Learning from Human Feedback - RLHF

A machine learning technique where AI models are trained to improve their performance by incorporating human evaluations (quality, safety, and desirability) of their outputs. This feedback is used to fine-tune the model’s behavior, aligning it more closely with human preferences, values, and real-world applications.

Resilience Engineering

A complexity science domain focused on designing systems that can adapt, recover, and thrive in the face of uncertainty or disruptions whether it’s a supply chain disruption or AI model failure. It helps systems evolve through volatility.
  • Core to enterprise risk and continuity planning
  • Builds fault-tolerant tech and operational ecosystems
  • Embraced in aerospace, pharma, and mission-critical systems

Retrieval Augmented Generation  - RAG

A machine learning technique where a language model improves its responses by first retrieving relevant information from external sources. This allows it to generate more accurate, context-aware answers using specialized or up-to-date knowledge beyond its training data.
S

Sampling

A subset of data drawn from a larger population. Used to estimate population characteristics.
  • Population Mean (μ): The true average value of a variable in the entire population.
  • Sample Mean (x̄): The average value of a variable in a sample. Used to estimate the population mean.
  • Confidence Interval: A range of values that is likely to contain the true population parameter (e.g., mean) with a certain level of confidence.

Self-Organization

The ability of a system to automatically arrange its internal structure or behavior without external control, often observed in complex adaptive systems such as teams, cities, or digital communities that often organize themselves based on shared incentives.
  • Drives bottom-up innovation and adaptability
  • Found in agile teams, marketplaces, and ecosystems
  • Useful for scaling without micromanagement

Semantic Layer

A contextual layer that connects raw data to business meaning usingontologies, metadata, and relationships. Powers explainability, traceability, and intelligent querying. A business-friendly abstraction that maps raw data to domain meaning. For example, mapping “cust_id” in raw logs to “Customer” in reports, ensuring consistency and context.
  • Improves traceability, queryability, and explainability
  • Powers data governance, BI tools, and AI pipelines
  • Core layer for intelligent decision systems

Statistical Significance

The probability of observing a statistically meaningful result, not due to chance.

Structured Data

Structured data is organized and formatted in a way that makes it easily searchable, typically found in relational databases.

Supervised Learning

Supervised Learning is a type of machine learning where the model is trained on a labeled dataset—which means each input has a corresponding correct output. The algorithm learns to map inputs to outputs by identifying patterns in the training data.

It is commonly used for:

  • Classification (e.g., spam detection, image recognition)
  • Regression (e.g., predicting house prices, stock forecasting)

Supervised learning is ideal when historical data with known outcomes is available and the goal is to make accurate future predictions.

Swarm Intelligence

A decentralized approach to problem-solving where collective behavior of agents, often modeled after biological systems, leads to adaptive and intelligent system outcomes. There is no central control, but the system is highly effective.
  • Used in routing, logistics, and robotic automation
  • Enables scalable, decentralized coordination
  • Mimics nature’s most efficient problem-solving systems

Systemic Risk

The potential for a failure in one part of a system to cascade and cause widespread disruption that is relevant in finance, supply chains, and infrastructure. An example is a bank collapse triggering a global crisis. Often invisible until it’s too late.
  • Common in financial systems, infrastructure, and ecosystems
  • Requires stress testing and network awareness
  • Critical concept for enterprise risk and strategy
T

T-Test

A statistical test used to compare the means of two groups, independent or paired is statistically significant or not.

Time Series Analysis

Methods for analyzing data points collected or recorded at specific time intervals.

Transformer

A Transformer is a breakthrough AI architecture that enables machines to process and generate human language with remarkable speed and accuracy. It powers modern Large Language Models (LLMs) and is the foundation for enterprise-grade Generative AI applications. Here are some if its salient features:
  • Uses self-attention to understand context across entire datasets, not just sequential inputs.
  • Enables parallel processing, making AI training faster and more scalable.
  • Backbone of LLMs like ChatGPT and Gemini.
  • Critical for NLP, automated decision-making, and unstructured data analysis.
  • Drives enterprise use cases in conversational AI, knowledge management, and AI copilots.
U

Unstructured Data

Unstructured data lacks a predefined format and includes content like emails, social media, video, audio, and text documents.

Unsupervised Learning

Unsupervised Learning is a type of machine learning where the model is trained on data without labeled outcomes. Instead of being told what to predict, the algorithm identifies hidden patterns, groupings, or structures in the input data on its own.

It is commonly used for:

  • Clustering (e.g., customer segmentation)
  • Dimensionality reduction (e.g., data visualization, noise reduction)
  • Anomaly detection (e.g., fraud detection)

Unsupervised learning helps uncover insights from data when explicit labels or outcomes are not available.

V

Variability

The difference exhibited by data points within a data set, as related to each other or as related to the mean. Three key measurements of variability are:
  • Range: The difference between the highest and lowest values in a dataset. Shows the spread of the data. (Formula: Maximum Value - Minimum Value)
  • Variance: The average squared deviation of all values from the mean. Measures how spread out the data is. (Formula: Σ(x - μ)² / n)
  • Standard Deviation (SD): The square root of the variance. Represents the average distance from the mean. (Formula: √(Σ(x - μ)² / n) )