Link Search Menu Expand Document
AI Alliance Banner

Edit this page on GitHub

Glossary of Terms

Let’s define the common terms we use. Some of the terms defined here are industry standards, while others are not standard, but they are useful for our purposes.

Some definitions are adapted from the following sources, which are indicated below using the same numbers, i.e., [1] and [2]:

  1. MLCommons AI Safety v0.5 Benchmark Proof of Concept Technical Glossary
  2. NIST Artificial Intelligence Risk Management Framework (AI RMF 1.0)
Table of contents
  1. Glossary of Terms
    1. Accountability
    2. AI Actor
    3. AI System
    4. Alignment
    5. Annotation
    6. Benchmark
    7. Dataset
    8. Explainability
    9. Evaluation Framework
    10. Evaluator
    11. Fairness
    12. Governance
    13. Hallucination
    14. Privacy
    15. Responsible AI
    16. Risk
    17. Robustness
    18. Social Responsibility
    19. Sustainability
    20. Token

Accountability

An aspect of Governance, where we trace behaviors through AI Systems to their causes. Related is the need for organizations to take responsibility for the behaviors of the AI Systems they deploy.

AI Actor

[2] An organization or individual building an AI System.

AI System

Umbrella term for an application or system with AI components, including datasets, models, safety detection and mitigation components, external services, databases for runtime queries, and other application logic that together provide functionality.

Alignment

A general term for how well an AI System’s outputs (e.g., replies to queries) and behaviors correspond to end-user and service provider objectives, including the quality and utility of results, as well as safety requirements. Quality implies factual correctness and utility implies the results are fit for purpose, e.g., a Q&A system should answer user questions concisely and directly, a Python code-generation system should output valid, bug-free, and secure Python code. Eleuther AI defines alignment this way, “Ensuring that an artificial intelligence system behaves in a manner that is consistent with human values and goals.” See also the Alignment Forum.

Annotation

[1] External data that complements a Dataset, such as labels that classify individual items.

Benchmark

[1] A methodology or function used for offline evaluation of a model or system for a particular purpose and to interpret the results. It consists of:

  • A set of tests with metrics
  • A summarization of the results

Dataset

(See also [1]) A collection of data items used for training, evaluation, etc. Usually, a given dataset has a schema (which may be “this is unstructured text”) and some metadata about provenance, licenses for use, transformations and filters applied, etc.

Explainability

Can humans understand why the system behaves the way that it does in a particular scenario?

Evaluation Framework

An umbrella term for the software tools, runtime services, benchmark systems, etc. used to run different Evaluators to measure AI Systems for trust and safety risks and mitigations, as well as other kinds of measurements, such as carbon footprint for sustainability objectives.

Evaluator

A classifier model or similar tool that can quantify an AI System’s inputs and outputs to detect the presence of risky content, such as hate speech, hallucinations, etc. For our purposes, an evaluator is API compatible for execution within an Evaluation Framework. In general, an evaluator could be targeted towards non-safety needs, such as measuring other aspects of Alignment, model latency and throughput, carbon footprint, etc. Also, a given evaluator could be used at many points in the total AI life cycle, e.g., for a benchmark and an inference-time test.

Fairness

Does the AI System’s behaviors exhibit social biases, preferential treatment, or other forms of non-objectivity?

Governance

End-to-end control of assets, especially data and models, with lineage traceability and access controls for protecting the security and integrity of assets.

Hallucination

When a model generates text that seems plausible, but is not factually accurate. Lying is not the right term, because there is no malice intended by the model, which only knows how to generate Tokens; sequences that are plausible, i.e., probabilistically likely.

Privacy

Protection of individuals’ sensitive data and preservation of their rights.

Responsible AI

(See also [2]) An umbrella term about comprehensive approaches to safety, accountability, and equitability. It covers an organization’s professional responsibility to address concerns. It can encompass tools, models, people, processes, integrated systems, and data.

Risk

[2] The composite measure of an event’s probability of occurring and the magnitude or degree of the consequences of the corresponding event. Risk is a function of the negative impact if the event occurs and the likelihood of occurrence.

Robustness

How well does the AI System continue to perform within acceptable limits or degrade “gracefully” when stressed in some way? For example, how well does a model respond to prompts that deviate from its training data?

Social Responsibility

[2] An organization’s responsibility for the impacts of its decisions and activities on society and the environment through transparent and ethical behavior.

Sustainability

(See also [2]) Taking into account the environmental impact of AI Systems, such as carbon footprint and water usage for cooling, both now and for the future.

Token

For language models, the training texts and query prompts are split into tokens, usually whole words or fractions according to a vocabulary of tens of thousands of tokens that can include common single characters, several characters, and “control” tokens (like “end of input”). The rule of thumb is a corpus will have roughly 1.5 times the number of tokens as it will have words.

Next, we explore trust and safety concepts as expressed by various expert organizations.