Glossary of terms

  • Updated

Core concepts

  • Actor – A dataset representing a business entity, for example, User, Account, or Document.
  • Alert – A message produced from a rule evaluation that interests a user. For example, an alert about the daily active user count dropping 20% below its 3-month average.
  • Analytic block (Block) – A building unit for composing arbitrarily complex analytical computations. Analytics offers a library of pre-defined block types.
  • Application – A logical grouping of objects for a particular business use case. Managing different use cases across teams or initiatives is an organizational convenience. Applications do not share any state among them.
  • Attribute – A column in an Analytics Dataset representing a qualitative or categorical property of the entity the dataset represents, for example, an age column in a user dataset or a device column in an event dataset.
  • Catalog – The repository of all objects organized into folders for an application. Users see parts of the catalog depending on their permissions. Each application has its own independent catalog structure.
  • Cohort – A group of actors sharing common properties or behaviors. For example, a cohort of enterprise users hosted two meetings within the first week of signup.
  • Dashboard – A collection of visualizations organized in a grid view. You typically use dashboards to monitor commonly used operational metrics.
  • Derived column – A calculated column of a dataset based on other columns of that dataset or related datasets. For example, a derived column of the User dataset called Engagement Cadence has daily, weekly, or monthly values for each user depending on their usage activity.
  • Warehouse-Native Analytics – Optimizely’s platform that connects directly to your cloud data warehouse (Snowflake, BigQuery, Redshift, Databricks), enabling direct querying and transformation with no data duplication.​
  • Connection – A connection acts as a bridge, enabling direct querying of a data warehouse from within the Analytics solution. It's essentially a secure pathway that lets Analytics to interact with your chosen data source, such as Snowflake, BigQuery, Amazon Redshift, or Databricks, to retrieve and analyze data.
  • Data source – The enterprise repository of data. Typically, it is a data warehouse such as Snowflake or BigQuery.
  • Exploration – Provides analytical insights, such as retention, conversion, churn, and engagement. It is built using an exploration template and includes a query definition and a resulting visualization.
  • Folder – Contains objects that you can save with a name. Folders logically organize objects such as datasets, explorations, dashboards, and so on. Folders are hierarchical, and a folder can contain sub-folders.
  • Funnel – Maps website visitors' flow to specific funnel steps that result in conversions or signups.
  • Group – A collection of users organized together for a business function. You can use groups to control users' access to objects and application privileges.
  • Measure – A column in an Analytics Dataset representing a quantitative measurement of business activity, such as the number of daily active users.
  • Notification – A notification is a message to a user about an alert triggered in the system. You receive notifications within the Analytics desktop browser or the Analytics Mobile application.
  • Organization – Represents the company that is a customer of Analytics.
  • Query – A query defines an analytical computation using SQL, NetScript, or templates. Executing a query produces results that you can visualize.
  • Rule – Defines one or more criteria for alert generation and the payload for generated alerts.
  • Sampling – A query performance optimization technique that uses a representative sample of data for computations instead of the entire data.
  • Segment – A category for breaking down analysis. For example, you can break down a retention analysis by segments of user cohorts or attributes.
  • Template – A pre-defined set of configurations for defining a particular object, such as an Exploration. Templates offer an easy way to author commonly used types of objects. Analytics offers templates for Explorations, Cohorts, Derived Columns, and Metrics.
  • User – A user is a person belonging to an organization that has access to the Analytics application.

Data modeling and configuration

  • Dataset – A tabular view (rows and columns) of data used for analytics. Datasets are logical views of data in your data source representing a business entity or activity, for example, User, Session, Ticket, Account, or Event.
  • Source Dataset – A representation of a logical mapping to a table, view, or SQL query in the data warehouse.
  • Derived Dataset – A dataset created by combining existing datasets (source or derived) to create a compound dataset.
  • Union Dataset – A logical combination of multiple warehouse tables exposed to users as a single logical dataset. A typical use case for union datasets is to create a single logical event dataset from a set of event tables in the warehouse, one for each event type.
  • User (Actor) Dataset – A dimension table representing users, with a relationship to the event dataset. Essential for event analysis.

  • Column Actor Dataset​ – A special dataset representing an actor without a dedicated table-defined purely by its presence in an event’s identifier column.

  • Event stream – A dataset representing a stream of events, such as user interactions in a mobile app.
  • Event type (Event) – A representation of values in an event stream dataset column that describes the event type.
  • Event stream annotation – Defines event datasets by identifying the following:

    • A timestamp column (for example, event_ts) to order events.

    • An event type column (for example, event_type) for display.

Infrastructure and performance

  • Service account and Materialization schema – A Service Account is used for secure access to your data warehouse, while a Materialization Schema defines where and how your data is stored in the warehouse after being processed.​
  • Materialization – The process of creating and storing materialized tables within a data warehouse to improve query performance. These materialized tables contain pre-computed results of frequently used queries, allowing for faster access to data when running reports and analyses.
  • Schema – The organization or structure of a database.

Experimentation and Metrics

  • Warehouse-Native Experimentation Analytics – Lets you integrate experiment data (Feature Experimentation or Web Experimentation) with warehouse-held business metrics to build a unified Experiment Scorecard-facilitating comparison of variations directly in your warehouse.

  • Decision Dataset – A dataset that captures variation assignment and decision events from Optimizely experiments.

  • Metrics – The reusable analytical measures built in Analytics, including the following:

    • Numeric aggregation (for example, SUM, AVG).

    • Conversion (event counts/users).

    • Ratio metrics (for example, event/event).

    • Custom formulas combining the previous.

  • Guardrail metrics – A secondary metric used to monitor the potential negative impact of an A/B test or experiment on other aspects of a product or user experience. They act as a safety net, alerting teams to unintended consequences that might arise from focusing solely on a primary success metric. Essentially, they help ensure that while optimizing for one thing, you're not inadvertently harming others.

  • Baseline – The reference variant in scorecards against which other variations are compared.

  • Experiment Scorecard – A visual report in Analytics showing raw values, percentage impact, statistical significance (through Stats Engine), and confidence intervals.

Stats engine and advanced analysis

  • Optimizely Stats Engine – A modern statistical approach used in Optimizely's experimentation platform. It combines sequential testing and false discovery rate control to provide faster and more accurate results in A/B testing and other experiments.
  • CUPED (Controlled-experiment Using Pre-Existing Data) – A statistical method that reduces variance in A/B tests, enhancing sensitivity and making it easier to detect differences between groups. CUPED helps you run smarter experiments that reach conclusions faster and with greater confidence.
  • Outlier management​ – A feature designed to improve the reliability of experiment results by mitigating the impact of extreme values (outliers) on metric calculations. It works by capping extreme values at a defined threshold or smoothing them based on percentiles, preventing a few unusual data points from skewing the overall results. This is particularly useful for metrics like revenue, where a few large purchases can disproportionately affect the average.
  • Sample Ratio Mismatch (SRM) detection​ – A feature that automatically monitors the distribution of traffic across different variations in an A/B test to identify potential imbalances. This helps detect early signs of issues in the experiment's configuration or implementation that could lead to biased results.