This quick start guide walks you through the end-to-end setup—from creating a service account and accessing the platform to connecting your data warehouse, defining datasets, and building your first exploration. See Terms and Concepts for explanations of these terms.
Log in to Optimizely Analytics
- Access the Optimizely Analytics web application URL on any browser, enter your email, and click Next.
- Enter the password, verify your account, and sign in.
If you encounter difficulties logging into Optimizely Analytics, write an email to Optimizely Support.
Create a service account with read/write schema
Optimizely Analytics uses the service account to query the warehouse. The schema is used to improve performance by caching common computations, such as the drop-down list of unique values in a column. This substantially improves the warehouse cost and performance.
- Create a service account for Optimizely Analytics in your data warehouse of choice (Snowflake, BigQuery, Databricks, or Amazon Redshift) to query the warehouse.
- Add a schema with read/write access granted.
- Note the database and schema name for enabling materialization in Optimizely Analytics later.
Configure your data warehouse for Optimizely Analytics
Setting up your data warehouse for analytics is a crucial step in ensuring seamless data integration and analysis. This process involves configuring your warehouse to efficiently handle data operations and connect with Optimizely Analytics. Learn how to configure the following warehouses:
- Configure your Snowflake warehouse
- Configure your BigQuery warehouse
- Configure your Amazon Redshift warehouse
- Configure your Databricks warehouse
Create an application and connect your data warehouse
- Create an application within your organization for storing data and generated analyses.
- Create a connection to your data warehouse. You can also connect to multiple warehouses and select one of them to be the primary connection. The following warehouse options are available:
Enable Materialization
Analytics creates materialized tables in the data warehouse, which contain intermediate results for improving performance.
-
Go to Settings > General Settings > Materialization and enable the feature.
- Configure the following fields:
- Database – The database name in the data warehouse where the materialized tables are created.
- Schema – The schema's name in the database where the materialized tables are created.
-
Refresh Cron Schedule – The refresh periodicity of the materialized tables using the cron syntax. The recommended periodicity is daily, so the schedule is
0 0 * * *.
Create datasets and establish relationships between them
You must create user and event datasets and a relationship between them before creating explorations in Optimizely Analytics.
Create event datasets
Before you create a dataset, you must understand whether your events are represented by one or many tables in your warehouse. These categories are two common possibilities, but it is also possible that you may not fall under any of these categories.
Create a source dataset if the data you want to evaluate your experiments against lives in one table. If multiple tables represent your event, you can create a union dataset.
Create actor datasets
Optimizely Analytics requires a dimension table of users with one row per user. If you have a users table, you can use it to create an actor dataset.
If you lack a dimension table in the warehouse and only have an identifier column, like visitor_id in an event table, you can use it to create a column actor dataset.
Create a relationship between the event and actor datasets
Creating a relationship between datasets tells Analytics how to join these tables together.
- Go to Related Datasets > + Add Related Dataset.
- Select the event dataset – ProductEvents.
- Select one to many for Cardinality because one user will have multiple events.
- Select the columns. In the Users table, select the id column — this is usually the primary key. In the Events table, choose the column representing the ID of the user who performed that event – here, user_id.
- Click Save.
Build an exploration
When your datasets are ready, you can build an exploration using them. You can manually create an exploration or ask Opal to generate an exploration for you.
- Go to Settings > Defaults and select the Users dataset for the Actors dataset field, and for the Event Stream field, choose Product Events. When you set these defaults, users are populated by default as the selected actor for this measure. Now, you are all set to build your exploration. This should be done before you create an exploration.
- Click + > Event Segmentation.
- Select the Count of unique actors that performed event measure.
- Click Select Events and choose Play Content.
- Click Run to display a chart in the visualization window.
- Name the exploration and click Save.
Learn more about Explorations in Optimizely Analytics.
Performance guidance
The following techniques are various approaches you can take to ensure that Optimizely Analytics runs at an optimal cost and performance profile:
- Ensure your events/conversions table is clustered by event date (not time), and event type (the same column selected in the Semantics tab of the dataset in Optimizely Analytics), in that order.
- (Experimentation Analytics only) Check if the warehouse is clustered by the experiment ID column and decision date (not time) if it has a separate decisions table.
- Ensure that you have created a new schema in your warehouse, give Optimizely Analytics read and write access to it, and then enter the name of that schema in the Optimizely Analytics app settings, in the Materialization section. This enables the materialization of repeated queries, which is a large cost/performance boost.
- Ensure your warehouse instance size is appropriate.
- Consider creating a new warehouse for Optimizely Analytics, if this same warehouse instance is also used by other workloads, so that Optimizely Analytics queries are isolated from other workloads.
Please sign in to leave a comment.