Light evaluations#
Available on registered community and paid plans
Light evaluations are available to registered community users and on all paid plans.
What are light evaluations?#
When building your workflow, you often want to test it with a handful of examples to get a sense of how it performs and make improvements. At this stage of workflow development, looking over workflow outputs for each example is often enough. The benefits of setting up more formal scoring or metrics don't yet justify the effort.
Light evaluation allows you to run the examples in a test dataset through your workflow one-by-one, writing the outputs back to your dataset. You can then examine those outputs next to each other, and visually compare them to the expected outputs (if you have them).
How it works#
Requires Google Sheets
Evaluations use Google Sheets to store the test dataset. To use evaluations, you must configure a Google Sheets credential.
Light evaluations take place in the 'Editor' tab of your workflow, although you’ll find instructions on how to set it up in the 'Evaluations' tab.
Steps:
- Create a dataset
- Wire the dataset up to the workflow
- Write workflow outputs back to dataset
- Run evaluation
The following explanation will use a sample workflow that assigns a category and priority to incoming support tickets.
1. Create a dataset#
Create a Google Sheet with a handful of examples for your workflow. Your sheet should contain column(s) for:
- The workflow input
- (Optional) The expected or correct workflow output
- The actual output
Leave the actual output column or columns blank, since you'll be filling them during the evaluation.

2. Wire the dataset up to your workflow#
Insert an evaluation trigger to pull in your dataset#
Each time the evaluation trigger runs, it will output a single item representing one row of your dataset.
Clicking the 'Evaluate all' button to the left of the evaluation trigger will run your workflow multiple times in sequence, once for each row in your dataset. This is a special behavior of the evaluation trigger.
While wiring the trigger up, you often only want to run it once. You can do this by either:
- Setting the trigger's 'Max rows to process' to 1
- Clicking on the 'Execute node' button on the trigger (rather than the 'Evaluate all' button)
Wire the trigger up to your workflow#
You can now connect the evaluation trigger to the rest of your workflow and reference the data that it outputs. At a minimum, you need to use the dataset’s input column(s) later in the workflow.
If you have multiple triggers in your workflow you will need to merge their branches together.

3. Write workflow outputs back to dataset#
To populate the output column(s) of your dataset when the evaluation runs:
- Insert the 'Set outputs' action of the evaluation node
- Wire it up to your workflow at a point after it has produced the outputs you're evaluating
- In the node's parameters, map the workflow outputs into the correct dataset column

4. Run evaluation#
Click on the Execute workflow button to the left of the evaluation trigger. The workflow will execute multiple times, once for each row of the dataset:
Review the outputs of each execution in the Google Sheet, and examine the execution details using the workflow's 'executions' tab if you need to.
Once your dataset grows past a handful of examples, consider metric-based evaluation to get a numerical view of performance. See also tips and common issues.