mirror of
https://github.com/Abdulazizzn/n8n-enterprise-unlocked.git
synced 2025-12-18 10:31:15 +00:00
feat(core): Add telemetry for Evaluation feature (no-changelog) (#15562)
Co-authored-by: Yiorgis Gozadinos <yiorgis@n8n.io>
This commit is contained in:
@@ -4,12 +4,14 @@ import { useUsageStore } from '@/stores/usage.store';
|
||||
import { useAsyncState } from '@vueuse/core';
|
||||
import { PLACEHOLDER_EMPTY_WORKFLOW_ID } from '@/constants';
|
||||
import { useCanvasOperations } from '@/composables/useCanvasOperations';
|
||||
import { useTelemetry } from '@/composables/useTelemetry';
|
||||
import { useToast } from '@/composables/useToast';
|
||||
import { useI18n } from '@/composables/useI18n';
|
||||
import { useRouter } from 'vue-router';
|
||||
import { useNodeTypesStore } from '@/stores/nodeTypes.store';
|
||||
import { useEvaluationStore } from '@/stores/evaluation.store.ee';
|
||||
import { computed } from 'vue';
|
||||
import { useNodeTypesStore } from '@/stores/nodeTypes.store';
|
||||
|
||||
import { computed, watch } from 'vue';
|
||||
import { N8nLink, N8nText } from '@n8n/design-system';
|
||||
import EvaluationsPaywall from '@/components/Evaluations.ee/Paywall/EvaluationsPaywall.vue';
|
||||
import SetupWizard from '@/components/Evaluations.ee/SetupWizard/SetupWizard.vue';
|
||||
@@ -21,10 +23,11 @@ const props = defineProps<{
|
||||
const workflowsStore = useWorkflowsStore();
|
||||
const usageStore = useUsageStore();
|
||||
const evaluationStore = useEvaluationStore();
|
||||
const nodeTypesStore = useNodeTypesStore();
|
||||
const telemetry = useTelemetry();
|
||||
const router = useRouter();
|
||||
const toast = useToast();
|
||||
const locale = useI18n();
|
||||
const nodeTypesStore = useNodeTypesStore();
|
||||
|
||||
const { initializeWorkspace } = useCanvasOperations({ router });
|
||||
|
||||
@@ -32,13 +35,18 @@ const evaluationsLicensed = computed(() => {
|
||||
return usageStore.workflowsWithEvaluationsLimit !== 0;
|
||||
});
|
||||
|
||||
const showWizard = computed(() => {
|
||||
const runs = Object.values(evaluationStore.testRunsById ?? {}).filter(
|
||||
const runs = computed(() => {
|
||||
return Object.values(evaluationStore.testRunsById ?? {}).filter(
|
||||
({ workflowId }) => workflowId === props.name,
|
||||
);
|
||||
return runs.length === 0;
|
||||
});
|
||||
|
||||
const hasRuns = computed(() => {
|
||||
return runs.value.length > 0;
|
||||
});
|
||||
|
||||
const showWizard = computed(() => !hasRuns.value);
|
||||
|
||||
// Method to run a test - will be used by the SetupWizard component
|
||||
async function runTest() {
|
||||
try {
|
||||
@@ -54,6 +62,14 @@ async function runTest() {
|
||||
}
|
||||
}
|
||||
|
||||
const evaluationsQuotaExceeded = computed(() => {
|
||||
return (
|
||||
usageStore.workflowsWithEvaluationsLimit !== -1 &&
|
||||
usageStore.workflowsWithEvaluationsCount >= usageStore.workflowsWithEvaluationsLimit &&
|
||||
!hasRuns.value
|
||||
);
|
||||
});
|
||||
|
||||
const { isReady } = useAsyncState(async () => {
|
||||
try {
|
||||
await usageStore.getLicenseInfo();
|
||||
@@ -83,6 +99,33 @@ const { isReady } = useAsyncState(async () => {
|
||||
}
|
||||
}
|
||||
}, undefined);
|
||||
|
||||
watch(
|
||||
isReady,
|
||||
(ready) => {
|
||||
if (ready) {
|
||||
if (showWizard.value) {
|
||||
telemetry.track('User viewed tests tab', {
|
||||
workflow_id: props.name,
|
||||
test_type: 'evaluation',
|
||||
view: 'setup',
|
||||
trigger_set_up: evaluationStore.evaluationTriggerExists,
|
||||
output_set_up: evaluationStore.evaluationSetOutputsNodeExist,
|
||||
metrics_set_up: evaluationStore.evaluationSetMetricsNodeExist,
|
||||
quota_reached: evaluationsQuotaExceeded.value,
|
||||
});
|
||||
} else {
|
||||
telemetry.track('User viewed tests tab', {
|
||||
workflow_id: props.name,
|
||||
test_type: 'evaluation',
|
||||
view: 'overview',
|
||||
run_count: runs.value.length,
|
||||
});
|
||||
}
|
||||
}
|
||||
},
|
||||
{ immediate: true },
|
||||
);
|
||||
</script>
|
||||
|
||||
<template>
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
import { describe, it, expect, beforeEach } from 'vitest';
|
||||
import { describe, it, expect, beforeEach, vi } from 'vitest';
|
||||
import { mock } from 'vitest-mock-extended';
|
||||
import { createTestingPinia } from '@pinia/testing';
|
||||
import { createComponentRenderer } from '@/__tests__/render';
|
||||
@@ -6,11 +6,30 @@ import EvaluationRootView from '../EvaluationsRootView.vue';
|
||||
|
||||
import { useWorkflowsStore } from '@/stores/workflows.store';
|
||||
import { useEvaluationStore } from '@/stores/evaluation.store.ee';
|
||||
import { useUsageStore } from '@/stores/usage.store';
|
||||
import { mockedStore } from '@/__tests__/utils';
|
||||
import type { IWorkflowDb } from '@/Interface';
|
||||
import { waitFor } from '@testing-library/vue';
|
||||
import type { TestRunRecord } from '@/api/evaluation.ee';
|
||||
import { PLACEHOLDER_EMPTY_WORKFLOW_ID } from '@/constants';
|
||||
import { useTelemetry } from '@/composables/useTelemetry';
|
||||
import { EVALUATION_NODE_TYPE, EVALUATION_TRIGGER_NODE_TYPE, NodeHelpers } from 'n8n-workflow';
|
||||
|
||||
vi.mock('@/composables/useTelemetry', () => {
|
||||
const track = vi.fn();
|
||||
return {
|
||||
useTelemetry: () => ({
|
||||
track,
|
||||
}),
|
||||
};
|
||||
});
|
||||
|
||||
const getNodeType = vi.fn();
|
||||
vi.mock('@/stores/nodeTypes.store', () => ({
|
||||
useNodeTypesStore: vi.fn(() => ({
|
||||
getNodeType,
|
||||
})),
|
||||
}));
|
||||
|
||||
describe('EvaluationsRootView', () => {
|
||||
const renderComponent = createComponentRenderer(EvaluationRootView);
|
||||
@@ -37,6 +56,21 @@ describe('EvaluationsRootView', () => {
|
||||
|
||||
beforeEach(() => {
|
||||
createTestingPinia();
|
||||
vi.clearAllMocks();
|
||||
|
||||
vi.spyOn(NodeHelpers, 'getNodeParameters').mockReturnValue({
|
||||
assignments: {
|
||||
assignments: [
|
||||
{
|
||||
id: 'xxxxx',
|
||||
name: '=',
|
||||
value: '',
|
||||
type: 'string',
|
||||
},
|
||||
],
|
||||
},
|
||||
options: {},
|
||||
});
|
||||
});
|
||||
|
||||
it('should initialize workflow on mount if not already initialized', async () => {
|
||||
@@ -91,4 +125,231 @@ describe('EvaluationsRootView', () => {
|
||||
|
||||
await waitFor(() => expect(container.querySelector('.setupContent')).toBeTruthy());
|
||||
});
|
||||
|
||||
describe('telemetry', () => {
|
||||
it('should send telemetry event on mount with setup view when no test runs exist', async () => {
|
||||
const workflowsStore = mockedStore(useWorkflowsStore);
|
||||
const evaluationStore = mockedStore(useEvaluationStore);
|
||||
const usageStore = mockedStore(useUsageStore);
|
||||
|
||||
workflowsStore.workflow = mockWorkflow;
|
||||
evaluationStore.testRunsById = {};
|
||||
usageStore.workflowsWithEvaluationsLimit = 10;
|
||||
usageStore.workflowsWithEvaluationsCount = 0;
|
||||
|
||||
// Mock no evaluation nodes in workflow
|
||||
getNodeType.mockReturnValue(null);
|
||||
|
||||
renderComponent({ props: { name: mockWorkflow.id } });
|
||||
|
||||
await waitFor(() => {
|
||||
expect(useTelemetry().track).toHaveBeenCalledWith('User viewed tests tab', {
|
||||
workflow_id: mockWorkflow.id,
|
||||
test_type: 'evaluation',
|
||||
view: 'setup',
|
||||
trigger_set_up: false,
|
||||
output_set_up: false,
|
||||
metrics_set_up: false,
|
||||
quota_reached: false,
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
it('should send telemetry event on mount with overview view when test runs exist', async () => {
|
||||
const workflowsStore = mockedStore(useWorkflowsStore);
|
||||
const evaluationStore = mockedStore(useEvaluationStore);
|
||||
const usageStore = mockedStore(useUsageStore);
|
||||
|
||||
workflowsStore.workflow = mockWorkflow;
|
||||
evaluationStore.testRunsById = {
|
||||
run1: mock<TestRunRecord>({ workflowId: mockWorkflow.id }),
|
||||
run2: mock<TestRunRecord>({ workflowId: mockWorkflow.id }),
|
||||
};
|
||||
usageStore.workflowsWithEvaluationsLimit = 10;
|
||||
usageStore.workflowsWithEvaluationsCount = 1;
|
||||
|
||||
renderComponent({ props: { name: mockWorkflow.id } });
|
||||
|
||||
await waitFor(() => {
|
||||
expect(useTelemetry().track).toHaveBeenCalledWith('User viewed tests tab', {
|
||||
workflow_id: mockWorkflow.id,
|
||||
test_type: 'evaluation',
|
||||
view: 'overview',
|
||||
run_count: 2,
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
it('should send telemetry event with trigger_set_up true when dataset trigger node exists', async () => {
|
||||
const workflowsStore = mockedStore(useWorkflowsStore);
|
||||
const evaluationStore = mockedStore(useEvaluationStore);
|
||||
const usageStore = mockedStore(useUsageStore);
|
||||
|
||||
const workflowWithTrigger = mock<IWorkflowDb>({
|
||||
...mockWorkflow,
|
||||
nodes: [
|
||||
{
|
||||
id: 'trigger1',
|
||||
name: 'Dataset Trigger',
|
||||
type: EVALUATION_TRIGGER_NODE_TYPE,
|
||||
typeVersion: 1,
|
||||
position: [0, 0],
|
||||
parameters: {},
|
||||
},
|
||||
],
|
||||
});
|
||||
|
||||
workflowsStore.workflow = workflowWithTrigger;
|
||||
evaluationStore.testRunsById = {};
|
||||
usageStore.workflowsWithEvaluationsLimit = 10;
|
||||
usageStore.workflowsWithEvaluationsCount = 0;
|
||||
|
||||
// Mock dataset trigger node type exists
|
||||
getNodeType.mockImplementation((nodeType) =>
|
||||
nodeType === EVALUATION_TRIGGER_NODE_TYPE ? { name: EVALUATION_TRIGGER_NODE_TYPE } : null,
|
||||
);
|
||||
|
||||
renderComponent({ props: { name: mockWorkflow.id } });
|
||||
|
||||
await waitFor(() => {
|
||||
expect(useTelemetry().track).toHaveBeenCalledWith('User viewed tests tab', {
|
||||
workflow_id: mockWorkflow.id,
|
||||
test_type: 'evaluation',
|
||||
view: 'setup',
|
||||
trigger_set_up: true,
|
||||
output_set_up: false,
|
||||
metrics_set_up: false,
|
||||
quota_reached: false,
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
it('should send telemetry event with output_set_up true when evaluation set output node exists', async () => {
|
||||
const workflowsStore = mockedStore(useWorkflowsStore);
|
||||
const evaluationStore = mockedStore(useEvaluationStore);
|
||||
const usageStore = mockedStore(useUsageStore);
|
||||
|
||||
const workflowWithOutputNode = mock<IWorkflowDb>({
|
||||
...mockWorkflow,
|
||||
nodes: [
|
||||
{
|
||||
id: 'output1',
|
||||
name: 'Set Outputs',
|
||||
type: EVALUATION_NODE_TYPE,
|
||||
typeVersion: 1,
|
||||
position: [0, 0],
|
||||
parameters: {
|
||||
operation: 'setOutputs',
|
||||
},
|
||||
},
|
||||
],
|
||||
});
|
||||
|
||||
vi.spyOn(NodeHelpers, 'getNodeParameters').mockReturnValue({
|
||||
operation: 'setOutputs',
|
||||
});
|
||||
|
||||
workflowsStore.workflow = workflowWithOutputNode;
|
||||
evaluationStore.testRunsById = {};
|
||||
usageStore.workflowsWithEvaluationsLimit = 10;
|
||||
usageStore.workflowsWithEvaluationsCount = 0;
|
||||
|
||||
// Mock evaluation node type exists
|
||||
getNodeType.mockImplementation((nodeType) =>
|
||||
nodeType === EVALUATION_NODE_TYPE ? { name: EVALUATION_NODE_TYPE } : null,
|
||||
);
|
||||
|
||||
renderComponent({ props: { name: mockWorkflow.id } });
|
||||
|
||||
await waitFor(() => {
|
||||
expect(useTelemetry().track).toHaveBeenCalledWith('User viewed tests tab', {
|
||||
workflow_id: mockWorkflow.id,
|
||||
test_type: 'evaluation',
|
||||
view: 'setup',
|
||||
trigger_set_up: false,
|
||||
output_set_up: true,
|
||||
metrics_set_up: false,
|
||||
quota_reached: false,
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
it('should send telemetry event with metrics_set_up true when evaluation metrics node exists', async () => {
|
||||
const workflowsStore = mockedStore(useWorkflowsStore);
|
||||
const evaluationStore = mockedStore(useEvaluationStore);
|
||||
const usageStore = mockedStore(useUsageStore);
|
||||
|
||||
const workflowWithMetricsNode = mock<IWorkflowDb>({
|
||||
...mockWorkflow,
|
||||
nodes: [
|
||||
{
|
||||
id: 'metrics1',
|
||||
name: 'Set Metrics',
|
||||
type: EVALUATION_NODE_TYPE,
|
||||
typeVersion: 1,
|
||||
position: [0, 0],
|
||||
parameters: {
|
||||
operation: 'setMetrics',
|
||||
},
|
||||
},
|
||||
],
|
||||
});
|
||||
|
||||
vi.spyOn(NodeHelpers, 'getNodeParameters').mockReturnValue({
|
||||
operation: 'setMetrics',
|
||||
});
|
||||
|
||||
workflowsStore.workflow = workflowWithMetricsNode;
|
||||
evaluationStore.testRunsById = {};
|
||||
usageStore.workflowsWithEvaluationsLimit = 10;
|
||||
usageStore.workflowsWithEvaluationsCount = 0;
|
||||
|
||||
// Mock evaluation node type exists
|
||||
getNodeType.mockImplementation((nodeType) =>
|
||||
nodeType === EVALUATION_NODE_TYPE ? { name: EVALUATION_NODE_TYPE } : null,
|
||||
);
|
||||
|
||||
renderComponent({ props: { name: mockWorkflow.id } });
|
||||
|
||||
await waitFor(() => {
|
||||
expect(useTelemetry().track).toHaveBeenCalledWith('User viewed tests tab', {
|
||||
workflow_id: mockWorkflow.id,
|
||||
test_type: 'evaluation',
|
||||
view: 'setup',
|
||||
trigger_set_up: false,
|
||||
output_set_up: false,
|
||||
metrics_set_up: true,
|
||||
quota_reached: false,
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
it('should send telemetry event with quota_reached true when evaluations quota is exceeded', async () => {
|
||||
const workflowsStore = mockedStore(useWorkflowsStore);
|
||||
const evaluationStore = mockedStore(useEvaluationStore);
|
||||
const usageStore = mockedStore(useUsageStore);
|
||||
|
||||
workflowsStore.workflow = mockWorkflow;
|
||||
evaluationStore.testRunsById = {};
|
||||
usageStore.workflowsWithEvaluationsLimit = 5;
|
||||
usageStore.workflowsWithEvaluationsCount = 5; // At limit
|
||||
|
||||
// Mock no evaluation nodes in workflow
|
||||
getNodeType.mockReturnValue(null);
|
||||
|
||||
renderComponent({ props: { name: mockWorkflow.id } });
|
||||
|
||||
await waitFor(() => {
|
||||
expect(useTelemetry().track).toHaveBeenCalledWith('User viewed tests tab', {
|
||||
workflow_id: mockWorkflow.id,
|
||||
test_type: 'evaluation',
|
||||
view: 'setup',
|
||||
trigger_set_up: false,
|
||||
output_set_up: false,
|
||||
metrics_set_up: false,
|
||||
quota_reached: true,
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
Reference in New Issue
Block a user