import type { SafetySetting } from '@google/generative-ai'; import { ChatGoogleGenerativeAI } from '@langchain/google-genai'; import { NodeConnectionTypes } from 'n8n-workflow'; import type { NodeError, INodeType, INodeTypeDescription, ISupplyDataFunctions, SupplyData, } from 'n8n-workflow'; import { getConnectionHintNoticeField } from '@utils/sharedFields'; import { getAdditionalOptions } from '../gemini-common/additional-options'; import { makeN8nLlmFailedAttemptHandler } from '../n8nLlmFailedAttemptHandler'; import { N8nLlmTracing } from '../N8nLlmTracing'; function errorDescriptionMapper(error: NodeError) { if (error.description?.includes('properties: should be non-empty for OBJECT type')) { return 'Google Gemini requires at least one dynamic parameter when using tools'; } return error.description ?? 'Unknown error'; } export class LmChatGoogleGemini implements INodeType { description: INodeTypeDescription = { displayName: 'Google Gemini Chat Model', name: 'lmChatGoogleGemini', icon: 'file:google.svg', group: ['transform'], version: 1, description: 'Chat Model Google Gemini', defaults: { name: 'Google Gemini Chat Model', }, codex: { categories: ['AI'], subcategories: { AI: ['Language Models', 'Root Nodes'], 'Language Models': ['Chat Models (Recommended)'], }, resources: { primaryDocumentation: [ { url: 'https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatgooglegemini/', }, ], }, }, inputs: [], outputs: [NodeConnectionTypes.AiLanguageModel], outputNames: ['Model'], credentials: [ { name: 'googlePalmApi', required: true, }, ], requestDefaults: { ignoreHttpStatusErrors: true, baseURL: '={{ $credentials.host }}', }, properties: [ getConnectionHintNoticeField([NodeConnectionTypes.AiChain, NodeConnectionTypes.AiAgent]), { displayName: 'Model', name: 'modelName', type: 'options', description: 'The model which will generate the completion. Learn more.', typeOptions: { loadOptions: { routing: { request: { method: 'GET', url: '/v1beta/models', }, output: { postReceive: [ { type: 'rootProperty', properties: { property: 'models', }, }, { type: 'filter', properties: { pass: "={{ !$responseItem.name.includes('embedding') }}", }, }, { type: 'setKeyValue', properties: { name: '={{$responseItem.name}}', value: '={{$responseItem.name}}', description: '={{$responseItem.description}}', }, }, { type: 'sort', properties: { key: 'name', }, }, ], }, }, }, }, routing: { send: { type: 'body', property: 'model', }, }, default: 'models/gemini-2.5-flash', }, // thinking budget not supported in @langchain/google-genai // as it utilises the old google generative ai SDK getAdditionalOptions({ supportsThinkingBudget: false }), ], }; async supplyData(this: ISupplyDataFunctions, itemIndex: number): Promise { const credentials = await this.getCredentials('googlePalmApi'); const modelName = this.getNodeParameter('modelName', itemIndex) as string; const options = this.getNodeParameter('options', itemIndex, { maxOutputTokens: 1024, temperature: 0.7, topK: 40, topP: 0.9, }) as { maxOutputTokens: number; temperature: number; topK: number; topP: number; }; const safetySettings = this.getNodeParameter( 'options.safetySettings.values', itemIndex, null, ) as SafetySetting[]; const model = new ChatGoogleGenerativeAI({ apiKey: credentials.apiKey as string, baseUrl: credentials.host as string, model: modelName, topK: options.topK, topP: options.topP, temperature: options.temperature, maxOutputTokens: options.maxOutputTokens, safetySettings, callbacks: [new N8nLlmTracing(this, { errorDescriptionMapper })], onFailedAttempt: makeN8nLlmFailedAttemptHandler(this), }); return { response: model, }; } }