import { OllamaEmbeddings } from '@langchain/ollama'; import { NodeConnectionTypes, type INodeType, type INodeTypeDescription, type ISupplyDataFunctions, type SupplyData, } from 'n8n-workflow'; import { logWrapper } from '@utils/logWrapper'; import { getConnectionHintNoticeField } from '@utils/sharedFields'; import { ollamaDescription, ollamaModel } from '../../llms/LMOllama/description'; export class EmbeddingsOllama implements INodeType { description: INodeTypeDescription = { displayName: 'Embeddings Ollama', name: 'embeddingsOllama', icon: 'file:ollama.svg', group: ['transform'], version: 1, description: 'Use Ollama Embeddings', defaults: { name: 'Embeddings Ollama', }, ...ollamaDescription, codex: { categories: ['AI'], subcategories: { AI: ['Embeddings'], }, resources: { primaryDocumentation: [ { url: 'https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.embeddingsollama/', }, ], }, }, inputs: [], outputs: [NodeConnectionTypes.AiEmbedding], outputNames: ['Embeddings'], properties: [getConnectionHintNoticeField([NodeConnectionTypes.AiVectorStore]), ollamaModel], }; async supplyData(this: ISupplyDataFunctions, itemIndex: number): Promise { this.logger.debug('Supply data for embeddings Ollama'); const modelName = this.getNodeParameter('model', itemIndex) as string; const credentials = await this.getCredentials('ollamaApi'); const headers = credentials.apiKey ? { Authorization: `Bearer ${credentials.apiKey as string}`, } : undefined; const embeddings = new OllamaEmbeddings({ baseUrl: credentials.baseUrl as string, model: modelName, headers, }); return { response: logWrapper(embeddings, this), }; } }