In-Memory Vector Store node#
Use the In Memory Vector Store node to store and retrieve embeddings in n8n's in-app memory.
On this page, you'll find the node parameters for the In Memory Vector Store node, and links to more resources.
Parameter resolution in sub-nodes
Sub-nodes behave differently to other nodes when processing multiple items using an expression.
Most nodes, including root nodes, take any number of items as input, process these items, and output the results. You can use expressions to refer to input items, and the node resolves the expression for each item in turn. For example, given an input of five name
values, the expression {{ $json.name }}
resolves to each name in turn.
In sub-nodes, the expression always resolves to the first item. For example, given an input of five name
values, the expression {{ $json.name }}
always resolves to the first name.
This node is different to AI memory nodes
The in-memory storage described here is different to the AI memory nodes such as Window Buffer Memory.
This node creates a vector database in the app memory.
Node parameters#
Operation Mode#
This Vector Store node has three modes: Get Many, Insert Documents, and Retrieve Documents. The mode you select determines the operations you can perform with the node and what inputs and outputs are available.
Get Many#
In this mode, you can retrieve multiple documents from your vector database by providing a prompt. The prompt will be embedded and used for similarity search. The node will return the documents that are most similar to the prompt with their similarity score. This is useful if you want to retrieve a list of similar documents and pass them to an agent as additional context.
Insert Documents#
Use Insert Documents mode to insert new documents into your vector database.
Retrieve Documents (For Agent/Chain)#
Use Retrieve Documents mode with a vector-store retriever to retrieve documents from a vector database and provide them to the retriever connected to a chain. In this mode you must connect the node to a retriever node or root node.
Get Many parameters#
- Memory Key: Enter the key to use to store the vector memory in the workflow data. n8n prefixes the key with the workflow ID to avoid collisions.
- Prompt: Enter the search query.
- Limit: Enter how many results to retrieve from the vector store. For example, set this to
10
to get the ten best results.
Insert Documents parameters#
- Memory Key: Enter the key to use to store the vector memory in the workflow data. n8n prefixes the key with the workflow ID to avoid collisions.
- Clear Store: Use this parameter to control whether to wipe the vector store for the given memory key for this workflow before inserting data (turned on).
Retrieve Documents (For Agent/Chain) parameters#
- Memory Key: Enter the key to use to store the vector memory in the workflow data. n8n prefixes the key with the workflow ID to avoid collisions.
Templates and examples#
Related resources#
Refer to LangChains's Memory Vector Store documentation for more information about the service.
View n8n's Advanced AI documentation.
AI glossary#
- completion: Completions are the responses generated by a model like GPT.
- hallucinations: Hallucination in AI is when an LLM (large language model) mistakenly perceives patterns or objects that don't exist.
- vector database: A vector database stores mathematical representations of information. Use with embeddings and retrievers to create a database that your AI can access when answering questions.
- vector store: A vector store, or vector database, stores mathematical representations of information. Use with embeddings and retrievers to create a database that your AI can access when answering questions.