Memory
LangChain Memory Nodes
Last updated
Was this helpful?
LangChain Memory Nodes
Last updated
Was this helpful?
Memory allow you to chat with AI as if AI has the memory of previous conversations.
Human: hi i am bob
AI: Hello Bob! It's nice to meet you. How can I assist you today?
Human: what's my name?
AI: Your name is Bob, as you mentioned earlier.
Under the hood, these conversations are stored in arrays or databases, and provided as context to LLM. For example:
By default, UI and Embedded Chat will automatically separate different users conversations. This is done by generating a unique chatId
for each new interaction.
You can separate the conversations for multiple users by specifying a unique sessionId
For every memory node, you should be able to see a input parameter Session ID
In the /api/v1/prediction/{your-chatflowid}
POST body request, specify the sessionId
in overrideConfig
GET /api/v1/chatmessage/{your-chatflowid}
DELETE /api/v1/chatmessage/{your-chatflowid}
sessionId
string
sort
enum
ASC or DESC
startDate
string
endDate
string
All conversations can be visualized and managed from UI as well:
For OpenAI Assistant, will be used to store conversations.