LangChain

Memori Cloud supports any LangChain chat model. Each class has its own registration keyword: ChatOpenAI for OpenAI, ChatAnthropic for Anthropic, ChatBedrock for BedRock, ChatGoogleGenerativeAI for Google Gen AI models.

TypeScript support for LangChain is coming soon. The TypeScript SDK currently supports OpenAI, Anthropic, and Gemini.

Quick Start

LangChain Integration
from langchain_openai import ChatOpenAI
from memori import Memori

client = ChatOpenAI(model="gpt-4o-mini")

mem = Memori().llm.register(chatopenai=client)
mem.attribution(entity_id="user_123", process_id="langchain_agent")

response = client.invoke("Hello!")
print(response.content)

Different Providers

PackageChat ModelRegistration Keyword
langchain-openaiChatOpenAIchatopenai=client
langchain-anthropicChatAnthropicchatanthropic=client
langchain-google-genaiChatGoogleGenerativeAIchatgooglegenai=client
langchain-awsChatBedrockchatbedrock=client
LangChain Providers
from langchain_anthropic import ChatAnthropic
from memori import Memori

client = ChatAnthropic(model="claude-sonnet-4-5-20250929")
mem = Memori().llm.register(chatopenai=client)

Supported Modes

ModeMethod
Syncclient.invoke()
Asyncawait client.ainvoke()
Streamedclient.stream()