π¦ ClawHub
getmem.ai Memory
by @nimblev2023
Persistent memory for AI agents via getmem.ai. Call mem.get() before each LLM call to inject context, and mem.ingest() after each turn to save the conversation.
π‘ Examples
import getmem_ai as getmem, osmem = getmem.init(os.environ["GETMEM_API_KEY"])
Before each LLM call β get relevant memory context
result = mem.get(user_id, query=user_message)
context = result["context"] # inject into system promptAfter each turn β save both user + assistant messages
mem.ingest(user_id, messages=[
{"role": "user", "content": user_message},
{"role": "assistant", "content": reply},
])
βοΈ Configuration
Set your API key in the environment:
export GETMEM_API_KEY=gm_live_YOUR_KEY_HERE
Get your key at https://platform.getmem.ai β $20 free credit on signup.
TERMINAL
clawhub install getmem