BytesAgain βΊ Skills βΊ Langcache Semantic Caching for OpenClaw
π¦ ClawHub
Langcache Semantic Caching for OpenClaw
by @manvinder01
This skill should be used when the user asks to "enable semantic caching", "cache LLM responses", "reduce API costs", "speed up AI responses", "configure LangCache", "search the semantic cache", "store responses in cache", or mentions Redis LangCache, semantic similarity caching, or LLM response caching. Provides integration with Redis LangCache managed service for semantic caching of prompts and responses.
Skills are meant to be used inside your own AI agent. Install it, run a quick smoke test, then ask your agent to apply it to your real task.
1
Install into your agentCopy the ClawHub install command and run it where your OpenClaw/agent environment is configured.
2
Run a smoke testUse the test prompt below to confirm the skill loads and understands the workflow before relying on it.
3
Use it in your own agentPaste your actual task into Manus, OpenClaw, Claude Code, Cursor, or another agent that supports skills.
I just installed the Langcache Semantic Caching for OpenClaw skill. Please run a quick smoke test: explain what this skill can do, ask me for the minimum input it needs, then produce one small sample output for a realistic task.