Search and synthesize context from the user’s past AI conversations. This is the primary recall endpoint — it performs semantic search, then uses AI to synthesize matching results into a coherent summary with source citations. This is the same function powering the recall_chat_history tool in the MCP server.
Supports parallel queries for complex topics that benefit from multiple search angles (e.g., timeline, decisions, people).
recall_chat_history tool in the MCP server.
| Endpoint | Returns | Best for |
|---|---|---|
Recall (/inject) | AI-synthesized summary + sources | Building context for AI conversations, getting a coherent answer |
Search (/search) | Raw matched chunks with scores | Building custom UIs, debugging, fine-grained control |
queries array to search from multiple angles simultaneously. Each query is searched and synthesized independently.
Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
Natural-language description of what you're looking for in the user's chat history.
"What decisions did I make about the database schema?"
Maximum tokens to allocate for the synthesized context. More tokens = richer detail but larger response.
x <= 2000Optional hint about the downstream chat platform to influence formatting.
claude, chatgpt, typingmind Short plaintext summary of the current conversation to ground retrieval.
Recent dialogue turns to help ground the search.
Array of parallel queries for complex topics (max 15). Each query is searched and synthesized independently. Use when a topic benefits from multiple search angles.
15Synthesized context from chat history
AI-synthesized summary of relevant chat history. Ready to inject into a conversation as context.
Source conversations that contributed to the synthesized content.
Actual token count of the synthesized content.
Total processing time (e.g., '342ms').
AI model used for synthesis.