From 13d0c1462ec44d2cedc49812adfd94fe928bc10d Mon Sep 17 00:00:00 2001 From: Harshith Mullapudi Date: Mon, 30 Jun 2025 23:35:58 +0530 Subject: [PATCH] Fix: add note regarding llama --- README.md | 46 ++++++++++++++++++------------------- mcp/package.json | 2 +- mcp/src/operations/graph.ts | 1 + 3 files changed, 25 insertions(+), 24 deletions(-) diff --git a/README.md b/README.md index ee27a60..5d295bb 100644 --- a/README.md +++ b/README.md @@ -4,8 +4,6 @@ - - ## C.O.R.E. **Contextual Observation & Recall Engine** @@ -17,9 +15,11 @@ C.O.R.E is built for two reasons: 1. To give you complete ownership of your memory, stored locally and accessible across any app needing LLM context. 2. To help SOL (your AI assistant) access your context, facts, and preferences for more relevant and personalized responses. -## Demo Video -[Check C.O.R.E Demo](https://youtu.be/iANZ32dnK60) +> **Note:** We are actively working on improving support for Llama models. At the moment, C.O.R.E does not provide optimal results with Llama-based models, but we are making progress to ensure better compatibility and output in the near future. +## Demo Video + +[Check C.O.R.E Demo](https://youtu.be/iANZ32dnK60) ## How is C.O.R.E different from other Memory Providers? @@ -37,7 +37,6 @@ With C.O.R.E, you see exactly what prices changed, who approved them, the contex Or ask: “What does Mike know about Project Phoenix?” and get a timeline of meetings, decisions, and facts Mike was involved in, with full traceability to those specific events. - ## C.O.R.E Cloud Setup 1. Sign up to [Core Cloud](https://core.heysol.ai) and start building your memory graph. @@ -97,30 +96,31 @@ Or ask: “What does Mike know about Project Phoenix?” and get a timeline of m 2. In Cursor, go to: Settings → Tools & Integrations → New MCP Server. 3. Add the CORE MCP server using the configuration format below. Be sure to replace the API_TOKEN value with the token you generated in step 1. - MCP configuration to add in Cursor - - ``` json - { - "mcpServers": { - "memory": { - "command": "npx", - "args": ["-y", "@redplanethq/core-mcp"], - "env": { - "API_TOKEN": "YOUR_API_TOKEN_HERE", - "API_BASE_URL": "https://core.heysol.ai", - "SOURCE": "cursor" - } - } - } - } - ``` + MCP configuration to add in Cursor + + ```json + { + "mcpServers": { + "memory": { + "command": "npx", + "args": ["-y", "@redplanethq/core-mcp"], + "env": { + "API_TOKEN": "YOUR_API_TOKEN_HERE", + "API_BASE_URL": "https://core.heysol.ai", + "SOURCE": "cursor" + } + } + } + } + ``` + 4. Go to Settings-> User rules -> New Rule -> and add the below rule to ensure all your chat interactions are being stored in CORE memory ``` After every interaction, update the memory with the user's query and the assistant's response to core-memory mcp. sessionId should be the uuid of the conversation ``` - + ## Connecting to the API You can also interact with C.O.R.E. programmatically via its APIs. diff --git a/mcp/package.json b/mcp/package.json index 065ff12..18f938f 100644 --- a/mcp/package.json +++ b/mcp/package.json @@ -1,6 +1,6 @@ { "name": "@redplanethq/core-mcp", - "version": "0.1.3", + "version": "0.1.5", "type": "module", "main": "dist/index.js", "bin": { diff --git a/mcp/src/operations/graph.ts b/mcp/src/operations/graph.ts index 4cf91ab..ca7cc1a 100644 --- a/mcp/src/operations/graph.ts +++ b/mcp/src/operations/graph.ts @@ -14,6 +14,7 @@ export async function ingestKnowledgeGraph(args: IngestKG) { const response = await axiosInstance.post(`/ingest`, { episodeBody: args.message, source: process.env.SOURCE, + referenceTime: new Date().toISOString(), sessionId: undefined, }); return response.data;