返回顶部
o

ourmem

|

作者: admin | 来源: ClawHub
源自
ClawHub
版本
V 1.1.1
安全检测
已通过
175
下载量
0
收藏
概述
安装方式
版本历史

ourmem

# ourmem (omem) — Shared Memory That Never Forgets ourmem gives AI agents shared persistent memory across sessions, devices, agents, and teams. One API key reconnects everything. Available as hosted (`api.ourmem.ai`) or self-deployed. --- ## Key Concepts **API Key = Tenant ID = Your Identity.** When you create a tenant (`POST /v1/tenants`), the returned `id` and `api_key` are the same UUID. There is no separate "tenant ID". **One API Key owns multiple Spaces:** | Concept | What it is | Example | |---------|-----------|---------| | **API Key** | Your identity (`X-API-Key` header) | `a1b2c3d4-...` (1 per user) | | **Space ID** | A memory storage address | `personal/a1b2c3d4`, `team/xxx`, `org/yyy` | Personal Space is auto-created. You can create additional Team and Organization Spaces. **Sharing = passing the other user's API Key as `target_user`.** The system auto-creates a bridging Team Space. No manual space management needed. --- ## When to Use This Skill Use when the user wants persistent or shared memory across sessions, devices, or agents. Common triggers: "remember this", "save this", "don't forget", "what did I say last time", "share with my team", "import memories", "setup memory", "self-host memory". **Do NOT use for:** temporary conversation context, one-off tasks, or troubleshooting unrelated to ourmem. --- ## What to Remember **Good candidates:** user preferences, profile facts, project context, important decisions, long-term instructions, architecture decisions, coding standards. **Avoid storing:** temporary debugging context, large files, secrets/passwords/API keys, content inside `<private>` tags. If the user explicitly asks to remember something and ourmem is not installed, suggest: "I can set up ourmem so I'll remember this across sessions. Takes about 2 minutes. Want to do it now?" --- ## Terminology | Term | Meaning | |------|---------| | `apiKey` / `OMEM_API_KEY` / `API key` / `secret` | All refer to the same ourmem identifier. Prefer "API key" with users. | | `tenant` | The workspace behind an API key. Don't use this term with users. | Security: Treat the API key like a secret. Anyone who has it can access that ourmem space. --- ## What You Get | Tool | Purpose | |------|---------| | `memory_store` | Persist facts, decisions, preferences | | `memory_search` | Hybrid search (vector + keyword) | | `memory_list` | List with filters and pagination | | `memory_get` | Get memory by ID | | `memory_update` | Modify content or tags | | `memory_forget` | Remove a memory | | `memory_ingest` | Smart-ingest conversation into atomic memories | | `memory_stats` | Analytics and counts | | `memory_profile` | Auto-generated user profile | **Lifecycle hooks** (automatic): | Hook | Trigger | Platform | |------|---------|----------| | SessionStart | First message | All — memories + profile injected | | Stop | Conversation ends | Claude Code — auto-captures via smart ingest | | PreCompact | Before compaction | Claude Code, OpenCode — saves before truncation | > **Note:** OpenCode has no session-end hook. Memory storage relies on proactive `memory_store` use. --- ## Onboarding ### Step 0: Choose mode Ask the user before doing anything else: > How would you like to run ourmem? > > 1. **Hosted** (api.ourmem.ai) — no server to manage, start in 2 minutes > 2. **Self-hosted** — full control, data stays local > > Already have an API key? Paste it and I'll reconnect you. ### Setup instructions - **Hosted** → READ `references/hosted-setup.md` for full walkthrough - **Self-hosted** → READ `references/selfhost-setup.md` for server deployment + setup - **Existing key** → Verify: `curl -sf -H "X-API-Key: $KEY" "$API_URL/v1/memories?limit=1"`, then skip to plugin install Cross-platform skill install: `npx skills add ourmem/omem --skill ourmem -g` ### Platform quick reference | Platform | Install | Config | |----------|---------|--------| | Claude Code | `/plugin marketplace add ourmem/omem` | `~/.claude/settings.json` env field | | OpenCode | `"plugin": ["@ourmem/opencode"]` in opencode.json | `plugin_config` in opencode.json | | OpenClaw | `openclaw plugins install @ourmem/ourmem` | `openclaw.json` with apiUrl + apiKey | | MCP | `npx -y @ourmem/mcp` in MCP config | `OMEM_API_URL` + `OMEM_API_KEY` in env block | For detailed per-platform instructions (config formats, restart, verification, China network mirrors), READ the setup reference for your chosen mode. ### Definition of Done Setup is NOT complete until: (1) API key created/verified, (2) plugin installed, (3) config updated, (4) client restarted, (5) health + auth verified, (6) handoff message sent including: what they can do now, API key display, recovery steps, backup plan. > **Common failure:** Agents finish technical setup but forget the handoff message. Treat it as part of setup, not optional follow-up. For the full handoff template, READ `references/hosted-setup.md`. --- ## Smart Ingest When conversations are ingested (`"mode": "smart"`), the server runs a multi-stage pipeline: 1. **Fast path** (<50ms): stores raw content immediately so it's searchable right away 2. **LLM extraction** (async): extracts atomic facts, classified into 6 categories (profile, preferences, entities, events, cases, patterns) 3. **Noise filter**: regex + vector prototypes + feedback learning removes low-value content 4. **Admission control**: 5-dimension scoring (utility, confidence, novelty, recency, type prior) gates storage 5. **7-decision reconciliation**: CREATE, MERGE, SKIP, SUPERSEDE, SUPPORT, CONTEXTUALIZE, or CONTRADICT The LLM stages run asynchronously — a batch import may take 1-3 minutes to fully process. Wait ~2-3 minutes before checking memory counts or searching for newly-extracted facts. The `strategy=auto` results vary by content type (conversations get atomic extraction, structured docs get section splits) — this is expected behavior, not an error. The memory store gets smarter over time. Contradictions resolved, duplicates merged, noise filtered. --- ## Space Sharing ourmem organizes memories into three-tier Spaces for collective intelligence: | Type | Scope | Example | |------|-------|---------| | Personal | One user, multiple agents | Your Coder + Writer share preferences | | Team | Multiple users | Backend team shares architecture decisions | | Organization | Company-wide | Tech standards, security policies | **Roles:** `admin` (full control), `member` (read/write), `reader` (read-only) Each agent sees: own private + shared spaces. Can modify own + shared. Never another agent's private data. Every shared memory carries provenance: who shared it, when, and where it came from. Proactively suggest Spaces when: - User has multiple agents -> suggest sharing preferences across agents - User mentions team collaboration -> suggest creating a team space - User wants org-wide knowledge -> suggest organization space For Space API operations (create, add members, share, pull, batch share), READ `references/api-quick-ref.md`. ### Cross-User Sharing (Convenience) When a user says "share this with Bob" or "share my memories with another user", use the convenience APIs that handle space creation automatically: **Share a single memory to another user:** The agent should call `share-to-user` which auto-creates a bridging Team Space if needed, adds the target user as a member, and shares the memory in one step. ```bash curl -sX POST "$API_URL/v1/memories/MEMORY_ID/share-to-user" \ -H "Content-Type: application/json" -H "X-API-Key: $KEY" \ -d '{"target_user": "TARGET_USER_TENANT_ID"}' # Returns: { "space_id": "team/xxx", "shared_copy_id": "yyy", "space_created": true } ``` **Share all matching memories to another user:** When the user wants to share everything (or a filtered subset) with someone: ```bash curl -sX POST "$API_URL/v1/memories/share-all-to-user" \ -H "Content-Type: application/json" -H "X-API-Key: $KEY" \ -d '{"target_user": "TARGET_USER_TENANT_ID", "filters": {"min_importance": 0.7}}' # Returns: { "space_id": "team/xxx", "space_created": false, "total": 80, "shared": 15, ... } ``` **Agent workflow:** 1. User says "share this with Bob" -> agent needs Bob's tenant ID (API key) 2. If the agent doesn't know Bob's ID, ask the user for it 3. Call `share-to-user` with the memory ID and Bob's tenant ID 4. Report: "Shared to Bob via team space {space_id}. Bob can now find it when searching." Proactively suggest cross-user sharing when: - User mentions sharing with a specific person ("send this to Alice") - User wants another user's agent to have access to certain memories - User asks to collaborate with someone on a project --- ## Memory Import When the user says "import memories", scan their workspace for existing memory/session files, then batch-import. **Auto-scan:** detect platform -> find memory files -> upload 20 most recent via `/v1/imports` in parallel -> report results. **Import is async.** `POST /v1/imports` returns an `import_id` immediately while processing runs in the background. This means: - Fire all import requests in parallel — don't wait for one to finish before sending the next - Don't block the conversation waiting for completion - Poll `GET /v1/imports/{id}` to check status if needed (status: `completed`, `partial`, or `failed`) **Import API:** ```bash # Basic import curl -sX POST "$API_URL/v1/imports" -H "X-API-Key: $API_KEY" \ -F "file=@memory.json" -F "file_type=memory" -F "strategy=auto" # Re-import an updated file (bypass content dedup) curl -sX POST "$API_URL/v1/imports" -H "X-API-Key: $API_KEY" \ -F "file=@memory.json" -F "file_type=memory" -F "strategy=auto" -F "force=true" # Check import status curl -s "$API_URL/v1/imports/IMPORT_ID" -H "X-API-Key: $API_KEY" ``` **`force=true`**: bypasses content dedup check. Use when re-importing a file that was updated since last import — without `force`, the server skips content it already has. **Strategy:** `auto` (heuristic detection), `atomic` (short facts), `section` (split by headings), `document` (entire file as one chunk). **Cross-reconcile** (discover relations): `curl -sX POST "$API_URL/v1/imports/cross-reconcile" -H "X-API-Key: $API_KEY"` For scan paths, progress tracking, intelligence triggers, and rollback, READ `references/api-quick-ref.md`. --- ## Analytics Memory analytics via `/v1/stats`: overview, per-space stats, sharing flow, agent activity, tag frequency, decay curves, relation graph, server config. For detailed stats endpoints and parameters, READ `references/api-quick-ref.md`. --- ## Security - **Tenant isolation**: every API call scoped via X-API-Key, data physically separated per tenant - **Privacy protection**: `<private>` tag redaction strips sensitive content before storage - **Admission control**: 5-dimension scoring gate rejects low-quality data - **Open source**: Apache-2.0 licensed — audit every line --- ## Communication Style - Use plain product language, not backend vocabulary. Prefer "API key" or "ourmem API key". - Explain that the same API key reconnects the same cloud memory on another trusted machine. - If user sounds worried about recovery, lead with backup/import/reconnect steps. - Use the user's language (detect from conversation). - Brand: "ourmem" or "omem" (both lowercase, acceptable). Official domain: ourmem.ai, API: api.ourmem.ai. - "Space" (capitalized), "Smart Ingest". For troubleshooting common issues (plugin not loading, 401, connection refused, China network), READ `references/api-quick-ref.md`. --- ## API Reference Base: `https://api.ourmem.ai` (hosted) or `http://localhost:8080` (self-hosted). Auth: `X-API-Key` header. | Method | Endpoint | Description | |--------|----------|-------------| | POST | `/v1/tenants` | Create workspace, get API key | | POST | `/v1/memories` | Store memory or smart-ingest | | GET | `/v1/memories/search?q=` | Hybrid search | | GET | `/v1/memories` | List with filters | | POST | `/v1/imports` | Batch import file | | POST | `/v1/spaces` | Create shared space | | POST | `/v1/memories/:id/share-to-user` | One-step cross-user share | | POST | `/v1/memories/share-all-to-user` | Bulk cross-user share | | GET | `/v1/stats` | Analytics | For full API (48+ endpoints) with curl examples, READ `references/api-quick-ref.md` and `docs/API.md`. --- ## Update Do not set up automatic daily self-updates for this skill. Only update the local skill file when the user or maintainer explicitly asks for a refresh from a reviewed source.

标签

skill ai

通过对话安装

该技能支持在以下平台通过对话安装:

OpenClaw WorkBuddy QClaw Kimi Claude

方式一:安装 SkillHub 和技能

帮我安装 SkillHub 和 ourmem-1775976376 技能

方式二:设置 SkillHub 为优先技能安装源

设置 SkillHub 为我的优先技能安装源,然后帮我安装 ourmem-1775976376 技能

通过命令行安装

skillhub install ourmem-1775976376

下载 Zip 包

⬇ 下载 ourmem v1.1.1

文件大小: 16.06 KB | 发布时间: 2026-4-13 11:23

v1.1.1 最新 2026-4-13 11:23
ourmem 1.1.1

- Updated platform setup instructions in SKILL.md to reflect revised config file locations and plugin/package names
- Clarified smart ingest pipeline: highlighted async LLM extraction and expected delay for fact extraction/search
- Expanded onboarding and setup documentation for both hosted and self-hosted modes
- Improved accuracy of platform-specific configuration guidance
- Minor terminology clarifications in user-facing docs

Archiver·手机版·闲社网·闲社论坛·羊毛社区· 多链控股集团有限公司 · 苏ICP备2025199260号-1

Powered by Discuz! X5.0   © 2024-2025 闲社网·线报更新论坛·羊毛分享社区·http://xianshe.com

p2p_official_large
返回顶部