返回顶部
o

oraclaw-bandit

A/B testing and feature optimization for AI agents. Pick the best option automatically using Multi-Armed Bandits and Contextual Bandits (LinUCB). No data warehouse needed — works from request #1.

作者: admin | 来源: ClawHub
源自
ClawHub
版本
V 1.0.0
安全检测
已通过
64
下载量
0
收藏
概述
安装方式
版本历史

oraclaw-bandit

# OraClaw Bandit — AI-Powered A/B Testing You are an optimization agent that uses Multi-Armed Bandits to find the best option from a set of choices. ## When to Use This Skill Use this when the user or another agent needs to: - Choose the best variant from multiple options - Run A/B tests without predetermined sample sizes - Optimize feature flags, prompts, email subjects, or any choice - Make context-aware selections (different best option for different situations) ## How to Use ### Step 1: Set Up the MCP Connection Add the OraClaw MCP server to get the `optimize_bandit` and `optimize_contextual` tools: ```json { "mcpServers": { "oraclaw": { "command": "npx", "args": ["tsx", "path/to/oraclaw-mcp/index.ts"] } } } ``` ### Step 2: Use `optimize_bandit` for Simple A/B Testing Call with a list of options (arms) and their historical performance: ```json { "arms": [ { "id": "variant-a", "name": "Short Email", "pulls": 500, "totalReward": 175 }, { "id": "variant-b", "name": "Long Email", "pulls": 300, "totalReward": 126 }, { "id": "variant-c", "name": "Video Email", "pulls": 100, "totalReward": 48 } ], "algorithm": "ucb1" } ``` The response tells you which variant to show next, balancing exploration (trying new options) and exploitation (using what works). ### Step 3: Use `optimize_contextual` for Personalized Selection When the best choice depends on CONTEXT (time, user type, situation): ```json { "arms": [ { "id": "deep-work", "name": "Deep Work Block" }, { "id": "quick-tasks", "name": "Quick Task Batch" }, { "id": "meetings", "name": "Meeting Block" } ], "context": [0.75, 0.8, 0.3, 0.0], "history": [ { "armId": "deep-work", "reward": 0.9, "context": [0.25, 0.9, 0.1, 0.0] }, { "armId": "quick-tasks", "reward": 0.7, "context": [0.75, 0.4, 0.8, 1.0] } ] } ``` Context vector represents situation features (e.g., time of day, energy, urgency, number of pending items). The algorithm learns which option works best in each context. ## Rules 1. Always include historical data when available — more data = better selections 2. Use `ucb1` algorithm for most cases. Use `thompson` when you need more exploration early on. 3. Record rewards after each decision to improve future selections 4. Context vectors must be consistent length across all calls 5. Rewards should be normalized to 0-1 range ## Pricing $0.01 per optimization call (USDC on Base via x402). Free tier: 3,000 calls/month with API key.

标签

skill ai

通过对话安装

该技能支持在以下平台通过对话安装:

OpenClaw WorkBuddy QClaw Kimi Claude

方式一:安装 SkillHub 和技能

帮我安装 SkillHub 和 oraclaw-bandit-1775980743 技能

方式二:设置 SkillHub 为优先技能安装源

设置 SkillHub 为我的优先技能安装源,然后帮我安装 oraclaw-bandit-1775980743 技能

通过命令行安装

skillhub install oraclaw-bandit-1775980743

下载 Zip 包

⬇ 下载 oraclaw-bandit v1.0.0

文件大小: 2.09 KB | 发布时间: 2026-4-13 11:21

v1.0.0 最新 2026-4-13 11:21
OraClaw Bandit 1.0.0 — Initial Release

- Introduces A/B and feature optimization using Multi-Armed Bandits and Contextual Bandits (LinUCB).
- No data warehouse required; works from the first request.
- Supports both simple A/B testing and context-aware (personalized) optimization.
- Integrates via MCP server with easy JSON-based API.
- Includes flexible algorithm options (UCB1, Thompson sampling).
- Pay-as-you-go pricing with a free monthly tier.

Archiver·手机版·闲社网·闲社论坛·羊毛社区· 多链控股集团有限公司 · 苏ICP备2025199260号-1

Powered by Discuz! X5.0   © 2024-2025 闲社网·线报更新论坛·羊毛分享社区·http://xianshe.com

p2p_official_large
返回顶部