返回顶部
p

ping-model

Measure and display AI model response latency. Use when the user types /ping or /ping followed by a model name to test round-trip time. Captures precise timing between command receipt and response generation, with smart duration formatting (ms, seconds, or minutes). Supports cross-model testing by temporarily switching models and measuring latency.

作者: admin | 来源: ClawHub
源自
ClawHub
版本
V 1.0.0
安全检测
已通过
263
下载量
0
收藏
概述
安装方式
版本历史

ping-model

# Ping Model Measure AI model response latency with consistent formatting. ## Quick Start ### Simple Ping (current model) ```bash bash command:"node {baseDir}/ping-model.js" ``` ### Ping Specific Model ```bash bash command:"node {baseDir}/ping-model.js --model minimax" ``` ### Compare Multiple Models ```bash bash command:"node {baseDir}/ping-model.js --compare kimi,minimax,deepseek" ``` ## Command Reference | Command | Description | |---------|-------------| | `/ping` | Ping current active model | | `/ping kimi` | Switch to kimi, ping, return | | `/ping minimax` | Switch to minimax, ping, return | | `/ping deepseek` | Switch to deepseek, ping, return | | `/ping all` | Compare all available models | ## Output Format **Required format - ALWAYS use this exact structure:** ``` 🧪 PING {model-name} 📤 Sent: {HH:MM:SS.mmm} 📥 Received: {HH:MM:SS.mmm} ⏱️ Latency: {formatted-duration} 🎯 Pong! ``` ### Latency Formatting Rules - **< 1 second**: Display as `XXXms` (e.g., `847ms`) - **≥ 1 second, < 60 seconds**: Display as `X.XXs` (e.g., `1.23s`) - **≥ 60 seconds**: Display as `X.XXmin` (e.g., `2.5min`) ### Examples **Fast response (< 1s):** ``` 🧪 PING kimi 📤 Sent: 09:34:15.123 📥 Received: 09:34:15.247 ⏱️ Latency: 124ms 🎯 Pong! ``` **Medium response (1-60s):** ``` 🧪 PING minimax 📤 Sent: 09:34:15.123 📥 Received: 09:34:16.456 ⏱️ Latency: 1.33s 🎯 Pong! ``` **Slow response (> 60s):** ``` 🧪 PING gemini 📤 Sent: 09:34:15.123 📥 Received: 09:35:25.456 ⏱️ Latency: 1.17min 🎯 Pong! ``` ## Cross-Model Testing When testing a non-active model: 1. Save current model context 2. Switch to target model 3. Execute ping 4. Measure latency 5. Restore original model 6. Display result **Critical:** Always return to the original model after testing. ## Comparison Mode ```bash bash command:"node {baseDir}/ping-model.js --compare kimi,minimax,deepseek,gpt" ``` Output format: ``` ══════════════════════════════════════════════════ 🧪 MODEL COMPARISON ══════════════════════════════════════════════════ 🥇 kimi 124ms 🥈 minimax 1.33s 🥉 deepseek 2.45s 4️⃣ openai 5.67s 🏆 Fastest: kimi (124ms) ``` ## Implementation The ping latency is measured as the time between: - T1: Message received by the agent - T2: Response ready to send This captures the model's internal processing time, not network latency.

标签

skill ai

通过对话安装

该技能支持在以下平台通过对话安装:

OpenClaw WorkBuddy QClaw Kimi Claude

方式一:安装 SkillHub 和技能

帮我安装 SkillHub 和 ping-model-1776235503 技能

方式二:设置 SkillHub 为优先技能安装源

设置 SkillHub 为我的优先技能安装源,然后帮我安装 ping-model-1776235503 技能

通过命令行安装

skillhub install ping-model-1776235503

下载 Zip 包

⬇ 下载 ping-model v1.0.0

文件大小: 4 KB | 发布时间: 2026-4-16 00:01

v1.0.0 最新 2026-4-16 00:01
Initial release: measure AI model response latency with smart duration formatting

Archiver·手机版·闲社网·闲社论坛·羊毛社区· 多链控股集团有限公司 · 苏ICP备2025199260号-1

Powered by Discuz! X5.0   © 2024-2025 闲社网·线报更新论坛·羊毛分享社区·http://xianshe.com

p2p_official_large
返回顶部