1.6 AI Dock Hands-on Demo
This section documents three real conversations from Studio runtime, demonstrating AI Dock's interaction patterns and tool invocation processes. Case 1 covers single-device status collection and reporting; Case 2 demonstrates end-to-end YOLO deployment with detection results observed via a web interface; Case 3 generates a community forum draft based on hands-on experience.
Case 1: Full Device Health Check
User Input
Please perform a full health check for me: temperature, CPU/BPU, memory, disk, network, and critical service status.
Full Conversation

Key observations:
- The agent breaks down the task into 6 numbered steps, each corresponding to a clear intent.
- Every tool invocation displays the command name, parameters, and raw output (which can be copied and re-run in a remote terminal for verification).
- The final output is organized into three tables: Hardware Status, Network Status, and Critical Service Status.
- A conclusion and recommended actions are provided at the bottom.
Case 2: End-to-End YOLO Deployment and Verification
This case consists of two conversation segments: the first asks the agent to consult documentation, plan, and initiate deployment; the second restarts the process and verifies detection results via a web interface.
2.1 Planning and Launch
User Input
Run a YOLO example on the current device following official procedures, and tell me the expected output.
Full Conversation

Key observations:
- The agent integrates three information sources: built-in knowledge (OpenClaw expert), online documentation retrieval, and on-device measurements.
- Critical commands are first presented as an "executable plan" with expected outputs before actual execution.
- An "Execute" button provides secondary confirmation, allowing developers to review all details before authorizing actions.
2.2 Restart and Web Verification
User Input
Stop the previous YOLO instance, relaunch YOLOv8n, and open the web page for me to view.
Full Conversation

Key observations:
- Within a single conversation, the agent completes: process termination, restart, web page opening, and result validation.
- The floating TogetheROS Web Display window is a browser preview automatically opened by Studio, showing real detection results.
- Labels like
keyboard/bedare actual objects detected by YOLOv8n; CPU usage, temperature, and FPS metrics are displayed at the bottom.
Case 3: Generating a Community Forum Draft from Hands-On Experience
User Input
Summarize the process of successfully running YOLO-World into a community forum draft, including environment setup, steps, real-world performance, common issues, and advanced tips.
Conversation Output
![AI Dock generating a community forum draft: Title reads "[Copy-Paste Ready Draft]," body titled "Running YOLO-World Open-Vocabulary Object Detection on RDK X5 in 5 Minutes." Sections include—Environment & Preparation (Device: RDK X5 8GB; OS: Official RDK Ubuntu 22.04 + TROS Humble; Camera: Standard USB plug-and-play 720P; Software: tros-humble-hobot-yolo-world v0.4.1); Complete Execution Steps (1. Install software / 2. Prepare config & environment / 3. Launch detection service—with separate bash commands for USB camera, MIPI camera, and local image test versions); Real-World Performance Table (Inference FPS ~6 / Per-frame latency ~150ms / BPU utilization ~60% / Full-partition resource usage 76bit); Viewing Results (via http://device-IP:8000 or studio_open_url for pop-up window); Common Issues & Fixes (5 scenarios: browser not opening, successful launch but blank page, camera not working, unstable bounding boxes, no detections—with root causes and troubleshooting commands); Advanced Tips (custom detection categories via yolo_world_texts, saving results, further development); Conclusion recommends directly copying these steps.](http://rdk-doc.oss-cn-beijing.aliyuncs.com/doc/img/rdk_studio/en/showcase-case3-forum-draft.png)
Key observations:
- The draft follows a standard community post structure: environment, steps, real-world results, troubleshooting, and advanced usage.
- Bash commands, paths, and package versions in the steps are directly sourced from the actual execution in Case 2.
- Real-world performance metrics (FPS, latency, BPU usage) are extracted from logs by the agent.
- "Common Issues & Fixes" are derived from the agent’s internalized experience and document retrieval—not hallucinated.
Common Observations Across All Three Cases
| Dimension | Observation |
|---|---|
| Task Decomposition | Case 1: 6-step health check; Case 2: "Plan → Confirm → Execute → Verify"; Case 3: "Experience Reorganization → Forum Format" |
| Tool Invocation | Involved tools include SSH execution, file I/O, official doc retrieval, OpenClaw expert, and web URL opening |
| Raw Output | stdout/logs from every command are streamed in the conversation panel and can be copied for reproduction |
| Secondary Confirmation | Action-oriented tasks (e.g., service restart, critical commands) require clicking an "Execute" button for confirmation |
| Output Formats | Diverse outputs: tables, step-by-step lists, forum drafts, live web previews |
| Interruption Methods | "End Current / Stop All" buttons in panel + Esc key |
For technical details on tool invocation and device awareness, see 3.2.2 Device Awareness and Tool Invocation.
Reproduction Requirements
- Complete 2.1 Installation and Login and connect an RDK board.
- Activate an accessible model as described in 2.5 Configure AI Model.
- Case 2 requires a USB camera (or follow instructions to switch to MIPI/local image versions).
- Enter the prompts from this section into AI Dock.
Agent responses may vary slightly due to model randomness, on-device model library versions, and system state.