Skip to content

🦞 OpenClaw - Fully Private Local AI Agent Platform AMD ROCm Edition

AMDOpenClaw

OpenClaw is a unified message processing and AI agent platform. Through its Gateway, it aggregates user messages from multiple channels (Lark, Telegram, iMessage, Slack, etc.) and routes them to Agents for processing, all running in isolated Workspaces. The platform adopts a modular design where Agents can flexibly invoke various tools and services.

OpenClaw (ClawX) project address: Link

OpenClaw Logo

Next, I will guide you step by step to deploy OpenClaw locally on an AMD 395 Max AI PC, building a fully private personal AI assistant!

Highlights

  • 🔒 Full Privacy: Local LLM + local OpenClaw, all data stays on your machine
  • 🦞 Multi-Channel Gateway: Unified access via Lark, Telegram, DingTalk, Slack, iMessage
  • 🧠 Agent Architecture: Modular design with memory, soul, skills context management
  • 🏠 Local Deployment: Run 35B parameter models on AMD 395 Max, no internet required

OpenClaw Platform Overview

OpenClaw has become an industry benchmark with its comprehensive architecture and ambitious "OS endgame" vision, sparking phenomenal interest in the AI Agent space. With over 430,000 lines of code, it covers core capabilities including message gateway, Agent runtime, tool invocation, and memory management.

Figure 5.6.1 OpenClaw Platform Architecture

Figure 5.6.1 OpenClaw Platform Architecture

Ecosystem Comparison

OpenClaw's success has also spawned community-derived projects:

ProjectPositioningFeatures
OpenClawFull-featured platform430K lines of code, OS-level architecture
NanobotMinimalismBy HKU team (HKUDS), lightweight approach
PicoClawEdge computingGo language, single binary, minimal resource usage
Figure 5.6.2 OpenClaw Ecosystem Comparison

Figure 5.6.2 OpenClaw Ecosystem Comparison

Context Usage

Even starting a new conversation, OpenClaw consumes 39K tokens of context — it carries extensive memory.md, soul.md, user.md along with skills descriptions and tool descriptions. OpenClaw heavily tests an Agent's long-context tool invocation and planning capabilities, so local deployment requires a machine with substantial VRAM!

Figure 5.6.3 OpenClaw Context Usage

Figure 5.6.3 OpenClaw Context Usage (39K tokens for a new conversation)

Step 1: Hardware Preparation - AMD 395 Max

The AMD 395 Max is ideal for both gaming and serving as a home AI hub, ensuring both model performance and local chat privacy.

Figure 5.6.4 AMD 395 Max Specifications

Figure 5.6.4 AMD 395 Max Hardware Specifications

The base environment for this guide is as follows:

----------------
AMD Ryzen™ AI 9 HX 395 Max 
LM Studio
Windows 11 / Linux
----------------

This guide assumes you are using an AMD 395 Max AI PC or other devices with AMD ROCm-supported GPUs.

Step 2: Deploy Local Model with LM Studio

LM Studio is a cross-platform desktop application designed for running large language models locally, supporting Windows, macOS, and Linux. It provides an intuitive GUI for searching, downloading, and managing open-source models (e.g., GGUF format), enabling offline conversations, inference testing, or API server hosting on personal devices without internet access.

Figure 5.6.5 LM Studio Interface

Figure 5.6.5 LM Studio Interface

2.1 Download and Install LM Studio

Visit LM Studio website to download and install.

This tutorial recommends using the gemma-4-26b-a4b model (MoE architecture, only 3B active parameters, 35B total).

Search and download the GGUF version in LM Studio, then note:

  • Base URL: http://127.0.0.1:12345/v1
  • Model ID: gemma-4-26b-a4b

For detailed LM Studio configuration, refer to Getting Started with ROCm Deploy

Step 3: Install OpenClaw (ClawX)

3.1 Download Installer

Go to the ClawX Release page and download the installer for your system (Windows or Mac).

Figure 5.6.6 ClawX Download Page

Figure 5.6.6 ClawX Release Download Page

3.2 Install ClawX

Double-click the installer and choose to install for the current user only:

Figure 5.6.7 Install ClawX - StartFigure 5.6.7 Install ClawX - User Selection

Figure 5.6.7 Double-click to install, select current user only

Choose the installation location and wait for completion:

Figure 5.6.8 Choose LocationFigure 5.6.8 Installing

Figure 5.6.8 Choose installation location and wait for completion

3.3 Initial Configuration

After installation, launch the app, select your language, and click Next:

Figure 5.6.9 Select Language

Figure 5.6.9 Select language and click Next

Wait for the environment check — any missing dependencies will be guided for installation, then click Next:

Figure 5.6.10 Environment Check

Figure 5.6.10 Environment check (missing components will be auto-guided)

3.4 Configure Local Model

Select Custom local model deployment and use the gemma-4-26b-a4b deployed via LM Studio:

Figure 5.6.11 Model Config - CustomFigure 5.6.11 Model Config - Verify

Figure 5.6.11 Select custom model and configure connection details

Enter the following configuration:

SettingValue
Base URLhttp://127.0.0.1:12345/v1
Model IDgemma-4-26b-a4b

After verification, click Next to complete the installation:

Figure 5.6.12 Installation Complete

Figure 5.6.12 Click Next, installation complete

3.5 Installation Success

Congratulations! Your fully local "lobster" is ready! 🦞

Figure 5.6.13 OpenClaw Ready

Figure 5.6.13 Local OpenClaw installation successful

Step 4: Using OpenClaw

4.1 Web Management Interface

Click "OpenClaw Page" in the bottom-left corner to access the web management interface, where you can chat and manage Agents. You can also chat and manage directly from the ClawX desktop client:

Figure 5.6.14 OpenClaw Web Interface

Figure 5.6.14 OpenClaw Web Management Interface

4.2 Configure Message Channels

In channel settings, you can configure Lark, DingTalk, and other messaging platforms step by step for unified multi-platform message access:

Figure 5.6.15 Configure Channel - LarkFigure 5.6.15 Configure Channel - DingTalk

Figure 5.6.15 Configure Lark/DingTalk message channels

Summary

Figure 5.6.16 Full Privacy Solution

Figure 5.6.16 Local LLM + Local OpenClaw = Fully Private Personal Assistant

Through this tutorial, you have successfully built a fully localized AI assistant solution on AMD 395 Max:

  • LM Studio running gemma-4-26b-a4b locally
  • OpenClaw (ClawX) providing Agent platform and multi-channel message gateway
  • All data stays on your machine, privacy fully under your control

FAQ

Q: How much VRAM is needed?

OpenClaw consumes 39K tokens of context for a new conversation. Combined with the model parameters, at least 24GB VRAM is recommended. The AMD 395 Max's shared memory architecture can meet this requirement.

Q: Which local models are supported?

Any local model compatible with the OpenAI API format can be used, including models deployed via LM Studio, Ollama, vLLM, etc. This tutorial recommends gemma-4-26b-a4b (MoE architecture, only 4B active parameters).

Q: Besides Lark and DingTalk, what other channels are supported?

OpenClaw supports Lark, Telegram, iMessage, Slack, DingTalk, and more. Configure them through the channel settings with guided setup.

References