Skip to content

LM Studio LLM Deployment from Scratch (Ubuntu 24.04 + ROCm 7+)

This section explains how to deploy LLMs on Ubuntu 24.04 using LM Studio + ROCm version llama.cpp, and provides performance examples for Qwen3-8B Q4_K_M.

Before starting this section, make sure you have completed the environment setup and correctly installed ROCm 7.1.0 (refer to env-prepare-ubuntu24-rocm7.md).


1. Using LM Studio (with ROCm Version llama.cpp Backend)

1.1 Download LM Studio AppImage

First, download the installer from the official website:

bash
https://lmstudio.ai/

Download the latest .AppImage file to your local machine.

Screenshot:


1.2 Extract the AppImage

Extract the AppImage contents into the squashfs-root directory:

bash
chmod u+x LM-Studio-*.AppImage
./LM-Studio-*.AppImage --appimage-extract

1.3 Fix chrome-sandbox Permissions

Navigate to the squashfs-root directory and set the appropriate permissions for the chrome-sandbox file (this binary is required for the application to run securely):

bash
cd squashfs-root
sudo chown root:root chrome-sandbox
sudo chmod 4755 chrome-sandbox

1.4 Launch LM Studio

Start the LM Studio application from the current directory:

bash
./lm-studio

2. Install the ROCm Version llama.cpp Backend

In LM Studio, select the ROCm version of the llama.cpp backend to install:

Note the supported architecture list for the ROCm version of llama.cpp currently provided by LM Studio (GPU architecture support status):


3. Qwen3-8B Q4_K_M Performance Example

Load the Qwen3-8B Q4_K_M model in LM Studio, set the context length to 4096. Actual test results:

  • Approximately 36 tokens/s

Screenshot example: