Overview
Adding Your First Node
Desktop (Windows/Mac/Linux)
- Download RemoteXPU from rxpu.nixosoft.com/download
- Install and launch โ it auto-starts in the system tray
- Sign in with Google โ your node appears here automatically
Headless Server (Linux/macOS)
curl -fsSL https://rxpu.nixosoft.com/setup-node.sh | bash
Then enter your cluster API key. Get it from rxpu.nixosoft.com โ Settings โ API Keys
Accessing a Remote Node's Dashboard
Once nodes are connected on VPN, click "Dashboard" on any node in the table below. This proxies through your local RXPU daemon โ VPN must be connected.
CPU Servers (LiteLLM Routers)
curl -fsSL https://rxpu.nixosoft.com/setup.sh | bash on your CPU server.
| Server | IP:Port | CPU | RAM | LiteLLM | Health | XPU Nodes | Status | Last Seen |
|---|---|---|---|---|---|---|---|---|
|
GB total
GB used
|
GPU Model Breakdown
| GPU Model | Count | Total VRAM |
|---|---|---|
| GB |
Online Nodes
Auto-refreshes every 10sXPU Nodes
| Node Name โ | Owner | GPU Model | VRAM | VPN IP โ | Mode | IP:Port | Status | Last Heartbeat | Remote Access | Actions |
|---|---|---|---|---|---|---|---|---|---|---|
|
OS Platform
Driver Version
CUDA Version
Registered
Mode
Pool Min VRAM
Pool GPU Filter
API Key
Node ID
User ID
|
||||||||||
Cluster Models โ
| Model | Size | Runs On | Compatible Devices | Speed / Quality | API Key | Enabled | Actions |
|---|---|---|---|---|---|---|---|
|
๐ GPU
๐ต CPU
๐ฃ NPU
|
/
/
โ
|
|
|
|
Assign XPU Node to CPU Server
Node:
Set Pool Filters
Node:
Add Cluster Model
What are you trying to do?
Manually configure a model. It will auto-download to XPU nodes when Ollama starts.
Edit Model โ
โก Cross-Device Benchmark โ
What benchmarks tell you
Run the same model on different nodes to compare real-world inference speed. Measures time-to-first-token (TTFT) and tokens/second.
Run Benchmark
Benchmark Results
| Node | Model | Compute | TTFT | Tok/sec | Total Time | Status | When | Actions |
|---|---|---|---|---|---|---|---|---|
Model Comparison
No comparison data available for this model.
๐ก๏ธ Remote Diagnostics โ
Nodes with Issues
No active alerts โ all nodes are healthy! ๐
Suggestion
Recent Diagnostic Logs
No diagnostic logs yet.
| Time | Node | Level | Message |
|---|---|---|---|
Send Message to Node
๐บ๏ธ Node Map
Node Locations
| Node | Location | ISP | GPU | Status |
|---|---|---|---|---|
| Online Offline |
Server Setup
One-Click Server Setup
Run this on any Linux CPU server. It installs LiteLLM, fetches all online XPU nodes from rxpu.nixosoft.com, and starts a unified OpenAI-compatible API endpoint on port 4000.
curl -fsSL https://rxpu.nixosoft.com/setup.sh | bash
What the script does
Registry Status
Firewall Setup
Open port 15029 to allow XPU contributors to connect:
sudo ufw allow 15029/tcp
sudo ufw reload
sudo firewall-cmd --permanent --add-port=15029/tcp
sudo firewall-cmd --reload
sudo iptables -A INPUT -p tcp --dport 15029 -j ACCEPT
Share with XPU Contributors
Send this to people who want to contribute their XPU:
๐ฅ๏ธ RemoteXPU Registry
Registry URL: https://rxpu.nixosoft.com
Download the app: https://nixosoft.com/downloads
1. Download and install RemoteXPU
2. Launch it
3. Select "Join an existing cluster"
4. Paste the registry URL above
5. Sign in with Google
shared-gpu-registry:
build: ./registry-service
container_name: shared-gpu-registry
restart: unless-stopped
ports:
- "15129:3456"
env_file:
- ./registry-service/.env
environment:
- PORT=3456
Environment Variables
Use Your Cluster
API Endpoint
API Keys
No enabled models.
Manage keys and models on the page.
Code Snippets
How Load Balancing Works
Your cluster automatically distributes requests across all online GPU nodes.
โข Multiple requests run in parallel โ one per GPU
โข LiteLLM routes to the least-busy node automatically
โข If a node goes offline, requests route to remaining nodes
โข Add more GPU nodes (same Google account) to increase capacity
โข Add more CPU servers to increase API throughput
Add a CPU Server (LiteLLM Router)
A CPU server acts as the API gateway โ it receives requests and routes them to your GPU nodes. You can have multiple CPU servers for redundancy.
Linux / macOS server:
curl -fsSL https://rxpu.nixosoft.com/setup.sh | bash
What it does:
โข Installs LiteLLM router
โข Connects to rxpu.nixosoft.com to discover your GPU nodes
โข Starts an OpenAI-compatible API on port 4000
โข Auto-updates its node list every restart
Multiple CPU servers:
Run the same command on any Linux server. Each becomes an independent API endpoint.
Useful for: different regions, redundancy, high availability.
Direct API test:
GPU Node Utilization
This shows which GPUs are available to serve your requests. Refreshes every 10s.
No online GPU nodes with VPN IPs found.
Settings
Current configuration loaded from environment variables. To modify, update the .env file and restart the service.
How to Update
1. SSH into the server hosting the Docker container
2. Edit the .env file in the registry-service directory
3. Restart the container:
docker compose up -d shared-gpu-registry