Dockhive · Cloud Infrastructure · 2025

Operators were accountable for infrastructure they couldn't see.

Dockhive is a unified cloud infrastructure platform for app hosting, databases, and workflow automation. As sole designer working with Dockhive's engineering team, I designed two layers: the deployment experience that gets developers to production in under 6 minutes, and the operator visibility layer that keeps them in control once it's live.

Dockhive · Platform Status

84 nodes · 312 workloads
Latest Deploy
my-next-app
Production · 4m ago · Live
Network Health
76 healthy
6 degraded
2 critical
Deploys today
124
Success rate
97.8%
Avg deploy time
5:42
Active regions
6

The Problem

A developer shipping a Next.js app with Postgres and background jobs had to configure Vercel, spin up Supabase, and wire n8n. Three dashboards, three billing pages, three mental models. Average time to first deployment: 45 minutes. Most gave up before finishing. Dockhive unified all three into one platform. The design problem: make deploying infrastructure feel like something a developer wants to do, not something they have to survive.

Deployment Flow

One entry point. Three services: app hosting, database, automation. Framework selection is a card grid, auto-detected from the repo. Three steps to deploy anything.

New Project · Service Selection
Step 1 of 3
App Hosting
Deploy React, Next.js, Vue and more. Git, Docker, or public URL.
Deploy →
🗄️
Database
PostgreSQL, MySQL, MongoDB, Redis and more. Managed backups.
Deploy →
🔄
N8N Automation
Visual workflow automation. Web3-focused templates included.
Deploy →

↑ Interactive — 89% of new users correctly identified which service they needed on first try, up from 34% with the previous dropdown.

The Operate Problem

Getting deployed is one problem. Understanding what's running after is the harder one. DePIN compute operators run nodes across multiple regions, earning rewards from AI workloads running autonomously 24/7. When something goes wrong, they find out when rewards stop.

Node Explorer

Before this, operators had a spreadsheet of node IDs. No geography, no topology. Degraded nodes in the same region is a power issue, not a hardware failure. That's a two-minute ISP call, not a truck roll.

Node Explorer · Global · 247 nodes
3 critical
Healthy (213)
Degraded (22)
Critical (3)
Offline (9)
Total nodes
247
Network uptime
98.4%
Active workloads
1,842
Rewards / hr
4.7 ETH

↑ Interactive — 28 geographic positions. Health, uptime, and earnings visible at a glance.

Failure & Response

The old interface showed one status: Offline. What caused it was left to the operator to reconstruct. The failure signature surfaces likely cause alongside state: thermal cascade, network drop, hardware stress. An operator responding to an alert at 3am needs the interface to manage cognitive load, not just display information.

Node Detail · NODE-047 · Lagos, NG
Critical · Intervention required
NODE-047
Lagos, NG · Rack B · Operator: Chukwuemeka A.
Critical
CPU
97%
Memory
94%
Uptime
94.1%
Temp
89°C
Network
42 Mbps
Rewards today
0.024 ETH
Failure signature — Thermal throttling cascade
Hardware overheating under sustained compute load
CPU and memory utilization elevated simultaneously with rising core temperature. Workloads are executing but at reduced throughput. If temperature exceeds 95°C, automatic shutdown triggers. Likely cause: cooling system underperforming or ambient temperature elevated.
Event timeline
Now
Thermal critical threshold — 89°C
14m ago
CPU sustained above 90% for 10+ min
31m ago
Memory utilization spike — 87%
1h ago
Large workload assigned — AI inference batch
Yesterday
All systems nominal · 99.2% uptime
01
Alert fires
02
Investigate
03
Intervene
04
Resolve
Alert fires — single critical fact
NODE-047 · Lagos — Thermal critical. CPU 97%, memory 94%, core temperature 89°C. Workload at risk of interruption.

↑ Interactive — Failure signature tells you what happened and why. Four-stage flow leads the operator through the incident.

Economic Layer

Operators aren't running infrastructure for fun. They're earning ETH from AI workloads. The economic layer surfaces earnings, performance, and the direct relationship between uptime and revenue. Degraded nodes cost real money. Now that cost is visible.

Economic Layer · Rewards & Performance
This month · 14.8 ETH earned
Total earned · all nodes
14.8 ETH
+12% vs last month
Month progress68%
Lost earnings · degraded nodes
0.84 ETH
Recoverable with intervention
Impact5%
Projected · month end
21.7 ETH
At current performance
vs target100%
Node
Earned
Performance
Status
NODE-012 · Frankfurt
0.48 ETH
↑ 99.9%
Healthy
NODE-031 · Seoul
0.44 ETH
↑ 99.1%
Healthy
NODE-047 · Lagos
0.011 ETH
↓ 41%
Critical
NODE-089 · Amsterdam
0.28 ETH
→ 71%
Degraded

↑ Interactive — Revenue per node, lost earnings from degraded nodes, projected payouts.

Design Decisions

Lagos-first design

If the interface works on a phone with 2G connectivity and one thumb, it works everywhere. This constraint shaped everything: tap targets, information density, progressive disclosure.

Failure signatures over status codes

"Offline" is not actionable. "Thermal throttling cascade, likely cooling system failure" is. The interface tells you what to do, not just what's wrong.

Four stages, not four tabs

Incident response is a flow, not a dashboard. Each stage surfaces exactly what the operator needs at that moment. No decisions about where to look.

Outcomes

89%

Service identification on first try, up from 34% with the previous dropdown.

↓40%

Reduction in operator churn.

84

Network nodes across 6 regions at launch.

<6min

Time to deploy, down from 45min.

Reflection

Early versions of the operator dashboard showed everything. Every metric, every log, every node. It looked comprehensive. In testing with Lagos-based operators on mobile connections, it was unusable. Pages took 12 seconds to load. Tap targets were too small for one-thumb use. The information hierarchy assumed a wide monitor and fast internet. Stripping it back to only what an operator needs at each moment (failure cause, affected nodes, recommended action) made the interface faster, clearer, and more decisive. The Lagos constraint wasn't a concession. It was the design principle that made the product work for everyone, including the San Francisco operators who never knew they needed it.