RunOS is a platform that lets you run your own cloud on any hardware. Today I'm sharing a new feature called Vibe Deploy that connects AI coding tools (currently Claude Code) to your infrastructure through an MCP server.
The flow looks like this: you describe your app in conversation, the AI writes the code, constructs a minimal runos.yaml file, and RunOS handles everything else. Server provisioning, Kubernetes setup, database creation, building, deploying. The runos.yaml for a typical app with a Postgres database is about 10 lines:
-----
app: news-api
port: 3000
requires:
news-api-db:
type: postgresql
class: postgresql.c0.tiny
config:
databaseName: newsapi
databaseUsername: newsapi
env:
url: DATABASE_URL
-----The MCP server uses lazy docs loading rather than stuffing the full RunOS documentation into the LLM context window upfront. The base instructions just tell the AI how to pull additional information on demand via REST API. It fetches only the docs it needs based on the conversation. Keeps context lean and responses accurate.
An accidental nice find was the debugging loop. When a deploy fails, the AI automatically checks build logs, reads app logs, finds the error, fixes the code and redeploys. I've watched it catch a TypeScript type error in a database pool config, fix it, redeploy, then catch a bad migration, fix that too, and get the app running. All without intervention. After deploy, the MCP connection stays useful. You can query your database in natural language, check object storage, read logs. No separate database client or CLI needed.
The cost angle matters too. A 10 node cluster on AWS runs roughly $2,500/month. Same thing on DigitalOcean is under $1,000. On Hetzner through RunOS it's under $250.
Where we're at honestly: it's late beta. Four services are available for AI provisioning right now (PostgreSQL, MySQL, Valkey, MinIO), more coming. There are probably bugs. We're heading to production in the next couple of months. The scope ranges from single node setups for prototypes to multi-cluster, multi-node environments. The goal is to be a serious enterprise offering, not just a toy for side projects.
One use case I'm particularly interested in: teams where developers want to spin up prototypes quickly without bothering DevOps every time. Provision a server in your own network, deploy through AI, iterate fast. No tickets, no waiting.
2 min demo at runos.com. Happy to answer any questions about the architecture or approach.