ModelReins – The Browser for AI Tools
Article URL: https://modelreins.com Comments URL: https://news.ycombinator.com/item?id=47635601 Points: 3 # Comments: 1
Stop juggling AI tools.Start orchestrating them.
Route tasks to the right model at the right cost. Claude, Codex, Ollama, or any AI tool. Direct API keys, no subscriptions required, no lock-in.
One dashboard, any provider.
+N
Scale on your terms. No artificial limits. Self-hosted: unlimited workers on your own hardware. SaaS: plans that grow with you. Every machine you own is a potential worker node.
Worker ↔ AI Providers
dashboard never sees your keys
Zero-knowledge. Your keys stay yours. Workers talk directly to providers using your API keys — not subscriptions that can be revoked. Prompts, code, responses never pass through our servers. We orchestrate. We don't eavesdrop. And we don't lock you in.
Before
18%
After
94%
No vendor lock-in. Ever. Vendors change terms overnight. Subscriptions get restricted. APIs get throttled. ModelReins workers use direct API keys — no subscriptions required, no vendor lock-in, no surprises.
Opus
rate limited
↓ auto-failover
Sonnet
routed
Rate limited? Work never stops. Hit a limit on Opus — the router instantly fails over to Sonnet. Sonnet full? Ollama picks it up locally for free. Zero downtime. Zero babysitting.
Live Signal Feed
What it looks like
One command. Your AI workforce is online.
worker — haiku-carbug
$ npx modelreins-worker
| / | ___ __| | ___| | _ \ _() __ ___ | |/| |/ _ \ / _
|/ _ \ | |_) / _ \ | '_ \/ __| | | | | (_) | (_| | __/ | _ < __/ | | | \__ \ |_| |_|\___/ \__,_|\___|_|_| \_\___|_|_| |_|___/ by MEDiAGATO__ |/ _ \ | |_) / _ \ | '_ \/ __| | | | | (_) | (_| | __/ | _ < __/ | | | \__ \ |_| |_|\___/ \__,_|\___|_|_| \_\___|_|_| |_|___/ by MEDiAGATO__Worker: haiku-carbug Provider: anthropic (haiku-4.5) Server: app.modelreins.com Tags: draft,triage,cheap,fast Session: spawn
[20:24:01] Ready — waiting for jobs... [20:24:17] >>> Job #803 claimed [20:24:17] Prompt: Write a product description for ModelReins... [20:24:17] Spawning: anthropic-cli "Write a product description..." [20:24:22] <<< Job #803 complete (exit 0, 4.8s) [20:24:27] >>> Job #804 claimed [20:24:27] Prompt: Triage this issue: auth middleware returns 403... [20:24:29] <<< Job #804 complete (exit 0, 1.2s) [20:24:34] Ready — waiting for jobs...|`
Google calls it:
"the shift from generative to agentic AI."
We just call it Tuesday.
ModelReins has been orchestrating multi-provider AI workforces while the industry was still writing trend reports about it.
Google Cloud AI Agent Trends 2026
Three steps
Run the server
Python + SQLite. Set a token, run python app.py. That's the control plane.
Connect workers
Install the SDK on any machine. npx modelreins-worker. It phones home and waits for work.
Dispatch jobs
Type a task in the dashboard. Pick a worker or let it auto-route. Output streams back in real time.
Capabilities
Multi-Agent Dispatch
Route tasks to any registered worker. Manual assignment or automatic.
Fleet Awareness
Define your infrastructure in YAML. Workers know what exists and what's healthy.
Context Policies
Control what each worker sees. Frontend tasks get URLs. Infra tasks get the map.
Secrets Brokering
Pointers, not passwords. Env vars or Vault. Workers get short-lived tokens.
Multi-Tenant RBAC
Complete data isolation. Admin, operator, viewer. Teams share one server safely.
Killswitch
File, URL, or dead man's switch. Halt all workers instantly. Independent of the server.
Signed Audit Trail
Every action HMAC-signed and logged. Verify integrity. Ship to your SIEM.
Zero-Knowledge Keys
Your API keys never touch the control plane. Workers fetch credentials locally.
Pricing
Free
$0/mo
-
2 workers
-
Unlimited jobs
-
Killswitch
-
Dashboard + streaming
-
Job chaining
-
Cost tracking
Start Free
Pro
$29/mo
-
10 workers
-
Unlimited jobs
-
Killswitch
-
Chain templates
-
Fleet context injection
-
Approval gates
-
Analytics dashboard
-
Priority support
Upgrade
Team
$79/mo
-
50 workers
-
Unlimited jobs
-
Killswitch
-
Chain templates
-
API access
-
Team members
-
Multi-user RBAC
-
Everything in Pro
Upgrade
Self-Hosted
Free
-
Unlimited everything
-
Your infrastructure
-
Full source (BSL 1.1)
-
Commercial license available
Source
Your agents. Your infrastructure. Your rules.
Start free. Upgrade when you need more workers.
Enterprise
Self-Hosted. Your Keys. Your Rules.
Run ModelReins on your own infrastructure. Bring your own API keys, keep data in-house, and manage AI workloads across teams with the controls you actually need.
-
Self-hosted deployment
-
Multi-tenant with role-based access
-
Fleet-aware worker routing
-
Approval gates and budget controls
-
SSO integration
-
Priority support and SLA
Let's Talk
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
model
The Pre-Flight Checklist: 7 Things I Verify Before Sending Any Prompt to Production
You wouldn't deploy code without running tests. So why are you sending prompts to production without checking them first? After shipping dozens of AI-powered features, I've settled on a 7-item pre-flight checklist that catches most problems before they reach users. Here it is. 1. Input Boundaries Does the prompt handle edge cases in the input? Empty strings Extremely long inputs (token overflow) Unexpected formats (JSON when expecting plain text) Quick test: Feed it the worst input you can imagine. If it degrades gracefully, you're good. 2. Output Format Lock Is the expected output format explicitly stated in the prompt? Bad: "Summarize this article." Good: "Summarize this article in exactly 3 bullet points, each under 20 words." Without format constraints, you get different shapes every r

Measuring AI's Role in Software Development: Evaluating Agency and Productivity in Low-Level Programming Tasks
The Role of AI in Low-Level Software Development: An Expert Analysis As a low-level programmer, I’ve witnessed the growing integration of AI tools like GitHub Copilot into software development workflows. The industry hype often portrays these tools as revolutionary, capable of transforming coding into a near-autonomous process. However, my firsthand experience reveals a more nuanced reality: AI serves as an accelerator and assistant, but its agency in handling complex, low-level tasks remains severely limited. This analysis dissects the mechanisms, constraints, and system instabilities of AI in this domain, contrasting practical contributions with exaggerated claims. Mechanisms of AI Integration in Low-Level Development 1. AI-Assisted Code Completion Impact → Internal Process → Observable

I built an npm middleware that scores your LLM prompts before they hit your agent workflow
The problem with most LLM agent workflows is that nobody is checking the quality of the prompts going in. Garbage in, garbage out but at scale, with agents firing hundreds of prompts per day, the garbage compounds fast. I built x402-pqs to fix this. It's an Express middleware that intercepts prompts before they hit any LLM endpoint, scores them for quality, and adds the score to the request headers. Install npm install x402-pqs Usage const express = require ( " express " ); const { pqsMiddleware } = require ( " x402-pqs " ); const app = express (); app . use ( express . json ()); app . use ( pqsMiddleware ({ threshold : 10 , // warn if prompt scores below 10/40 vertical : " crypto " , // scoring context onLowScore : " warn " , // warn | block | ignore })); app . post ( " /api/chat " , ( re
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.





Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!