I didn't set out to build an AI builder. I set out to stop hating side-project setup.
Every time I had a new idea - a job board, a habit tracker, a small CRM for a friend's agency - I'd lose the first two evenings to the same ritual: npm init, auth boilerplate, schema design, Tailwind config, deploy pipeline. By the time I got to the actual idea, the spark was gone.
So I started building a tool to skip the ritual. That tool became Xandhi OS.
The 90-second pitch
You type:
"A SaaS dashboard with team workspaces, billing integration, and a public marketing page."
Xandhi OS routes that prompt through nine layers and hands you back a complete, downloadable application - frontend, styling, interactivity, and structure - in minutes. Real code. Not a sandbox. Not a snippet.
Why most AI builders frustrated me
Three things made me allergic to the existing options:
They were either too shallow or too magical. Some give beautiful components but no backend structure. Others give working apps but lock you to expensive paid models with metered tokens.
They didn't think about cost discipline. If I'm going to prototype 20 ideas before finding the one, I cannot pay $2 per prototype.
The output felt like a black box. I want the actual code, in my hands, deployable anywhere.
I wanted something different: deep, transparent, and cheap to experiment with.
The 9-layer architecture
The core insight was that "AI builder" isn't one job - it's a pipeline. Each stage has different requirements (latency, reasoning depth, structured output, creativity), which means each stage benefits from a different approach.
Here's the pipeline I ended up with:
1. Intent Parser - what does the user actually want?
2. Spec Generator - turn the intent into a structured spec
3. Architecture Planner - choose stack, modules, data model
4. Component Composer - UI layout, page tree, design tokens
5. Code Generator - write the actual files
6. Linter - check for syntax and style issues
7. Auto-Debugger - fix errors automatically
8. Security Scanner - check for vulnerabilities
9. Packager - bundle everything for download
Each layer has a defined input/output contract. The orchestrator is a state machine walking the prompt through them.
Multi-model routing: the unsung trick
Here's what makes Xandhi OS efficient to run.
Through OpenRouter, the system routes through 13 AI models from 6 providers. Instead of being locked to one expensive model, it picks the best model for each task:
- Planning stages use models good at reasoning and structured output
- Code generation uses models specialized in code quality
- Debugging uses models with strong error analysis capabilities
- Simple tasks use lightweight, fast models
A simplified routing concept:
ROUTING = {
"intent": ["qwen/qwen-2.5-72b:free", "meta-llama/llama-3.3-70b:free"],
"spec": ["deepseek/deepseek-chat:free"],
"code": ["deepseek/deepseek-coder:free", "meta-llama/llama-3.3-70b:free"],
"debug": ["meta-llama/llama-3.3-70b:free"],
}
async def route(task, prompt):
for model in ROUTING[task]:
try:
return await call_openrouter(model, prompt)
except:
continue # try next model in fallback chain
raise Exception(f"All models failed for: {task}")
That's the core concept. Add retries, circuit breakers, and observability around it, and you have a production routing layer.
The 7-step auto-debug pipeline
This was the feature that changed everything. Instead of handing users broken code and saying "fix it yourself," Xandhi OS runs every generated file through:
- Lint - check syntax
- Error detection - find logical issues
- Auto-fix - feed errors back to AI for correction
- Re-lint - verify fixes worked
- Security scan - check for XSS, injection, etc.
- Per-file validation - ensure each file is complete
- Final verification - confirm the build is clean
This eliminates roughly 60% of "broken build" complaints. Just feeding the error back into the model with one more turn fixes most generation bugs.
What surprised me
Free models are stunningly good in 2025. They match paid model output quality around 85-90% of the time on structured code tasks.
The bottleneck isn't the model - it's the prompt scaffolding. Same model, better system prompts, 2x output quality.
Self-healing eliminates most complaints. Auto-feeding build errors back into the model is the single highest-ROI feature I built.
People want the code more than the deploy. I assumed deploy-first. Users overwhelmingly asked for "give me the downloadable files."
The tech stack
| Layer | Technology |
|---|---|
| Frontend | Next.js 15, React, TypeScript |
| Backend API | Go (Fiber framework) |
| AI Engine | Python (FastAPI), OpenRouter |
| Database | PostgreSQL 16 |
| Caching | Redis 7 |
| Proxy | Nginx (SSL, gzip, rate limiting) |
| Hosting | Hetzner Cloud (Germany) |
| Payments | Razorpay (India + International) |
Results so far
- 22+ page types and templates live
- 13 AI models across 6 providers
- 7-step auto-debug pipeline
- Average build time: 3-5 minutes
- Cost per build for free tier users: zero
- Active Discord community
What I learned (the unsexy version)
Pipelines beat monoliths. One big prompt to "build me an app" is a coin flip. Nine small prompts, each evaluated independently, is engineering.
Routing matters more than model choice. Picking which model when matters more than picking the best model for everything.
Make the output ownable. Users trust tools that hand them the code, not tools that hide it behind a paywall.
Ship in public. I underestimated how much momentum Discord plus daily build-logs creates.
Be honest about limitations. Users forgive imperfect code generation. They don't forgive fake metrics or misleading claims.
Try it
If any of this resonated, Xandhi OS is free to try:
- Website: xandhi.com
- Discord: discord.gg/uAxufdAnD
- Twitter: @xandhios
- GitHub: github.com/xandhiai/xandhi-os
Tell me what you'd build. I might ship a template for it this week.
-- Built with too much chai in New Delhi
Top comments (0)