Readit News logoReadit News
evara-ai commented on Elevated Errors in Claude.ai   status.claude.com/inciden... · Posted by u/LostMyLogin
evara-ai · 10 days ago
This is a real operational problem when you're building client-facing automation systems on top of these APIs. I build chatbots, workflow automation, and AI agent systems for clients — and the hardest conversation is explaining that your system's uptime is fundamentally capped by your LLM provider's uptime.

Patterns that have helped in production:

1. Multi-provider fallback. For conversational systems, route to Claude by default, fall back to GPT-4 on 5xx errors. The response quality difference is usually acceptable for the 2-3% of requests that hit the fallback. This turns a hard outage into a slight quality degradation.

2. Async queuing for non-real-time workflows. If you're processing documents, generating reports, or running batch analysis — don't call the API synchronously. Queue the work, retry with exponential backoff, and let the system self-heal when the API recovers. Most of our automation pipelines run with a 15-minute SLA, not a 500ms one.

3. Graceful degradation in real-time systems. For chatbots and voice agents, have a scripted fallback path. "I'm having trouble processing that right now — let me transfer you to a human" is infinitely better than a hung connection or error message.

The broader issue: we're all building on infrastructure where "four nines" isn't even on the roadmap yet. That's fine if you architect for it — treat LLM APIs like any other unreliable external dependency, not like a database query.

evara-ai commented on Parallel coding agents with tmux and Markdown specs   schipper.ai/posts/paralle... · Posted by u/schipperai
evara-ai · 10 days ago
The comment about having "3 to 6 hours per day" to work directly with code is the key insight here. I run a small AI consultancy and use Claude Code daily to deliver client projects — chatbots, automation pipelines, API integrations — and the spec-driven approach described in this post is what makes it actually work at scale.

The pattern I've converged on: spend the first 30 minutes writing detailed markdown specs (inputs, outputs, edge cases, integration points), then let Claude Code chew through the implementation while I review, test, and iterate. For a typical automation project — say a WhatsApp bot that handles booking flows and integrates with a client's CRM — this cuts delivery time roughly in half compared to writing everything manually.

The biggest practical lesson: the spec quality is everything. A vague spec produces code you'll spend more time debugging than you saved. A good spec with explicit error handling expectations, API response formats, and state transitions produces code that's 80-90% production-ready on the first pass.

Where I disagree slightly with the parallel agent approach: for client-facing work where correctness matters more than speed, I've found 2-3 focused agents (one on backend, one on frontend, one on tests) more reliable than 6-8 competing agents that create merge conflicts. The overhead of resolving conflicts and ensuring consistency across parallel outputs eats into the productivity gains fast.

evara-ai commented on A case for Go as the best language for AI agents   getbruin.com/blog/go-is-t... · Posted by u/karakanb
evara-ai · 10 days ago
Having built a bunch of production AI agent systems (workflow automation, conversational agents, data pipelines with LLM-in-the-loop) in Python and Node.js — my take is that language choice is genuinely the least important decision you'll make.

The bottleneck in agent systems is almost never your language runtime. It's LLM API latency (200-2000ms per call), external service I/O, and retry/error handling across unreliable tool calls. Whether your orchestration loop runs in 2ms (Go) or 15ms (Python) is irrelevant when you're waiting 800ms for Claude to respond.

What actually matters for production agent systems: (1) state management across multi-step workflows that can fail at any point, (2) graceful degradation when one tool in a chain times out, (3) observability into what the agent decided and why. These are design problems, not language problems.

Python wins on ecosystem breadth — every LLM provider ships a Python SDK first, every embedding model has Python bindings, and the tooling around prompt engineering and evaluation is Python-native. When you're iterating on agent behavior (which is 80% of the work), that ecosystem advantage compounds fast.

That said, Go's argument is strongest for the "agent runtime" layer — the part that manages concurrency, schedules tool calls, and handles streaming. If you separate the orchestration runtime from the AI logic, Go for the former and Python for the latter isn't a bad split.

evara-ai commented on Show HN: I built a sub-500ms latency voice agent from scratch   ntik.me/posts/voice-agent... · Posted by u/nicktikhonov
evara-ai · 10 days ago
Great writeup. I've been building production voice agents and automation systems with the same stack (Twilio + Deepgram + ElevenLabs + LLM APIs) for client-facing use cases — appointment booking, lead qualification, guest concierge, and workflow orchestration.

The "turn-taking problem, not transcription problem" framing is exactly right. We burned weeks early on optimizing STT accuracy when the actual UX killer was the agent jumping in mid-sentence or waiting too long. Switching from fixed silence thresholds to semantic end-of-turn detection was night and day.

One dimension I'd add: geography matters even more when your callers are in a different region than your infrastructure. We serve callers in India connecting to US-East, and the Twilio edge hop alone adds 150-250ms depending on the carrier. Region-specific deployments with caller-based routing helped a lot.

The barge-in teardown is the part most people underestimate. It's not just canceling LLM + TTS — if you have downstream automation (updating booking state, triggering webhook workflows, writing to DB), you need to handle the race condition where the system already committed to a response path that's now invalid. We had a bug where a barged-in appointment confirmation was still triggering the downstream booking pipeline.

evara-ai commented on Ask HN: Who wants to be hired? (March 2026)    · Posted by u/whoishiring
evara-ai · 10 days ago
Location: India Remote: Yes (worldwide, flexible timezones) Willing to relocate: No Technologies: Python, Node.js, TypeScript, Twilio, ElevenLabs, Deepgram, OpenAI API, Anthropic API, PostgreSQL, Supabase, Redis, Docker, AWS Resume/CV: https://www.linkedin.com/in/priyanshu-axiom Email: priyanshukumarmaurya2224@gmail.com

Founder of Evara AI (incubated at IIT Bhubaneswar). I specialize in Voice AI agents (inbound/outbound call automation, lead qualification, appointment scheduling) and custom automation systems. Code-first approach -- Python/Node.js pipelines, no platform lock-in.

Recent work includes an AI concierge system for hospitality and automation delivery for a digital marketing agency. Looking for contract or freelance engagements, available ~20 hrs/week.

Dead Comment

u/evara-ai

KarmaCake day5March 3, 2026
About
Founder, Evara AI (incubated at IIT Bhubaneswar). I build Voice AI agents, WhatsApp chatbots, and custom automation systems. Python/Node.js, code-first approach.

Available for contract/freelance work. Email: priyanshukumarmaurya2224@gmail.com LinkedIn: www.linkedin.com/in/priyanshu-axiom

View Original