Hacker Newsnew | past | comments | ask | show | jobs | submit | vampiregrey's commentslogin

I've been using Claude Code as my primary agent orchestration tool for a few weeks now. Running cron-style loops for browser automation through a remote Playwright instance. Haven't tried Codex at this depth yet, but from what I've seen it's more focused on code generation while Claude Code is more of a general-purpose agent runtime. The things that make Claude Code powerful for me are the persistent session with tool access, the ability to run background tasks on schedules, and the hook/skill system for customization. If Codex matches that plus adds better async execution (Claude Code's scheduled tasks die when the session ends), it could be compelling. But for now Claude Code handles my use case better than anything else I've tried.


I mostly use claude code for coding purposes. According to my experience, claude code is unbeatable in terms of coding and UI. But I feel like chat gpt is better for architectural decisions. But I haven't tried codex, BUT I am listening a lot now that its limits are far better than claude code and that it is good in terms of value for money.


Interesting approach. I've been running a similar setup using Slack as the control plane for browser automation agents. The main challenge with messaging apps as control planes is handling long-running tasks and state management. Telegram's bot API is actually nicer than Slack's for this since you get more granular message editing.


agreed. the long-running tasks i handle with immediate-background and continually edit message with updates, along with fork-and-reconcile semantics so you don't have to queue messages. also having a supergroup with separate topics for different sessions works really well.

the separate topics approach along with allowing other processes to enqueue messages into the session stream solves state pretty tidily, see "the session as decision-maker" in the linked report


Location: India

Remote: Yes

Willing to relocate: Yes (Worldwide)

Technologies: Node.js, NestJS, Python, Go, Agentic AI, LLM Integration, Vector Embeddings, React, Next.js, TypeScript, AWS, Docker, PostgreSQL, GitLab CI/CD

Résumé/CV: https://zerobitflip.com

Email: shubham.maurya.co@outlook.com

Backend Engineer at British Council, architecting Agentic AI systems for one of the world's largest cultural organizations. Building AI agent orchestration layers and LLM-powered backend services that automate complex examination workflows across 100+ countries. Previously built a FinTech SaaS from zero at Marquee Equity and solo-shipped an AI-powered interview prep platform (ProTechStack.com)


Thank you !


Well deserved! this is a huge win for TS backend stack


I am still waiting for Mac Mini with M5


and MacStudio!


What do you use the Mac Studio for?

I’ve always felt they weren’t really worth it for performance per dollar spent. For C++ work I just use a non-Mac workstation. For lighter workloads the Mac Mini is very capable already.


The Studio (Stud IO™) is the new Mac Pro - it's not "worth it" unless you need the most performance period - or you have money to spare.

Or you really, really need to drive eight displays from a single machine.

For "home user" stuff a Mac mini or MacBook is going to do everything you ever need (in fact, they have the problem where the M1 systems are still perfectly capable, six years later).


studio with m5 ultra this week might have me pulling the trigger.


You think they will skip M4 ultra? May be they slowly plan to launch ultra chips alternate years since the development costs are high and demand is niche.

If they do a 1TB m5 ultra, I too would be configuring one for sure.


I hope so! I already have the M4 mini pro - would like to bump up the prompt processing time and the memory bandwidth at the same time with the new M5 matmul changes.


Location: India (Remote)

Remote: Yes

Willing to relocate: Yes

Technologies: Go, Node.js/NestJS, Python, AWS, Docker, PostgreSQL, Playwright, React/Next.js

Résumé/CV: https://zerobitflip.com

Email: sm@zerobitflip.com

Senior Backend Engineer with 5+ years building scalable systems across FinTech, EdTech, and AI. Currently at British Council architecting exam platforms for millions of concurrent users. Solo-built a Go-based scraping engine that extracts structured data from 200M+ company records, and an AI interview prep platform with 100K+ users (ProTechStack.com). No over-engineered stacks — Go, PostgreSQL, Playwright, and headless Chrome. Looking for early-stage roles.


Hey HN,

I’ve been building Solidclaw, a small local credential broker + model proxy for OpenClaw.

I started it because I was increasingly uncomfortable with:

API keys living in random .env files

Provider keys sitting inside OpenClaw configs

Tool secrets bundled directly into plugins

Credentials duplicated across services

It felt messy and fragile — especially when experimenting with multiple providers and tools.

So Solidclaw sits in front and handles:

Keeping provider keys out of OpenClaw

Isolating tool secrets

Scoping model access via tokens

Injecting secrets only at runtime

The idea is simple: clean separation between orchestration and secrets. Less key sprawl. Smaller blast radius. More explicit access control.

It’s very early stage and evolving fast. I’m iterating based on real usage while building AI infra tools.

If you're working with OpenClaw — or generally building local-first AI infra — I’d genuinely appreciate feedback, criticism, or ideas.

Happy to share architecture details if there’s interest.


I’m not sure I buy this framing.

I agree that reading every dependency isn’t realistic. But “not reading the code” as a principle feels risky.

In my experience, abstractions hold until they don’t. The first time you hit a production incident and the docs stop helping, reading the source stops being academic and starts being survival.

We once had a performance issue caused by a library making assumptions about concurrency that weren’t obvious from the API. The fix only became clear after stepping through the source.

I think the real skill isn’t avoiding reading code, it’s knowing when to escalate from trust to understanding.

For glue code or low stakes utilities, sure. For auth, billing, or core infra, I’d argue reading at least the critical paths pays dividends.


So last week I tried Gemini pro 3, Opus 4.6, GLM 5, Kimi2.5 so far using Kimi2.5 yeilded the best results (in terms of cost/performance) for me in a mid size Go project. Curious to know what others think ?


I predict Gemini Flash will dominate when you try it.

If you're going for cost performance balance choosing Gemini Pro is bewildering. Gemini Flash _outperforms_ Pro in some coding benchmarks and is the clear parento frontier leader for intelligence/cost. It's even cheaper than Kimi 2.5.

https://artificialanalysis.ai/?media-leaderboards=text-to-im...


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: