Hacker Newsnew | past | comments | ask | show | jobs | submit | gck1's commentslogin

Opus 4.7 refuses to work on the scraper that opus 4.6 wrote. I can assure you that my scraper is configured to be as polite and nice to the target as possible. It definitely is way nicer than any ANT/OAI scraper our there.

Curiously, the reason why I started using Claude about a year ago was that OAI models were refusing to answer even the most benign questions, which nobody but someone with paranoia consider a dual intent.


And with Anthropic introducing KYC requirements, this is essentially a lifetime ban.

Fun times.


I have a fun little agent in my tmux agent orchestration system - Socratic agent that has no access to codebase, can't read any files, can only send/receive messages to/from the controlling agent and can only ask questions.

When I task my primary agent with anything, it has to launch the Socratic agent, give it an overview of what are we working on, what our goals are and what it plans to do.

This works better than any thinking tokens for me so far. It usually gets the model to write almost perfectly balanced plan that is neither over, nor under engineered.


Sounds pretty neat! Is there an written agent.md for that you could share for that?

I even have a specific, non-negotiable phase in the process where model MUST interview me, and create an interview file with everything captured. Plan file it produces must always include this file as an artifact and interview takes the highest precedence.

Otherwise, the intent gets lost somewhere in the chat transcript.


The raw Q&A is essential. I think Q & Q works so we'll because it reveals how the model is "thinking" about what you're working on, which allows for correction and guidance upfront.

Are these your own skills files or are you using something off the shelf like bmad or specify-kit?

This is interesting, can you link any more details on it?

Not GP, but BMAD has several interview techniques in its brainstorming skill. You can invoke it with /bmad-brainstorming, briefly explain the topic you want to explore, then when it asks you to if you want to select a technique, pick something like "question storming". I've had positive experience with this (with Opus 4.7).

While we're giving away bags, I'd like HDD manufacturers to get some too.

> They explicitly explain your workflow needs to change

How about - don't break my workflow unless the change is meaningful?

While we're at it, either make y in x.y mean "groundbreaking", or "essentially same, but slightly better under some conditions". The former justifies workflow adjustments, the latter doesn't.


Cache ttl on max subscriptions is 1h, FYI.

Only if you set `ENABLE_PROMPT_CACHING_1H`, which was mentioned in the release notes for a recent Claude Code release but doesn't seem to be in the official docs.

subusers supposedly get it automatic again after the fix (and now also with `DISABLE_TELEMETRY=1`)

but if you are api user you must set `ENABLE_PROMPT_CACHING_1H` as i understood

and when using your own api (via `ANTHROPIC_BASE_URL`) ensure `CLAUDE_CODE_ATTRIBUTION_HEADER=0` is set as well... https://github.com/anthropics/claude-code/issues/50085

and check out the other neckbreakers ive found pukes lots of malicious compliance by feels... :/

[BUG] new sessions will *never* hit a (full)cache #47098 https://github.com/anthropics/claude-code/issues/47098

[BUG] /clear bleeds into the next session (what also breaks cache) #47756 https://github.com/anthropics/claude-code/issues/47756

[BUG] uncachable system prompt caused by includeGitInstructions / CLAUDE_CODE_DISABLE_GIT_INSTRUCTIONS -> git status https://github.com/anthropics/claude-code/issues/47107


Bruh. It's getting hard to track down all these MAKE_IT_ACTUALLY_WORK settings that default to off for no reason.

For me it's gotten to the point where I have a wrapper script that applies like 5 environment variables and even patches the system prompt strings prior to every Claude Code invocation.

After the Claude Code source code leak someone discovered that some variables are read directly from the process environment. Can't even trust that setting them in ~/.claude/settings.json will work!

I've actually started asking Claude itself to dissect every Claude Code update in order figure out if it broke some part of the Rube Goldberg machine I was forced to set up.


That's the beginning of Googlification of feature evolution, via population statistics rather than quality.

If it increases a KPI by 5% for 95% of users but torpedos the experience for 5%? Ship it.


no way, I didn't realise this worked.

My attention span is such that I get side tracked and wind up taking longer than 5 mins quite a bit :D


That'd be awesome but it doesn't reflect what I see. Do you have a source for that? What I see is if take a quick break the session loses ~5% right at the start of the next prompt processing. (I'm currently on max 5x)

Not at my workstation right now, but simply ask claude to analyze jsonl transcript of any session, there are two cache keys there, one is 5m, another 1h. Only 1h gets set. There are also some entries there that will tell you if request was a cache hit or miss, or if cache rewrite happened. I've had claude test another claude and on max 5x subscription, cache miss only happened if message was sent after 1h, or if session was resumed using /resume or --resume (this is a bug that exists since January - all session resumes will cause a full cache rewrite).

However, cache being hit doesn't necessarily mean Anthropic won't just subtract usage from you as if it wasn't hit. It's Anthropic we're talking about. They can do whatever they want with your usage and then blame you for it.


I have heard that if you have telemetry disabled the cache is 5 minutes, otherwise 1h. No clue how true that is however my experience (with telemetry enabled) has been the 1h cache.

They've acknowledged that as a bug and have fixed it.

It's true as far as I can tell, just by my own checking using `/status`. You can also tell by when the "clear" reminder hint shows up. Also if you look at the leaked claude code you can see that almost everything in the main thread is cached with 1H TTL (I believe subagents use 5 minute TTL)

Anthropic is playing a strange game. It's almost like they want you to cancel the subscription if you're an active user and only subscribe if you only use it once per month to ask what the weather in Berlin is.

First they introduce a policy to ban third party clients, but the way it's written, it affects claude -p too, and 3 months later, it's still confusing with no clarification.

Then they hide model's thinking, introduce a new flag which will still show summaries of thinking, which they break again in the next release, with a new flag.

Then they silently cut the usage limits to the point where the exact same usage that you're used to consumes 40% of your weekly quota in 5 hours, but not only they stay silent for entire 2 weeks - they actively gaslight users saying they didn't change anything, only to announce later that they did, indeed change the limits.

Then they serve a lobotomized model for an entire week before they drop 4.7, again, gaslighting users that they didn't do that.

And then this.

Anthropic has lost all credibility at this point and I will not be renewing my subscription. If they can't provide services under a price point, just increase the price or don't provide them.

EDIT: forgot "adaptive thinking", so add that too. Which essentially means "we decide when we can allocate resources for thinking tokens based on our capacity, or in other words - never".


OpenAI and Anthropic have realized that their entire business is exactly one open weight model drop away from Chinese labs that matches Opus 4.5 performance.

They realized they have no product or ground to stand on. Once such model drops and once chip manufacturers catch up with demand, they are dead, if their only product is inference.

So OpenAI decided to do weird things like buying up all the hardware that exists or will exist in the next 2 years to buy time to build the product. Then launch things like Sora, ChatGPT shopping, ads etc. They seem to be struggling with this.

Anthropic, being late to the game of hoarding up all the hardware, decided to "buy time" by hiding CoT, implementing KYC (especially for Chinese users), to delay the efforts of distillation. The products they build in the interim are SaaS clones designed from the POV of AI agents and tight integrations with their models.

And it seems like Google is just sitting aside, watching things unfold, since their business model doesn't stand on inference.

The most likely scenario is that OpenAI and Anthropic will still crash and burn when such open model is released.

Figma's survival is still questinable though. Most likely scenario is likely that there's going to be an open source alternative that has AI integration at the core level, rather than an afterthought.


Exactly this. There are now many great open source coding agents. All we lack is a good model to point them at.

Models are going to commodities - just switch the most affordable.

Longer term, running models locally is going to be increasingly available.


I'm curious, is there anything out there that will get you the 80% of logo work? Claude opus always came closest to me as a non-designer, but it was always still something far from usable.

Haven’t found it yet. The component parts seems to be close. Opus is much better at drawing SVG than it used to be.

Get the draft using pixel generators and convert the end result with svgai.org.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: