[0]: https://github.com/openclaw/openclaw/commit/d378a504ac17eab2...
1. Take the oauth credentials and roll your own agent -- this is NOT allowed
2. Run your agentic application directly in Claude Code -- this IS allowed
When OpenClaw says "Open-Claw style CLI usage", it means literally running OpenClaw in an official Claude Code session. Anthropic has no problems with this, this is compliant with their ToS.
When you use Claude Code's oauth credentials outside of the claude code cli Anthropic will charge you extra usage (API pricing) within your existing subscription.
Which I would not even try and test though if Anthropic did not ban my account. The shadiest thing I did was to use it with opencode for a while I think. Never installed claw or used CC tokens somewhere else.
This is a weird company doing weird shit.
Still early days, but code is available, sort of works if you squint, and welcomes PRs: https://github.com/rcarmo/vibes/tree/go
I have had some ideas for a custom harness (like embedding some tools OOTB and replacing slow tooling) but these policies throw me off. Instead I use local models.
Problem is API costs are insane. I have toyed with the idea of running a local model that works with Claude Sonnet or even Haiku, and I know this has been done by others.
Contrast that to what GitHub did which was to pause new customers to ensure quality remained and things were stable.
Somewhat suspicious that if I do this without an official Anthropic notice I'll lose my precious Max $200/mo account so I'll sit tight perhaps for a while.
Anthropic staff have had contradictive statements in Twitter and have corrected each other. Their intent for clarifications lead to confusion.
> OpenClaw treats Claude CLI reuse and claude -p usage as sanctioned for this integration unless Anthropic publishes a new policy.
Oh cool, so everything is back to business now, until they all or sudden update their policy tomorrow that retracts everything.
Anthropic have proved themselves to be be unreliable when it comes to CC. Switching to other providers is the best way to go, if you want to keep your insanity.
Last year I was excited about the constant forward progress on models but since February or so its just been a mess and I want off this ride.
Either way I’m going to wait for “official” word from Anthropic, which I guess at this point will probably be a “Tell HN” or Reddit text post or a Xitter from some random employee’s personal account, because apparently that’s the state of corporate communication now.
I am specifically talking about switching because of the harness, not model quality. Anyone else match my experience?
I wonder how many other people recently did the same. It would be prudent of Anthropic to let people use Pro/Max OAuth tokens with other harnesses I think. Even though I get why they want to own the eyeballs.
Use something else.
One day you're experimenting just fine. The next, everything breaks.
And I'd gladly use their web containerized agents instead (it would pretty much be the same thing), but we happen to do Apple stuff. So unless we want to dive into relying on ever-changing unreliable toolchains that break every time Apple farts, we're stuck with macOS.
The most recent Anthropic announcement was not that people would be banned for using subscriptions with OpenClaw, but that it would be charged as extra usage. I think the reason this was changed three days after that announcement is that being charged for extra usage meant people would not be banned for using their subscription OAuth tokens directly against the Anthropic API with a third party harness, as they had been before. But rather both that usage, and the more recent claude -p usage both be charged as extra usage.
Release notes and announcements are a well-known agentic anti-pattern.
If you're doing them, you're doing agentic wrong. /s-ish-also-cry
That erosion pushed me to try Codex. I signed up for their most expensive pro plan. Now I'm about to experiment with Kimi. I'm not saying they're better (well, sometimes they are). But here's the thing - what Anthropic did is they made me look. They made a loyal customer start shopping around. And I think that's the worst thing you can do.
Having said that - as an LLM provider for my product, we're staying with Claude. I still trust in their ethics. Please don't prove me wrong.
I agree with GP that this is hard to take seriously.
I had an idea on a whim to vibe-engineer an irccloud replacement for myself.
Started with claude web + Opus 4.7 and continued with Claude Code. Ate up two full cycles of my quota in maybe 6-10 prompts.
Then I iterated on that with pi.dev+codex for HOURS, managed to use 50% of my Codex Pro subscription.
I used to use GLM mostly and had a Claude Pro subscription for occasional review and clean up.
Now I just use GLM.
I do think Claude Max is value for money. But it's more value than I personally need and I like Anthropic less and less.
Anthropic was, even to me, “one of the better ones” until recently. They have made many questionable/poor decisions the last 6-8 weeks and people are right to call them out for it, especially when they want our money.
Best and most applicable typo ever ʕ ´ • ᴥ •̥ ` ʔ
I remember when I’d periodically rage quit from Uber One to Lyft Pink and back again every time I had a terrible customer-service experience. In the end, I realized picking a demon and getting familiar with its quirks was the better way to go.
I’m currently sticking with Claude, in part because I’m not exposed to this nonsense due to OpenClaw, in larger part because of the Hegseth-Altman DoD nonsense. More broadly, however, I’m not sure if any of Google, Anthropic or OpenAI are coming across as stars in AI communication and customer service.
I don't see anything on this page that claims something different from that, or that addresses that claim at all.
But the bills comes thru, one has to pay AWS cause you need the servers, but pay AI agents that make mistake and everyone hopes they work just by typing and saying do x or y. And now they actually invented and engineering and deploy something called Adaptive Thinking and the models can allocate allocate zero reasoning tokens. Its game over, but it was over regardless, there is nothing special about models and they trained them now even with YouTube and soon to be Twitter(X), TikTok and bullshit. Now all those Nvidia GPUs interconnected via NVLink definitely powerful super computers, but the "software" let alone the "AI" is not there yet and OpenAI is worth close to 1 Trillions Dollars ... I mean come on!
Oh no. They won't update the policy. Boris or Thariq will casually mention in a random off-hand commebt on Twitter that this is banned now, and then will gaslight everyone that this has always been the case.
They don't ban Openclaw prompts, each custom LLM application provides a client application id (this is how e.g. Openrouter can tell you how popular Openclaw is, and which models are used the most).
Anthropic just checks for that.
Had to stop because they don't like us proxying requests anymore.
Anthropic models write much better code, they are easy to follow, reasonable and very close to what I would done if I had the time... OpenAI's on the other hand generate extremely complex solutions to the simplest problems.
I was so disappointed by non-Anthropic models, that for a couple of weeks I only used Anthropic models, but based on this thread, I'll go back and give it another try. It's good to go back and try things again every couple of weeks.
Of course, I was annoyed that they lobotomized 4.6, the difference was day and night, and Anthropic is certainly not a company I trust. In my opinion, it shows their willingness to rugpull, so I'm looking at other approaches. Since 4.7, things went back to normal, things you'd expect to work just work.
Some negative signal for better overall view on things: I'm still with Anthropic and will probably stay with them for the foreseeable future.
I think after DoD/DoW shenanigans (which in of itself felt like a reasonable take on the part of Anthrpic) they got a bunch of visibility and new users, so them hitting some scaling limits is pretty much inevitable - so some service disruption is inevitable. Couple this with the tokenizer changes and seeming decrease in model performance (adaptive thinking etc.), and lots of people will be rightfully pissed off, alongside increased downtime (doesn't matter that much for me, definitely does matter for anything time-sensitive).
At the same time, in practice I've only seen it do stupid things across 8 million tokens about 5 times (confusing user/assistant roles, not reading files that should be obvious for a given use case, and picking trivially wrong/stupid solutions when planning things), alongside another 4 times that tests/my ProjectLint tool caught that I would have missed. The error rate is still arguably lower than mine, though I work in a very well known and represented domain (webdev with a bunch of DevOps and also some ML stuff, and integration with various APIs etc.).
At the same time, the 85 EUR they gave to me for free has been enough to weather the instability in regards to pricing changes and peak usage. They've fixed most of the issues I had with Claude Code (notably performance), and the sub-agent support is great and it's way better than OpenCode in my experience. They also keep shipping new features that are pretty nice, like Dispatch and Routines and Design, those features also seem nice and not like something completely misdirected, so that's nice. The Opus 4.7 model quality with high reasoning is actually pretty nice as well and works better than most of the other models I've tried (OpenAI ones are good, I just prefer Claude phrasing/language/approaches/the overall vibe, not even sure what I'd call it exactly, all the stuff in addition to the technical capabilities).
At the same time, if they mess too much with the 100 USD tier, I bet I could go to OpenAI or try out the GLM 5.1 subscription without too many issues. For now they're replacing all the other providers for me. Oh also I find the subscription vs API token-based payment approach annoying, but I guess that's how they make their money.
All these models and agents are shortcuts for all of us to be lazy and play games and watch YouTube or Netflix because we use them to work-less, well the party will be over soon.
I'm confused by the comments being full of people swearing off Claude, feels like real HN bubble stuff.
If I'm paying for compute, why should it matter whether I use Anthropic's harness (e.g., Claude Code) or a 3rd-party harness?
For a while there I had both Opus 4.6 and Codex access and I frequently pitted them against each other, I never once saw Opus come out ahead. Opus was good as a reviewer though, but as an implementer it just felt lazy compared to 5.4 xhigh.
One feature that I haven’t seen discussed that much is how codex has auto-review on tool runs. No longer are you a slave to all or nothing confirmations or endless bugging, it’s such a bad pattern.
Even in a week of heavy duty work and personal use I still haven’t been able to exhaust the usage on the $200 plan.
I’ll probably change my mind when (not IF) OpenAI rug pull, but for spring ‘26, codex is definitely the better deal.
I still have their subscription, but am using pi now, mainly because something happened that made my opencode sessions unusable (cannot continue them, just blanks out, I assume something in the sqlite is fucked), and I cannot be bothered to debug it.
For what I use the agents, the Chinese models are enough
I still VERY occasionally use it (as I'm friggin able to anyway) but it's definitely nowhere near my usage previously. And I refuse to give them money, and besideswhich have no goddamn notion of whether it would even be worth it on the lowest paid tier.
Ah well. The free ride was fun but I knew it had a shelf life.
This was widely reported, and happened to me. You probably can’t reproduce it or see it in docs because they seem to have changed the policy.
The other criticism I see is "ask it what happened in 1989" but as a my use case isn't writing a high school history essay I simply don't care. Or believe one should seek those kind of answers from any AI. (If you're curious it simply cuts off the reply).
I fully appreciate that YMMV and what sits right for others will not align with what's acceptable to me. Anthropic and OpenAI both are in my badbooks as much as Z.ai. pick your poison as they say.
> This is slightly different from what OpenCode was banned from doing; they were a separate harness grabbing a user’s Claude Code session and pretending to be Claude Code.
> OpenClaw was still using Claude Code as the harness (via claude -p)[0]. I understand why Anthropic is doing this (and they’ve made it clear that building products around claude -p is disallowed) but I fear Conductor will be next.
It's just OpenClaw people claiming "Anthropic told us it's fine".
What's not allowed is grabbing the oauth tokens and using these for your own custom agent, which is what was (and still is) banned.
Nothing has changed, this appears to just be a giant misunderstanding (and probably a poor choice of words from Openclaw).
I understand why they have to charge more, but not many are gonna be able to afford even $100 a month, and that doesn't seem to be sufficient.
It has to come with some combination of better algorithms or better hardware.
Google when they merged YouTube and Google+, Reddit multiple times, Facebook after countless scandals. Microsoft destroying windows and pushing ads.
At the end of the day a solid product and company can withstand online controversy.
With Claude it's a constant battle of typing /usage after every iteration and trying to guess if it's enough for the next task or not =)
I will say that Codex high/x-high has consistently performed the best for me, but YMMV
Not that they don't bring value, I'm just not convinced they'll be able to sell their products in a sticky enough way to make up the prices they'll have to extract to make up for the absurd costs.
hn is not a monolith. People here routinely disagree with each other, and that's what makes it great
With Claude Code they can predict what the traffic would look like with third party harness they cannot.
The models and tools levelling out is great for users because the cost of switching is basically nil. I'm reading people ITT saying they signed up for a year - big mistake. A year is a decade right now.
Codex is abysmal for UI design imo.
From what I understand, they still had the Claude Code harness available, but were mostly fully integrated on the pi agent framework, using Claude Code's oauth credentials directly,
At least the only action I was still able to perform was to refund the user, or paypal would have just kept the money.
Anthropic is constantly destroying goodwill and now seems to be in panic mode.
Now with Opus 4.7 of course the “burden” of adjusting reasoning effort has been taken away from you even at the API level.
In my experience people don’t change the thinking level at all.
But if you go information architecture first and have that codified in some way (espescially if you already have the templates), then you can nudge any agent to go straight into CSS and it will produce something reasonable.
Plus I like being able to switch a model.
They also absolutely blocked OpenClaw system prompts from this path in the prior weeks, based purely on keyword detection. Seems they’ve undone that now.
I'd agree with you, except I've heard this argument before. Amazon, Google, Facebook all burned lots of cash, and folks were convinced they would fail.
On the other hand plenty burned cash and did fail. So could go either way.
I expect, once the market consolidates to 2 big engines, they'll make bonkers money. There will be winners and losers. But I can't tell you which is which yet.
I'm currently making an effort to switch to local for stuff that can be local - initially stand alone tasks, longer term a nice harness for mixing. One example would be OCR/image description - I have hooks from dired to throw an image to local translategemma 27b which extracts the text, translates it to english, as necessary, adds a picture description, and - if it feels like - extra context. Works perfectly fine on my macbook.
Another example would be generating documentation - local qwen3 coder with a 256k context window does a great job at going through a codebase to check what is and isn't documented, and prepare a draft. I still replace pretty much all of the text - but it's good at collecting the technical details.
Who’s wrong?
Unless you compare with the reported cash burn or projected losses.
> they’ll raise effective prices some more while Claude diffuses into the economy, sounds like a money printer
But the problem is, they have no moat. Even if Claude diffuses into the economy (still to be seen how much it can effectively penetrate sectors other than engineering, spam, marketing/communications), there is no moat, all providers are interchangeable. If Antrhopic raise the prices too much, switch out to the OpenAI equivalent products.
Edit: I’d also consider waiting for WWDC, they are supposed to be launching the new Mac Studio, an even if you don’t get it, you might be able to snag older models for cheaper
They have a metric called Model-Harness Index:
MHI = 0.50 × ToolCalling + 0.30 × HumanEval + 0.20 × MMLU (scale 0-100)
> Smart Cloud Routing > > Large-context requests auto-route to a cloud LLM (GPT-5, Claude, etc.) when local prefill would be slow. Routing based on new tokens after cache hit. --cloud-model openai/gpt-5 --cloud-threshold 20000
Half the articles are paywalled but the free ones outline the financial situation of the SOTA providers and he has receipts
I disagree very strongly with this, both anecdotally and in the data - subscriptions are growing in all frontier providers; anecdata is right here in HN when you look around almost everyone is talking about CC, codex is a distant second, and completely anecdotally I personally strictly prefer GPT 5.3+ models for backend work and Opus for frontend; Gemini reviews everything that touches concurrency or SQL and finds issues the other models miss.
My general opinion is that models cannot be replaceable, because a model which can replace every other provider must excel at everything all specialist models excel at and that is impossible to serve at scale economically. IOW everyone will have at least two subscriptions to different frontier labs and more likely three.
If tomorrow Kimi release a model better at something, you'd switch to it.
I postulate in practice this won't matter since the space of use cases is so large if Kimi released the absolutely best model at everything they wouldn't be able to serve it (c.f. Mythos).
Anthropic builds the Claude model family and provides access via an API and Claude CLI. In OpenClaw, Anthropic API keys and Claude CLI reuse are both supported. Existing legacy Anthropic token profiles are still honored at runtime if they are already configured.
Best for: standard API access and usage-based billing. Create your API key in the Anthropic Console.
openclaw onboard
# choose: Anthropic API key
# or non-interactive
openclaw onboard --anthropic-api-key "$ANTHROPIC_API_KEY"
{
env: { ANTHROPIC_API_KEY: "sk-ant-..." },
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-6" } } },
}
adaptive thinking in OpenClaw when no explicit thinking level is set./think:<level>) or in model params: agents.defaults.models["anthropic/<model>"].params.thinking.OpenClaw’s shared /fast toggle also supports direct public Anthropic traffic, including API-key and OAuth-authenticated requests sent to api.anthropic.com.
/fast on maps to service_tier: "auto"
/fast off maps to service_tier: "standard_only"
Config default:
{ agents: { defaults: { models: { "anthropic/claude-sonnet-4-6": { params: { fastMode: true }, }, }, }, }, }
Important limits:
api.anthropic.com requests. If you route anthropic/* through a proxy or gateway, /fast leaves service_tier untouched.serviceTier or service_tier model params override the /fast default when both are set.usage.service_tier. On accounts without Priority Tier capacity, service_tier: "auto" may still resolve to standard.OpenClaw supports Anthropic’s prompt caching feature. This is API-only; legacy Anthropic token auth does not honor cache settings.
Use the cacheRetention parameter in your model config:
| Value | Cache Duration | Description |
|---|---|---|
none |
No caching | Disable prompt caching |
short |
5 minutes | Default for API Key auth |
long |
1 hour | Extended cache |
{
agents: {
defaults: {
models: {
"anthropic/claude-opus-4-6": {
params: { cacheRetention: "long" },
},
},
},
},
}
When using Anthropic API Key authentication, OpenClaw automatically applies cacheRetention: "short" (5-minute cache) for all Anthropic models. You can override this by explicitly setting cacheRetention in your config.
Use model-level params as your baseline, then override specific agents via agents.list[].params.
{
agents: {
defaults: {
model: { primary: "anthropic/claude-opus-4-6" },
models: {
"anthropic/claude-opus-4-6": {
params: { cacheRetention: "long" }, // baseline for most agents
},
},
},
list: [
{ id: "research", default: true },
{ id: "alerts", params: { cacheRetention: "none" } }, // override for this agent only
],
},
}
Config merge order for cache-related params:
agents.defaults.models["provider/model"].paramsagents.list[].params (matching id, overrides by key)This lets one agent keep a long-lived cache while another agent on the same model disables caching to avoid write costs on bursty/low-reuse traffic.
amazon-bedrock/*anthropic.claude*) accept cacheRetention pass-through when configured.cacheRetention: "none" at runtime.cacheRetention: "short" for Claude-on-Bedrock model refs when no explicit value is set.Anthropic’s 1M context window is beta-gated. In OpenClaw, enable it per model with params.context1m: true for supported Opus/Sonnet models.
{
agents: {
defaults: {
models: {
"anthropic/claude-opus-4-6": {
params: { context1m: true },
},
},
},
},
}
OpenClaw maps this to anthropic-beta: context-1m-2025-08-07 on Anthropic requests. This only activates when params.context1m is explicitly set to true for that model. Requirement: Anthropic must allow long-context usage on that credential. Note: Anthropic currently rejects context-1m-* beta requests when using legacy Anthropic token auth (sk-ant-oat-*). If you configure context1m: true with that legacy auth mode, OpenClaw logs a warning and falls back to the standard context window by skipping the context1m beta header while keeping the required OAuth betas.
The bundled Anthropic claude-cli backend is supported in OpenClaw.
claude -p usage as sanctioned for this integration unless Anthropic publishes a new policy.claude -p, and Anthropic staff told us OpenClaw-style Claude CLI usage is allowed again. We are treating that guidance as settled unless Anthropic publishes a new policy change.claude -p when available.401 errors / token suddenly invalid
No API key found for provider “anthropic”
openclaw models status.No credentials found for profile anthropic:default
openclaw models status to see which auth profile is active.No available auth profile (all in cooldown/unavailable)
openclaw models status --json for auth.unusableProfiles.More: /gateway/troubleshooting and /help/faq.