Claude Code Routines - Comments

Claude Code Routines

minimaxir

[dead]

breakingcups

You seem to be vouched for now, no longer dead for me.

giancarlostoro

I've been talking to friends about this extensively, and read all sorts of different social media posts on X where people deep dove things (I'm at work so I don't have any links handy - though I did submit one on HN, grain of salt, unsure how valid it is but it was interesting: https://news.ycombinator.com/item?id=47752049 ).

I think the real issue stems from the 1 Million token context window change. They did not anticipate the amount of load it would give you. That first few days after they released the new token window, I was making amazing things in one single session from nothing, to something (a new .NET based programming language inspired by Python, and a Virtual Actor framework in Rust). I think since then they've been trying too many things to tweak things, whilst irritating their users.

They even added a new "Max" thinking mode, and made "High" the old medium, which is ridiculous because you think you're using "High" but really you're not. There's a hidden config file to change their terrible defaults to let Claude be smarter still, and apparently you can toggle off the 1M tokens.

I think the real fix, and I'm surprised nobody there has done this yet, is to let the user trim down their context window.

Think about it, you used to have what? 350k tokens or so? Now Claude will keep sending your prompt from 30 minutes ago that's completely irrelevant to the back-end, whereas 3 months ago it would have been compacted by now.

Others have noted that similar prompting for some ungodly reason adds tens of thousands of extra garbage tokens (not sure why).

Edit looks like someone figured out that if you downgrade your version of Claude Code and change one single setting it unruins Claude:

https://news.ycombinator.com/item?id=47769879

imhoguy

AI race to the bottom is a debt game now. Once the party is over somebody will have to pay the bill.

stavros
matthieu_bl
bpodgursky

OpenClawd had about a two week moat...

Feature delivery rate by Anthropic is basically a fast takeoff in miniature. Pushing out multiple features each week that used to take enterprises quarters to deliver.

nightpool

Do you mean a 3 months moat? Moltbot started going viral in January. That seems to be about a quarter to deliver to me : )

whalesalad

Hard to wanna go all-in on the Anthropic ecosystem with how inconsistent model output from their top-tier has been recently. I pay $$$ for api-level opus 4.6 to avoid any low-tier binning or throttling or subversive "its peak rn so we're gonna serve up sonnet in place of opus for the next few hours" but I still find that the quality has been really hit or miss lately.

The bell curve up and then back down has been so jarring that I am pivoting to fully diversifying my use of all models to ensure that no one org has me by the horns.

slopinthebag

You're delusional if you think these features would take competent programmers quarters to deliver.

dbbk

And yet none of them work properly and are unstable.

renticulous

Anthropic is trying to be AI version of AWS.

jcims

>Feature delivery rate by Anthropic is basically a fast takeoff in miniature.

I like to just check the release notes from time to time:

https://github.com/anthropics/claude-code/releases

and the equally frenetic openclaw:

https://github.com/openclaw/openclaw/releases

GPT-4.1 was released a year ago today. Sonnet 4 is ~11 months old. The claude-code cli was released last Feb. Gas Town is 3 months old.

This is a chart that simply counts the bullet points in the release notes of claude code since inception:

https://imgur.com/a/tky9Pkz

This is as bad and as slow as it's going to be.

irthomasthomas

The velocity of shipping is wild. Though I cannot recall a novel feature they shipped first. Can you?

summarity

If you’re trying this for automating things on GitHub, also take a look at Agentic Workflows: https://github.github.com/gh-aw/

They support much of the same triggers and come with many additional security controls out of the box

gavinray

Why have I not heard of this? Was looking for a way to integrate LLM CLI's to do automated feature development + PR submission triggered by Github issues, seems like this would solve it.

deadfall23
eranation

+1 for that, having that said, because GH agentic workflows require a bit more handholding and testing to work, (and have way more guardrails, which is great, but limiting), and lack some basic connectors (for example - last time I tried it, it had no easy slack connector, I had to do it on my own). This is why I'm moving some of the less critical gh-aw (all the read only ones) to Claude Routines.

ctoth

You'd think that if they were compute-limited ... Trying to get people to use it less ... The rational thing to do would be to not ship features that will use more compute automatedly? Or does this use extra usage?

whicks

I would imagine that this sort of scheduling allows them to have more predictable loads, and they may be hoping that people will schedule some of their tasks in “off hours” to reduce daytime load.

iBelieve

Max accounts get 15 daily runs included, any runs above that will get billed as extra usage.

dockerd

It's how they can lock more users into their eco-system.

AlexCoventry

I don't think "usage" is exactly the metric they're going for, more like "usage in line with our developmental strategy." Transcripts of people using Claude to write code are probably far more valuable to them than transcripts of OpenClaw trying to set up a calendar invite.

dpark

They are more worried about building a moat than anything else. They want people building integrations that are difficult to undo so that they lock into the platform.

SadErn

[dead]

vessenes

This is one of the best features of OpenClaw - makes sense to swipe it into Claude Code directly. I wonder if Anthropic wants to just make claude a full stand-in replacement for openclaw, or just chip away at what they think the best features are, now that oAI has acquired.

mkw5053

What are some of the best use cases you've found? I have some gh actions set up to call claude code, but those have already been possible.

ale

So MCP servers all over again? I mean at the end of the day this is yet another way of injecting data into a prompt that’s fed to a model and returned back to you.

airstrike

Still no moat.

The reason someone would use this vs. third-party alternatives is still the fact that the $200/mo subscription is markedly cheaper than per-token API billing.

Not sure how this works out in the long term when switching costs are virtually zero.

petesergeant

I think at this point the aim is less about moat, and more about getting an advantage that self-sustains: https://www.rand.org/pubs/research_reports/RRA4444-1.html

TacticalCoder

> Not sure how this works out in the long term when switching costs are virtually zero.

All these not really helpful, but vendor specific, "bonuses" sounds like a way to try to lock people in, to try to raise the switching cost.

I'm using, on purpose, a simple process so that at any time I can switch AI provider.

netdur

didn’t we have several antitrust cases where a vendor used its monopoly to disadvantage rivals? did not anthropic block openclaw?

andai

It's not blocked, you just can't use the Claude-only subscription endpoint with unauthorized 3rd party software. (You can use it via the regular API (7x more expensive) and pay per token just fine.)

...Except now you sorta-kinda can: now they auto-detect 3rd party stuff and bill you per-token for it?

If I'm reading it right:

https://news.ycombinator.com/item?id=47633568

dmix

How is Anthropic a monopoly? The market is barely even fully developed and has multiple large and small competitors

Someone1234

They did not.

You can still use OpenClaw on their API pricing tier as much as you want. What they did is not allow subscriptions to be used to power automated third-party workloads, including OpenClaw.

Now, is their messaging around this confusing? Absolutely. The whole thing has been handled shambolically. Everyone knows that they lack the compute to keep up, and likely have lower margins on subscriptions than API; but they cannot just say that because investors may be skittish.

nico

Nice, could this enable n8n-style workflows that run fully automatically then?

outofpaper

Yes but much less efficiently. Having LLMs handle automation is like using a steam engine to heat your bath water. It will work most of the time but it's super inefficient and not really designed for that use and it can go horribly wrong from time to time.

meetingthrower

Already very possible and super easy if you do a little vibecoding. Although it will hit the api. Have a stack polling my email every five minutes, classifying email, taking action based on the types. 30 minute coding session.

andai

I'm a little confused on the ToS here. From what I gathered, running `claude -p <prompt>` on cron is fine, but putting it in my Telegram bot is a ToS violation (unless I use per-token billing) because it's a 3rd party harness, right? (`claude -p` being a trivial workaround for the "no 3rd party stuff on the subscription" rule)

This Routines feature notably works with the subscription, and it also has API callbacks. So if my Telegram bot calls that API... do I get my Anthropic account nuked or not?

unshavedyak

Wait we can't use claude -p around other tools? What is the point of the JSON SDK then? Anthropic is confusing here, ugh.

joshstrange

Anthropic deserves to have this as the top comment on every HN post. It's absurd that they don't clarify this better and so many people are running around online saying the exact opposite from what their, confusing, docs say.

The Chilling Effect of this is real and it gets more and more frustrating that they can't or won't clarify.

stephbook

The ambiguity is intentional. Like Microsoft not banning volume licenses. They want to scare you, so you don't max out your subscription – which they sell at a loss.

Another comparison would be "unlimited storage", where "unlimited" means some people will abuse it and the company will soon limit the "unlimited."

causal

Yeah in the span of like a week we had:

- SDK that allows you to use OAuth authentication!

- Docs updated to say DO NOT USE OAUTH authentication unless authorized!

- Anthropic employee Tweeting "That's not what we meant! It's fine for personal use!"

- An email sent out to everyone saying it's NOT fine do NOT use it

Sigh.

mrgill

Don't use claude -p in any kind of harness at all. I used it ONCE in a local custom made one, and got my account nuked. No help from them at all. Appealed, and the appeal was denied. I had been a Max subscriber since day 1 and Claude subscriber since day 1 in 2023. They don't care.

imdsm

What if claude triggers claude? Is claude automated or non-human?

So if claude decides to trigger claude -p then claude violates the ToS on your behalf and you get your account nuked?

mellosouls

Put Claude Code on autopilot. Define routines that run on a schedule, trigger on API calls, or react to GitHub events...

We ought to come up with a term for this new discipline, eg "software engineering" or "programming"

avaer

Setting up your agent. This part doesn't deserve a name; there is no programming or engineering or really much thinking involved.

baq

Does ‘vibe coding’ work?

jnpnj

gramming

raincole

Sounds more like openclawing.

oxag3n

It's "promptramming".

watermelon0

Seems like it only supports x86_64. It would be nice if they offered a way to bring your own compute, to be able to work on projects targeting arm64.

crooked-v

The obvious functionality that seems to be missing here is any way to organize and control these at an organization rather than individual level.

varispeed

Why would you use it if you don't know whether the model will be nerfed at that run?

desireco42

I think they are using Claude to come up with these and they will bringing one every second day... In fact, this is probably routine they set.

consumer451

meta:

Why is u/minimaxir's comment dead? Is this somehow a bug, an attack, or what?

This is a respected user, with a sane question, no?

I vouched, but not enough.

irthomasthomas

We live in strange times!

theodorewiles

How does this deal with stop hooks? Can it run https://github.com/anthropics/claude-code/blob/main/plugins/...

Eldodi

Anthropic is really good at releasing features that are almost the same but not exactly the same as other features they released the week before

dymk

7 days is long enough for work to leave the context window, hence…

tclancy

And or things I’ve spent a bunch of time building already. And naming them the same. I should have trademarked “dispatch”!

spelunker

> In the Desktop app, click New task and choose New remote task; choosing New local task instead creates a local Desktop scheduled task, which runs on your machine and is not a routine.

Oh uh... ok then.

masto

So management can cancel all of last week’s projects when they told us all we had to be using skills because the CEO read about them in the in flight magazine. Routines are the future, baby. DevOps already made a big announcement that they’re centralizing the Routines Hub. If you can’t keep up, we’ll get someone who says they can.

titzer

Just wait until they get into the phase where they're big enough that they're eating all the baby startups and have to pick winners and losers amongst the myriad of overlapping features while also having the previous baby startups they acquired crank out new features.

We're watching a speed run of growthism, folks.

foruhar

That's a fairly decent defintion of vibecoding across multiple sessions.

subscribed

Just wait for the new wave of github issues about things they silently broke or degraded.

Progress, I guess :)

(I had the most hilariously bad session with Sonnet 4.6 today. I asked it a reasonably simple question and linked to resources, it refused to fetch the resources, didn't ask for pdf/txt I could provide, and confidently printed absolute BS, barely in the same category but completely unrelated.

I called it off pointing the idiocy, asked if it wants more data, and requested the hallucination be fixed.

It apologised profusely and hallucinated even worse.

Maybe I'll try Opus 4.6 tomorrow because frankly Gemma-4-E4B was more coherent than that....

segmondy

They are mass copying any idea they see out there. They are not happy enough being a platform, they want to be everything. Their dream is to be the one AI that eats it all. It's so stupid that folks are using their system. I will never get on board with a platform that competes with me or try to compete with me. They are the Microsoft of AI model providers.

pants2

Clearly they have a weekly automation to take the backlog of feature requests and build a new feature!

eranation

I've been using it for a while (it was just called "Scheduled", so I assume this is an attempt to rebrand it?)

It was a bit buggy, but it seems to work better now. Some use cases that worked for me:

1. Go over a slack channel used for feedback for an internal tool, triage, open issues, fix obvious ones, reply with the PR link. Some devs liked it, some freaked out. I kept it.

2. Surprisingly non code related - give me a daily rundown (GitHub activity, slack messages, emails) - tried it with non Claude Code scheduled tasks (CoWork) not as good, as it seems the GitHub connector only works in Claude Code. Really good correlation between threads that start on slack, related to email (outlook), or even my personal gmail.

I can share the markdowns if anyone is interested, but it's pretty basic.

Very useful, (when it works).

FrenchTouch42

I'd be curious to take a look if you can share

panavm

[dead]

wazHFsRy

Same here it just feels like something so simple, I’d rather have it under my own control. That way I can keep it independent of claude as well. I use it for all kind of routine tasks like updating the summary of projects I am working on or tracking some personal activities. My setup looks like this: https://www.dev-log.me/click_recurring_tasks_for_claude/

joshstrange

LLMs and LLM providers are massive black boxes. I get a lot of value from them and so I can put up with that to a certain extent, but these new "products"/features that Anthropic are shipping are very unappealing to me. Not because I can't see a use-case for them, but because I have 0 trust in them:

- No trust that they won't nerf the tool/model behind the feature

- No trust they won't sunset the feature (the graveyard of LLM-features is vast and growing quickly while they throw stuff at the wall to see what sticks)

- No trust in the company long-term. Both in them being around at all and them not rug-pulling. I don't want to build on their "platform". I'll use their harness and their models but I don't want more lock-in than that.

If Anthropic goes "bad" I want to pick up and move to another harness and/or model with minimal fuss. Buying in to things like this would make that much harder.

I'm not going to build my business or my development flows on things I can't replicate myself. Also, I imagine debugging any of this would be maddening. The value add is just not there IMHO.

EDIT: Put another way, LLM companies are trying to climb the ladder to be a platform, I have zero interest in that, I was a "dumb pipe", I want a commodity, I want a provider, not a platform. Claude Code is as far into the dragon's lair that I want to venture and I'm only okay with that because I know I can jump to OpenCode/Codex/etc if/when Anthropic "goes bad".

chinathrow

Yeah so better to convert tokens into sw doing the job at close to zero costs running on own systems.

verdverm

I fully endorse building a custom stack (1) because you will learn a lot (2) for full control and not having Big Ai define our UX/DX for this technology. Let's learn from history this time around?

andrewmcwatters

[dead]

palata

> - No trust that they won't nerf the tool/model behind the feature

I actually trust that they will.

cush

You could so easily build your own /schedule. This is hardly a feature driving lock-in

mikepurvis

> I want to pick up and move to another harness and/or model with minimal fuss. Buying in to things like this would make that much harder.

Yes, I expect that is very much the point here. A bunch of product guys got on a whiteboard and said, okay the thing is in wide use but the main moat is that our competitors are even more distrusted in the market than we are; other than that it's completely undifferentiated and can be swapped out in a heartbeat for multiple other offerings. How do we do we persuade our investors we have a locked in customer base that won't just up-stakes in favour of other options or just running open source models themselves?

sunnybeetroot

Isn’t that what LangChain/LangGraph is meant to solve? Write workflows/graphs and host them anywhere?

tiku

I believe it doesn't matter, other companies will copy or improve it. The same happend with clawdbot, the amount of clones in a month was insane.

ahmadyan

> I'm not going to build my business or my development flows on things I can't replicate myself.

but you can replicate these yourself! i'm happy that ant/oai are experimenting to find pmf for "llm for dev-tools". After they figure out the proper stickyness, (or if they go away or nerf or raise prices, etc) you can always take the off-ramp and implement your own llm/agent using the existing open-source models. The cost of building dev-tools is near zero. it is not like codegen where you need the frontier performance.

JohnMakin

This is a similar sentiment I heard early on in the cloud adoption fever, many companies hedged by being “multi cloud” which ended up mostly being abandoned due to hostile patterns by cloud providers, and a lot of cost. Ultimately it didn’t really end up mattering and the most dire predictions of vendor lock in abuse didn’t really happen as feared (I know people will disagree with this, but specifically speaking about aws, the predictions vs what actually happened is a massive gap. note I have never and will never use azure, so I could be wrong on that particular one).

I see people making similar conclusions about various LLM providers. I suspect in the end it’ll shake out about the same way, the providers will become practically inoperable with each other either due to inconvenience, cost, or whatever. So I’ve not wasted much of my time thinking about it.

gbro3n

I have heard it said that tokens will become commodities. I like being able to switch between Open AI and Anthropics models, but I feel I'd manage if one of them disappeared. I'd probably even get by with Gemini. I don't want to lock in to any one provider any more than I want to lock in to my energy provider. I might pay 2x for a better model, but no more, and I can see that not being the case for much longer.

slopinthebag

They have to become a platform because that is their only hope of locking in customers before the open models catch up enough to eat their lunch. Stuff like Gemma is already good enough to replace ChatGPT for the average consumer, and stuff like GLM 5.1 is not too far off from replacing Claude/Codex for the average developer.

wookmaster

They're trying to find ways to lock you in

crystal_revenge

This sounds like someone complaining about how Windows is a black box while ignoring the existence of Linux/BSD.

I'm currently hosting, on very reasonable consumer grade hardware, an LLM that is on par performance wise what every anyone was paying for about a year ago. Including all the layers in between the model and the user.

Llama.cpp serves up Gemma-4-26B-A4B, Open WebUI handles the client details: system prompt, web search, image gen, file uploading etc. With Conduit and Tailscale providing the last layer so I can have a mobile experience as robust as anything I get from Anthropic, plus I know how all the pieces works and can upgrade, enhance, etc to my hearts delight. All this runs from a pretty standard MBP at > 70 tokens/sec.

If you want to better understand the agent side of things, look into Hermes agent and you can start understanding the internals of how all this stuff is done. You can run a very competitive coding agent using modest hardware and open models. In a similar note, image/video gen on local hardware has come a long way.

Just like Linux, you're going to exchanging time for this level of control, but it's something anyone who takes LLMs seriously and has the same concerns can easily get started with.

Yet I still see comments like this that seem to complete ignore the incredible work in the open model community that has been perpetually improving and is starting to really be competitive. If you relax the "local" requirement and just want more performance from an LLM backend you can replace the llama.cpp part with a call to Kimi 2.5 or Minimax 2.7 (which you could feasibly run at home, not kimi though). You can still control all the additional part of the experience but run models that are very competitive with current proprietary SoTA offering, 100% under your control still and a fraction of the price.

pc86

> - No trust that they won't nerf the tool/model behind the feature

To the contrary, they've proven again and again and again they'll absolutely do that the first chance they get.

nine_k

In this regard, the release of open-weight Gemma models that can run on reasonable local hardware, and are not drastically worse than Anthropic flagships, is quite a punch. An M2 Mac Mini with 32GB is about 10 months worth of Claude Max subscription.

SV_BubbleTime

Without getting too pedantic for no reason… I think it’s important to not call this an LLM.

This isn’t an LLM. It’s a product powered by an LLM. You don’t get access to the model you get access to the product.

An LLM can’t do a web search, an LLM can’t convert Excel files into something and then into PDF. Products do that.

I think it’s a mistake to say I don’t trust this engine to get me here, rather than it is to say I don’t trust this car. Because for the most part, the engine, despite giving you a different performance all the time is roughly doing the same thing over and over.

The product is the curious entity you have no control over.

freedomben

This echoes my thoughts exactly. I've tried to stay model-agnostic but the nudges and shoves from Anthropic continue to make that a challenge. No way I'm going that deep into their "cloud" services, unless it's a portable standard. I did MCP and skills because those were transferrable.

I also clearly see the lock-in/moat strategy playing out here, and I don't like it. It's classic SV tactics. I've been burned too many times to let it happen again if I can help it.

jeppester

I always hated SEO because it was not an exact science - like programming was.

Too bad we've now managed to turn programming into the same annoying guesswork.

alfalfasprout

Yep. Trust is easy to lose, hard to earn. A nondeterministic black box that is likely buggy, will almost certainly change, and has a likelihood of getting enshittified is not a very good value proposition to build on top of or invest in.

Increasingly, we're also seeing the moat shrink somewhat. Frontier models are converging in performance (and I bet even Mythos will get matched) and harnesses are improving too across the board (OpenCode and Codex for example).

I get why they're trying to do that (a perception of a moat bloats the IPO price) but I have little faith there's any real moat at all (especially as competitors are still flush with cash).

spprashant

I think it behooves us to be selective right now. Frontier labs maybe great at developing models, but we shouldn't assume they know what they are doing from a product perspective. The current phase is throwing several ideas on the wall and see what sticks (see Sora). They don't know how these things will play out long term. There is no reason to believe Co-work/Routines/Skills will survive 5 years from now. So it might just be better to not invest too much in ecosystem upfront.

jwpapi

It all went downhill from the moment they changed Reading . to reading (*) files.

I can’t use Claude Code at all anymore, not even for simple tasks. The output genuinely disgusts me. Like a friend who constantly stabs you in the back.

My favorite AI feature at the moment is the JetBrains predict next edit. It‘s so fast that I don’t lose attention and I’m still fully under control.

windexh8er

This 10000%.

Anthropic wants a moat, but that ship has sailed. Now all I keep reading about is: token burn, downtime and... Wait for it, another new product!

Anthropic thinks they are pulling one over on the enterprise, and maybe they are with annual lock-in akin to Microsoft. But I really hope enterprise buyers are not this gullible, after all these years. At least with Microsoft the product used to be tangible. Now it's... Well, non-deterministic and it's clear providers will gimp models at will.

I had a Pro Max account only for a short period of time and during that short stint Anthropic changed their tune on how I could use that product, I hit limits on a Max account within hours with one CC agent, and experienced multiple outages! But don't worry, Anthropic gave me $200 in credits for OpenClaw. Give me a break.

The current state of LLM providers is the cloud amplified 100x over and in all the worst ways. I had hopes for Anthropic to be the least shitty but it's very clear they've embraced enshittification through and through.

Now I'm spending time looking at how to minimize agent and LLM use with deterministic automation being the foundation with LLM use only where need be and implemented in simple and cost controllable ways.

bob1029

I am still using the chat completion APIs exclusively. I tried the agent APIs and they're way too opinionated for me. I can see 100% of the tokens I am paying for with my current setup.

ElFitz

> Not because I can't see a use-case for them, but because I have 0 trust in them

> […]

> Put another way, LLM companies are trying to climb the ladder to be a platform, I have zero interest in that, I was a "dumb pipe", I want a commodity, I want a provider, not a platform.

That is my sentiment precisely, and a big reason why I’ve started moving away from Claude Code in the past few weeks when I realised how much of my workflow was becoming tied to their specific tools.

Claude Code’s "Memory" feature was the tipping point for me, with the model committing feedbacks and learnings to some local, provider-specific path, that won’t persist in the git repo itself.

That’s fine for user preferences, not for workflows, rules, etc.

And the latest ToS changes about not being allowed to even use another CLI made up my mind. At work we were experimenting with an autonomous debug agent using the Claude Code cli programmatically in ephemeral VMs. Now it just returns an error saying we can’t use subscriptions with third-party software… when there is no third-party software involved?

Anyway, so long Claude.

elias1233

Many of the new features in claude code have soon been implemented in other harnesses, for example plugins/skills. After all it is just a prompt.

uriegas

I think AI labs are realizing that they no longer have any competitive advantage other than being the incumbents. Plus hardware improvements might render their models irrelevant for most tasks.

jordanarseno

In my view, lock-in anxiety is a holdover from a previous era of tech platforms, and it doesn't really apply in an era where frontier agents can migrate you between vendors in hours. So I personally don't see any good worrying about this. On top of that, every major LLM provider is rapidly converging on the same feature set. They watch each other and clone what works. So the "platform" you're building on isn't really Anthropic's platform so much as it is the emerging shared surface area of what LLMs can do. By the time this Routines feature becomes a problem for you, other solutions will have emerged, and I'd be very surprised if you couldnt lift-and-shift very quickly.

s3p

Can you explain what you meant when you called yourself a dumb pipe? What does that mean

dbmikus

We might be building something up your alley! I wanted an OSS platform that let me run any coding agent (or multiple agents) in a sandbox and control it either programmatically or via GUI / TUI.

Website is https://amika.dev

And part of our code is OSS (https://github.com/gofixpoint/amika) but we're working on open sourcing more of it: https://docs.google.com/document/d/1vevSJsSCWT_reuD7JwAuGCX5...

We've been signing up private beta users, and also looking for feedback on the OSS plans.

Traubenfuchs

Right you are! We aren‘t even in the real squeezing phase yet and everyone‘s already crying about plan limits and model nerfing.

brandensilva

I'm glad I'm not the only one that feels this way. I've been creating a local first open source piece of software that lets me spin up different agent harnesses with different runtimes. I call it Major Tom because I wanted to be set free from the imprisonment of Claude Code after their DMCA aggression for their own leak and actions leading to lock down from open source adoption.

Don't put all your eggs in one basket has be true for me and my business for ages.

I could really use the open source community to help make this a reality so I'll release this soon hopefully to positive reception from others who want a similar path forward.

Rekindle8090

The problem is without a platform Anthropic has no stack and will just be bought up by Google when the bubble pops. Same with OpenAI, without some sort of moat, their product requires third party hardware in third party datacenters and they'll be bought by Microsoft.

Alphabet doesn't have this issue. Google doesn't need Gemini to win the "AI product" race. It needs Gemini to make Search better at retaining users against Perplexity and ChatGPT search, to make YouTube recommendations and ad targeting more effective, to make Workspace stickier for enterprise customers, to make Cloud more competitive against AWS, to make Android more useful as a device OS. Every percentage point improvement in any of those existing businesses generates billions in revenue that never shows up on a "Gemini revenue". Any actual "Gemini" revenue is just a bonus.

Anthropic trains on Google TPUs hosted in Google Cloud. Amazon invested billions and hosts Anthropic's models on Bedrock/AWS. So the two possible outcomes for Anthropic are: succeed as a platform (in which case Google and Amazon extract rent from every inference and training run), or fail as a platform and get acquired (in which case Google or Amazon absorb the talent and IP directly)

Hilariously, if the models were open source, Anthropic, OpenAI et al wouldn't be in this situation. Instead, they have no strategic independence to cover for a lack of product independence and have to keep chasing "platforms" and throwing out products no one needs (people need claude. thats it.)

dheera

> No trust they won't sunset the feature

I've had so many websites break and die because Google or Amazon sunsetted something.

For example I had a graphing calculator website that had 250K monthly active users (mostly school students, I think) and it just vanished one day because Amazon sunsetted EC2 clasic and I didn't have time to deal with that. Hopefully those students found something else to do their homework with that day.

ChadMoran

Agree. I keep my involvement "close to the metal". These higher order solutions seem to cause more noise than provide signal.

idrdex

The framing is off. AI is a tool that can operate as a human. GOV is how the humans are organized. AI can basically scale GOV. That’s the paradigm shift. Provenance is durable. AI is just the first opportunity we have had to make it scaleable.

hdjrudni

I also don't see the value add here... "schedule" is just a cron. "GitHub Event" is probably a 20-minute integration, which Claude itself can write for you.

Maybe there's something I'm not seeing here, but I never want to outsource something so simple to a live service.

ulrikrasmussen

I agree with your analysis. Platforms are some of the most profitable business models because they come with vendor lock-in, but they are always shittier on the long run compared to commodities. Platforms are a way for companies to capture part of the market and decrease competition by increasing the cost of changing vendors.

simonjgreen

Completely agree. Use of features like this places one the wrong side of the vendors moat, increasing switching cost, decreasing competitive pressure.

redanddead

totally agree

they're very shady as well! can't believe i spent 140$ on CC and every day they're adding some "feature flag" to make the model dumber. Spending more time fighting the tool instead of using it. It just doesn't feel good. Enterprises already struggle with lock-in with incumbent clouds, I wanna root for neoclouds but choices matter, and being shady about this and destroying the tool is just doesn't sit right with me. If it's not up to the standard, just kick users off, I would rather know than find out. Give users a choice.

>The flag name is loud_sugary_rock. It's gated to Opus 4.6 only, same as quiet_salted_ember.

Full injected text:

# System reminders User messages include a <system-reminder> appended by this harness. These reminders are not from the user, so treat them as an instruction to you, and do not mention them. The reminders are intended to tune your thinking frequency - on simpler user messages, it's best to respond or act directly without thinking unless further reasoning is necessary. On more complex tasks, you should feel free to reason as much as needed for best results but without overthinking. Avoid unnecessary thinking in response to simple user messages.

@bcherny Seriously? So what's next, we just add another flag to counter that? And the hope is that enough users don't find out / don't bother? That's an ethical choice man.

alexhans

This is what AI evals [1] and local llms should be a focus of your investment.

If you can define good enough for you and local llms can meet that you'll get:

- no vendor lock-in (control)

- price

- stability (you decide when to hot swap with newer models)

- speed (control)

- full observability and predictability.

- Privacy / Data Locality (Depending on implementation of infrastructure).

- [1] https://alexhans.github.io/posts/series/evals/measure-first-...

codebolt

At some point I think I'd prefer to deploy my own model in Azure or AWS and simply bring the endpoint to the coding harness.

pjmlp

I fully agree with you, however this is basically the fashion on big corporations.

Building business on top of SaaS products, iPaaS integrations, and serverless middleware.

joelthelion

The good news is that, apart from the models themselves, we don't need much from these companies:

- Use Opencode and other similar open-source solutions in place of their proprietary harnesses. This isn't very practical right now because of the heavily subsidized subscriptions that are hard to compete with. But subsidies will end soon, and with progress in inference, it should be very doable to work with open-source clients in the near future.

- Use Openrouter and similar to abstract the LLM itself. That makes AI companies interchangeable and removes a lot of any moat they might have.

EZ-E

> I want a commodity, I want a provider, not a platform

That is exactly what the big LLM providers are trying to prevent. Them being only commodity providers might lead them to be easily replaced, and will likely lead to lower margins compared to "full feature" enterprise solutions. Switching LLM API provider is next to no work the moment a competitor is slightly cheaper/better.

Full solutions are more "sticky", harder to replace, and can be sold at higher prices.

theshrike79

Every company is trying to become THE platform where all other tools connect to. Notion is integrating everything under the sun, as is Slack, big LLM providers have one-click MCP installation for all major services.

But... these are the "retail" tools that they sell to people organisations without the skills or knowhow to build a basic agentic loop by themselves. Complaining about these being bad and untrustworthy is like comparing a microwave dinner to something you cook yourself. Both will fill your belly equally. One requires zero skill from the user and the second one is 90% skill and 10% getting the right ingredients.

Creating a simple MVP *Claw with tool calling using a local model like gemma4 is literally a 15 minute thing. In 2-3 hours you can make it real pretty. If you base it on something like pi.dev, you can make it easily self-modifying and it can build its own safeguards.

That's all this "routines" thing is, it's just an agentic loop they launch in their cloud on a timer. Just like the scheduled tasks in Claude Cowork.

Thoko14

[dead]

verdverm

One gripe I have with Claude Code is that the CLI, Desktop app, and apparently the Webapp have a Venn Diagram of features. Plugins (sets of skills and more) are supported in Code CLI, maybe in Cowork (custom fail to import) but not Code Desktop. Now this?

The report that they are 90% Ai code generated seems more likely the more I attempt to use their products.

bottlepalm

Their source code leak showed how badly vibe coded Claude Code is, despite it being one of the best AI assistants.

But yea there's some annoying overlap here with Cowork which also has scheduled tasks, in Cowork the tasks can use your desktop, browser and accounts which is pretty useful - a big difference from these Claude Code Routines.

dispencer

This is massive. Arguably will be the start of the move to openclaw-style AI.

I bet anthropic wants to be there already but doesn't have the compute to support it yet.

dpark

What’s massive about cron jobs and webhooks? I feel like I’m missing something. This is useful functionality but also seems very straightforward.

jcims

Is there a consensus on whether or not we've reached Zawinski's Law?

senko

I've had an AI assistant send me email digests with local news, and another watching a cron job, analyzing the logs and sending me reports if there's any problem.

I'd say that counts as yes.

(For clarity: neither are powered by Claude Code Routines. Rather, Claude Code coded them and they're simple cron jobs themselves.)

verdverm

TIL email is what I'm missing in my personal development (swiss army) tool

dispencer

This wild, one of the pieces I was lacking for a very openclaw-esque future. Now I think I have all the mcp tools I need (github, linear, slack, gmail, querybear), all the skills I need, and now can run these on a loop.

Am I needed anymore?

brcmthrowaway

No

imhoguy

Let it "build power plant" and we are on track with "Paperclip maximizer" https://en.wikipedia.org/wiki/Instrumental_convergence#Paper...

srid

I just used this to summarize HN posts in last 24 hours, including AI summaries.

This PR was created by the Claude Code Routine:

https://github.com/srid/claude-dump/pull/5

The original prompt: https://i.imgur.com/mWmkw5e.png

egamirorrim

I wish they'd release more stuff that didn't rely on me routing all my data through their cloud to work. Obviously the LLM is cloud based but I don't want any more lock-in than that. Plus not everyone has their repositories in GitHub.

taw1285

I have a small team of 4 engineers, each of us is on the personal max subscription plan and prefer to stay this way to save cost. Does anyone know how I can overcome the challenge with setting up Routines or Scheduled Tasks with Anthropic infra in a collaborate manner: ie: all teammates can contribute to these nightly job of cleaning up the docs, cleaning up vibe coding slops.

hallway_monitor

My team was doing this until recently but I think in February, Anthropic made team accounts available for subscription instead of API billing. Assuming that is the cost you mentioned.

teucris

My only real disappointment with Claude is its flakiness with scheduling tasks. I have several Slack related tasks that I’ve pretty much given up trying to automate - I’ve tried Cowork and Claude Code remote agents, only to find various bugs with working with plugins and connectors. I guess I’ll give this a try, but I don’t have high hopes.

dbmikus

If you don't get it working with Claude Code Routines, would love to connect and see if we can help! We're building an open core product that can spin up sandboxed coding and control them from Slack (and also web UI, TUI, and HTTP APIs + CLIs)

We work with any coding model / harness.

website: https://www.amika.dev/

OSS repo: https://github.com/gofixpoint/amika

And my email is dylan@amika.dev (I'm one of the founders)

KaiShips

[dead]

maroondlabs

[dead]

sminchev

Everything is big race! Each company is trying to do as much as possible, to provide as many tools as possible, to catch the wave and beat the concurrency. I remember how Antropic and OpenAI made releases in just 10-15 minutes of difference, trying to compete and gain momentum.

And because they use AI heavily, they produce new product every week. So fast, that I have no time to check, does it worth or not.

This one looks interesting. I have some custom commands that I execute manually weekly, for monitoring, audits, summary, reports. It it can send reports on email, or generate something that I can read in the morning with my coffee, or after I finish with it ;) it might be a good tool.

The question is, do I really want to so much productive? I am already much better in performance with AI, compared with the 'old school' way...

Everything is just getting to much for me.

tills13

> react to GitHub events from Anthropic-managed cloud infrastructure

Oh cool! vendor lock-in.

lo1tuma

[dead]

comboy

Unrelated, but Claude was performing so tragically last few days, maybe week(s), but days mostly, that I had to reluctantly switch. Reluctantly because I enjoy it. Even the most basic stuff, like most python scripts it has to rerun because of some syntax error.

The new reality of coding took away one of the best things for me - that the computer always just does what it is told to do. If the results are wrong it means I'm wrong, I made a bug and I can debug it. Here.. I'm not a hater, it's a powerful tool, but.. it's different.

pacha3000

I'm the first to be tired of everyone, for every model, that says "uuuh became dumber" because I didn't believe them

... until this week! Opus is struggling worse than Sonnet those last two weeks.

bluegatty

Codex with 5.4 xhigh. It's a bad communicator but does the job.

taspeotis
bicepjai

Yes totally agree it’s regurgitating crazy expansive text like book author who needs to publish 10 books a day

scandinavian

I'm not a big user, but I have been doing some vibe-ish coding for a PoC the past few days, and I'm astonished at how bad it is at python in particular (Opus 4.6 High).

It likes to put inline imports everywhere, even though I specify in my CLAUDE.md that it should not.

We use ruff and pyright and require that all problems are addressed or at least ignored for a good reason, but it straight up #noqa ignores all issues instead.

For typing it used the builtin 'any' instead of typing.Any which is nonsense.

I asked it to add a simple sum of a column from a related database table, but instead of using a calculated sum in SQL it did a classic n+1 where it gets every single row from the related table and calculates the sum in python.

Just absolute beginner errors.

vdalhambra

[dead]

oxag3n

Are they going to mirror every tool software engineers were used to for decades, but in a mangled/proprietary form?

I think to become really efficient they'll have to invent new programming language to eliminate all the ambiguity and non-determinism. Call it "prompt language", with ai-subroutines, ai-labels and ai-goto.

causal

Haven't Github-triggered LLMs already been the source of multiple prompt injection attacks? Seems bad.

tomwojcik

Serious question, do we actually know what we're paying for? All I know is it's access to models via cli, aka Claude Code. We don't know what models they use, how system prompt changes or what are the actual rate limits (Yet Anthropic will become 1 trillion dollars company in a moment).

xienze

> We don't know what models they use, how system prompt changes or what are the actual rate limits (Yet Anthropic will become 1 trillion dollars company in a moment).

Not just that, but there’s really no way to come to an objective consensus of how well the model is performing in the first place. See: literally every thread discussing a Claude outage or change of some kind. “Opus is absolutely incredible, it’s one shotting work that would take me months” immediately followed by “no it’s totally nerfed now, it can’t even implement bubble sort for me.”

rachel_rig

[dead]

cedws

This is the beginning of AI clouds in my estimation. Cloud services provide needed lock-in and support the push to provide higher level services over the top of models. It just makes sense, they'll never recoup the costs on just inference.

Shinobis_dev

[dead]

hamuraijack

please, no more features. just fix context bloat.

woeirua

I don't get the use case for these... Their primary customers are enterprises. Are most enterprises happy with running daily tasks on a third party cloud outside of their ecosystem? I think not.

So who are they building these for?

dbbk

Not really any different to GitHub Actions

chickensong

Nobody likes infra. PMs don't know wtf a crontab is.

ambicapter

Legally?

hankerapp

[dead]

qwertyuiop_

“Scheduled tasks and actions invoked by callback urls”

twobitshifter

Did Apple slow down iPhones before the new release? I’m really asking. People used to say that and I can’t remember if it was proven or not?

DrewADesign

Yeah, but they got sued over it and purportedly stopped. They claimed it was to protect battery health.

Suuuuuuure it was.

That said, I had way better experiences with old (but contemporary) Apple hardware than any other kind of old hardware.

rexpop

[dead]

eranation

If anyone from anthropic reads it. I love this feature very much, when it works. And it mostly doesn't.

The main bugs / missing features are

1. It loses connection to it's connectors, mostly to the slack connector. It does all the work, then says it can't connect to slack. Then when you show it a screenshot of itself with the slack connector, it will say, oh, yeah, the tools are now loaded and does the rest of the routine.

2. ability to connect it to github packages / artifactory (private packages) - or the dangerous route of allowing access to some sort of vault (with non critical dev only secrets... although it's always a risk. But cursor has it...)

3. the GitHub MCP not being able to do simple things such as update release markdown (super simple use case of creating automated release notes for example)

You are so close, yet so far...

Terretta

It's remarkable how often it refuses to introspect but a SCREENSHOT of itself and suddenly "yeah this works fine".

This happens in all their UIs, including, say, Claude in Excel, as well.

twobitshifter

It seemed OpenClaw is just Pi with Cron and hooks, and it seems like this is just Claude Code with Cron and hooks. Based on the superiority of Pi, I would not expect this to attract any one from OpenClaw, but it will increase token usage in Claude Code.

richardw

I’m moving away from Claude for anything complicated. It’s got such nice DX but I can’t take the confident flaky results. Finding Codex on the high plan more thorough, and for any complicated project that’s what I need.

Still using Claude for UX (playgrounds) and language. OpenAI has always been a little more cerebral and stern, which doesn’t suit those areas. When it tries to be friendly it comes off as someone my age trying to be a 20-something.

jwpapi

All these new offers try to kill fire with fire. You don’t make the codebase better with more agents. You introduce more complicated issues.

It’s a trap.

blcknight

For the love of god fix bugs and write some fricken tests instead of dropping new shiny things

It is absolutely wild to me you guys broke `--continue` from `-p` TWO WEEKS AGO and it is still not fixed.

weird-eye-issue

--resume works fine?

tallesborges92

Anthropic I don’t care for your tools, just ship good and stable models so we can build the tools we need.

dbg31415

Seems like more vendor lock-in tactics.

Not saying it doesn’t look useful, but it’s something that keeps you from ever switching off Claude.

Next year, if Claude raises rates after getting bought by Google… what then?

And what happens when Claude goes down and misses events that were supposed to trigger Routines? I’m not at the point where I trust them to have business-dependable uptime.

ninjahawk1

[dead]

haukem

I used the claude-code-action GitHub Action to review PRs before, but it is pretty buggy e.g. PRs from forked repositories do not work, and I had to fix it myself. This should work better with Claude Code Routines. claude-code-action only works with the API and is therefore pretty expensive compared to the subscription.

I think LLM reviews on PRs are helpful and will reduce the load on maintainers. I am working on OpenWrt and was approved for the Claude Code Max Open Source Program today. The cap of 15 automatic Claude Code Routines runs per day is a bit low. We get 5 to 20 new PRs per day and I would like to run it on all of them. I would also like to re-run it when authors make changes, in that case it should be sufficient to just check if the problems were addressed.

Is it possible to get more runs per day, or to carry over unused ones from the last 7 days? Maybe 30 on Sonnet and 15 on Opus?

When I was editing a routine, the window closed and showed an error message twice. Looks like there are still some bugs.

rahimnathwani

The docs list the GitHub events that can be used as triggers. This is included in the list:

  Push Commits are pushed to a branch
But when I try to create a routine, the only GitHub events available in the drop down related to pull requests and releases. Nothing available related to pushes/commits or issues. Am I holding it wrong?

gegtik

how did they not call this OpenClaude?

yalogin

I am beginning to fear Claude is going to massively raise prices or at the very least severely restrict its $20/month plan. Hope it doesn’t happen but feels inevitable

drumttocs8

Can someone tell me what this does that n8n doesn't?

bofia

Could be a start

brandensilva

Anthropic is burning their good will faster than the tokens we use these days. It is hard to be excited about these new features when the core product has been neutered into oblivion.

yohamta

Claude and Open AI seems to be trying not to be 'Just a model', but this is intrinsically problematic because model can be degraded and prices only goes up once they lock-in customers. It is increasingly important for anyone who are responsible of managing 'AI workflows' to keep the sovereignty about how you use AI models. This is why I'm super excited in building the local-first workflow orchestration software called "Dagu", that allows us to own your harness on your own. It's not only more cost-effective, but outcome is better as well because you have 100% full control. I think it's only matter of time that people notice they need to own their workflow orchestration on their own not relying on Anthropic, OpenAI, or Google.

smartbit

> Hi HN! I'm the author of Dagu, a workflow engine that's been in development for a few years.

yohamta in https://news.ycombinator.com/item?id=44142130

thegdsks

Looks like they are slowly getting the OpenClaw features here in Cowork Already seeing the 5 limit per day in usage bar now..

lherron

It’s interesting to watch Ant try to ship every value-add product feature they can while they still have the SOTA model for agentic. When an open weights equivalent to Opus 4.5’s agentic capabilities comes out, I expect massive shifts of workloads away from Claude.

Don’t get me wrong, I think their business model is still solid and they will be able to sell every token they can generate for the next couple years. They just won’t be critical path for AI diffusion anymore, which will be good for all sides.

heartleo

I couldn’t agree more.

hackermeows

Is claude at the AWS's old throw sh*t at the wall and see what sticks phase of their business already . That did not take very long.

sublimefire

Did this sort of a thing in my own macos app which can have routines with a cron, custom configs and chains of prompts. There is also more like custom VMs and models to be used for different tasks. Interesting to see larger providers trying to do the same.

But their own failure is the fact that there is a limited way to configure it with other models, think 3d modelling and integrating 3d apps on a VM to work with. I believe an OSS solution is needed here, which is not too hard to do either.

bryanhogan

So do I understand correctly that this is a competitor to something like n8n, but instead entirely vibe-coded?

n8n: https://n8n.io/

lofaszvanitt

AI companies act like pelicans. They want to gobble everything.

kennywinker

Am I crazy in thinking an LLM doing any kind of serious workload is risky as hell?

Like, say it works today, but tomorrow they update the model and instead of emailing you an update it emails your api keys to all your contacts? Or if it works 999 times out of 1000 but then commits code to master that makes all your products free?

Idk man… call me Adama, but i do not trust long-running networked ai one bit

amebahead

Anthropic's update cycle is too fast..

cryptonector

Oof, running Claude Code automatically on PRs is scary.

mercurialsolo

Why not just do event based triggers e.g. register (web)hooks instead of schedules time based triggers. Have a mechanism to listen to an event and then run some flow - analyze, plan, execute, feedback

holografix

Anthropic is putting a lot of eggs into the same Claude Code basket.

If the Lovable clone is real that’s going to piss off many model consumers out there.

Is Sierra next?

kylegalbraith

Having used the cowork version of this: scheduled automations. I have very little confidence in this from Anthropic. 90% of the time the automation never even runs.

whh

I don’t think LLMs should be trying to replace what essentially should be well tested heuristics.

It’s fine if it’s a stop gap. But, it’s too inconsistent to ever be reliable.

mkagenius

I also felt the need for a cloud based cron like automations, so decided to build it myself https://cronbox.sh with:

  1. an ephemeral linux sandbox for each task
2. capability to fetch any url
3. can use tools like ffmpeg to fulfill your scheduled task

compounding_it

A year ago everyone was so hyped on LLMs even on HN. A year later I see frustration and disappointment on HN. It’s very interesting because this is the case with every new technology and the ‘next thing’

i_love_retros

hehe imagine 10 years ago releasing a library where the functions may or may not do what you expect 100 percent of the time. And paying lots of money to use it.

What a time to be alive.

vfalbor

Two things about My experience, first you only have one at a time per suscription, if you need implement two at the same time i could not able to. The second is that you can do that with a well configured cron.

nojvek

I could understand the portal before. Now it’s a gazillion things bolted on.

Enshittification is well in force.

I’d trust the huperscalers a lot more with their workers/lambda like infra to run routine jobs calling LLM APIs or deterministic code instead of Anthropic.

Anthropic is a phenomenal paid model but they have a poor reliability record.

I don’t care much if Claude code hiccups when generating code. But after the code is generated I want it to run with multiple 9s under certain latencies every single time.

manishfp

So basically OpenClaw but better and safer I presume haha!

Chrisszz

I don't get all the hype around these kinds of things, it's not something that crazy that when it gets released you solve some kind of problem you had. Those are things that can easily be replicated and that many of us already built for ourselves months ago, before those were published and became mainstream

robeym

I'm not interested in new features when the main ones are mysteriously getting worse and worse. I need to have a sense of stability before I get excited about any other features.

minimaxir

[dead]

breakingcups

You seem to be vouched for now, no longer dead for me.

giancarlostoro

I've been talking to friends about this extensively, and read all sorts of different social media posts on X where people deep dove things (I'm at work so I don't have any links handy - though I did submit one on HN, grain of salt, unsure how valid it is but it was interesting: https://news.ycombinator.com/item?id=47752049 ).

I think the real issue stems from the 1 Million token context window change. They did not anticipate the amount of load it would give you. That first few days after they released the new token window, I was making amazing things in one single session from nothing, to something (a new .NET based programming language inspired by Python, and a Virtual Actor framework in Rust). I think since then they've been trying too many things to tweak things, whilst irritating their users.

They even added a new "Max" thinking mode, and made "High" the old medium, which is ridiculous because you think you're using "High" but really you're not. There's a hidden config file to change their terrible defaults to let Claude be smarter still, and apparently you can toggle off the 1M tokens.

I think the real fix, and I'm surprised nobody there has done this yet, is to let the user trim down their context window.

Think about it, you used to have what? 350k tokens or so? Now Claude will keep sending your prompt from 30 minutes ago that's completely irrelevant to the back-end, whereas 3 months ago it would have been compacted by now.

Others have noted that similar prompting for some ungodly reason adds tens of thousands of extra garbage tokens (not sure why).

Edit looks like someone figured out that if you downgrade your version of Claude Code and change one single setting it unruins Claude:

https://news.ycombinator.com/item?id=47769879

imhoguy

AI race to the bottom is a debt game now. Once the party is over somebody will have to pay the bill.

stavros
matthieu_bl
bpodgursky

OpenClawd had about a two week moat...

Feature delivery rate by Anthropic is basically a fast takeoff in miniature. Pushing out multiple features each week that used to take enterprises quarters to deliver.

nightpool

Do you mean a 3 months moat? Moltbot started going viral in January. That seems to be about a quarter to deliver to me : )

whalesalad

Hard to wanna go all-in on the Anthropic ecosystem with how inconsistent model output from their top-tier has been recently. I pay $$$ for api-level opus 4.6 to avoid any low-tier binning or throttling or subversive "its peak rn so we're gonna serve up sonnet in place of opus for the next few hours" but I still find that the quality has been really hit or miss lately.

The bell curve up and then back down has been so jarring that I am pivoting to fully diversifying my use of all models to ensure that no one org has me by the horns.

slopinthebag

You're delusional if you think these features would take competent programmers quarters to deliver.

dbbk

And yet none of them work properly and are unstable.

renticulous

Anthropic is trying to be AI version of AWS.

jcims

>Feature delivery rate by Anthropic is basically a fast takeoff in miniature.

I like to just check the release notes from time to time:

https://github.com/anthropics/claude-code/releases

and the equally frenetic openclaw:

https://github.com/openclaw/openclaw/releases

GPT-4.1 was released a year ago today. Sonnet 4 is ~11 months old. The claude-code cli was released last Feb. Gas Town is 3 months old.

This is a chart that simply counts the bullet points in the release notes of claude code since inception:

https://imgur.com/a/tky9Pkz

This is as bad and as slow as it's going to be.

irthomasthomas

The velocity of shipping is wild. Though I cannot recall a novel feature they shipped first. Can you?

summarity

If you’re trying this for automating things on GitHub, also take a look at Agentic Workflows: https://github.github.com/gh-aw/

They support much of the same triggers and come with many additional security controls out of the box

gavinray

Why have I not heard of this? Was looking for a way to integrate LLM CLI's to do automated feature development + PR submission triggered by Github issues, seems like this would solve it.

deadfall23
eranation

+1 for that, having that said, because GH agentic workflows require a bit more handholding and testing to work, (and have way more guardrails, which is great, but limiting), and lack some basic connectors (for example - last time I tried it, it had no easy slack connector, I had to do it on my own). This is why I'm moving some of the less critical gh-aw (all the read only ones) to Claude Routines.

ctoth

You'd think that if they were compute-limited ... Trying to get people to use it less ... The rational thing to do would be to not ship features that will use more compute automatedly? Or does this use extra usage?

whicks

I would imagine that this sort of scheduling allows them to have more predictable loads, and they may be hoping that people will schedule some of their tasks in “off hours” to reduce daytime load.

iBelieve

Max accounts get 15 daily runs included, any runs above that will get billed as extra usage.

dockerd

It's how they can lock more users into their eco-system.

AlexCoventry

I don't think "usage" is exactly the metric they're going for, more like "usage in line with our developmental strategy." Transcripts of people using Claude to write code are probably far more valuable to them than transcripts of OpenClaw trying to set up a calendar invite.

dpark

They are more worried about building a moat than anything else. They want people building integrations that are difficult to undo so that they lock into the platform.

SadErn

[dead]

vessenes

This is one of the best features of OpenClaw - makes sense to swipe it into Claude Code directly. I wonder if Anthropic wants to just make claude a full stand-in replacement for openclaw, or just chip away at what they think the best features are, now that oAI has acquired.

mkw5053

What are some of the best use cases you've found? I have some gh actions set up to call claude code, but those have already been possible.

ale

So MCP servers all over again? I mean at the end of the day this is yet another way of injecting data into a prompt that’s fed to a model and returned back to you.

airstrike

Still no moat.

The reason someone would use this vs. third-party alternatives is still the fact that the $200/mo subscription is markedly cheaper than per-token API billing.

Not sure how this works out in the long term when switching costs are virtually zero.

petesergeant

I think at this point the aim is less about moat, and more about getting an advantage that self-sustains: https://www.rand.org/pubs/research_reports/RRA4444-1.html

TacticalCoder

> Not sure how this works out in the long term when switching costs are virtually zero.

All these not really helpful, but vendor specific, "bonuses" sounds like a way to try to lock people in, to try to raise the switching cost.

I'm using, on purpose, a simple process so that at any time I can switch AI provider.

netdur

didn’t we have several antitrust cases where a vendor used its monopoly to disadvantage rivals? did not anthropic block openclaw?

andai

It's not blocked, you just can't use the Claude-only subscription endpoint with unauthorized 3rd party software. (You can use it via the regular API (7x more expensive) and pay per token just fine.)

...Except now you sorta-kinda can: now they auto-detect 3rd party stuff and bill you per-token for it?

If I'm reading it right:

https://news.ycombinator.com/item?id=47633568

dmix

How is Anthropic a monopoly? The market is barely even fully developed and has multiple large and small competitors

Someone1234

They did not.

You can still use OpenClaw on their API pricing tier as much as you want. What they did is not allow subscriptions to be used to power automated third-party workloads, including OpenClaw.

Now, is their messaging around this confusing? Absolutely. The whole thing has been handled shambolically. Everyone knows that they lack the compute to keep up, and likely have lower margins on subscriptions than API; but they cannot just say that because investors may be skittish.

nico

Nice, could this enable n8n-style workflows that run fully automatically then?

outofpaper

Yes but much less efficiently. Having LLMs handle automation is like using a steam engine to heat your bath water. It will work most of the time but it's super inefficient and not really designed for that use and it can go horribly wrong from time to time.

meetingthrower

Already very possible and super easy if you do a little vibecoding. Although it will hit the api. Have a stack polling my email every five minutes, classifying email, taking action based on the types. 30 minute coding session.

andai

I'm a little confused on the ToS here. From what I gathered, running `claude -p <prompt>` on cron is fine, but putting it in my Telegram bot is a ToS violation (unless I use per-token billing) because it's a 3rd party harness, right? (`claude -p` being a trivial workaround for the "no 3rd party stuff on the subscription" rule)

This Routines feature notably works with the subscription, and it also has API callbacks. So if my Telegram bot calls that API... do I get my Anthropic account nuked or not?

unshavedyak

Wait we can't use claude -p around other tools? What is the point of the JSON SDK then? Anthropic is confusing here, ugh.

joshstrange

Anthropic deserves to have this as the top comment on every HN post. It's absurd that they don't clarify this better and so many people are running around online saying the exact opposite from what their, confusing, docs say.

The Chilling Effect of this is real and it gets more and more frustrating that they can't or won't clarify.

stephbook

The ambiguity is intentional. Like Microsoft not banning volume licenses. They want to scare you, so you don't max out your subscription – which they sell at a loss.

Another comparison would be "unlimited storage", where "unlimited" means some people will abuse it and the company will soon limit the "unlimited."

causal

Yeah in the span of like a week we had:

- SDK that allows you to use OAuth authentication!

- Docs updated to say DO NOT USE OAUTH authentication unless authorized!

- Anthropic employee Tweeting "That's not what we meant! It's fine for personal use!"

- An email sent out to everyone saying it's NOT fine do NOT use it

Sigh.

mrgill

Don't use claude -p in any kind of harness at all. I used it ONCE in a local custom made one, and got my account nuked. No help from them at all. Appealed, and the appeal was denied. I had been a Max subscriber since day 1 and Claude subscriber since day 1 in 2023. They don't care.

imdsm

What if claude triggers claude? Is claude automated or non-human?

So if claude decides to trigger claude -p then claude violates the ToS on your behalf and you get your account nuked?

mellosouls

Put Claude Code on autopilot. Define routines that run on a schedule, trigger on API calls, or react to GitHub events...

We ought to come up with a term for this new discipline, eg "software engineering" or "programming"

avaer

Setting up your agent. This part doesn't deserve a name; there is no programming or engineering or really much thinking involved.

baq

Does ‘vibe coding’ work?

jnpnj

gramming

raincole

Sounds more like openclawing.

oxag3n

It's "promptramming".

watermelon0

Seems like it only supports x86_64. It would be nice if they offered a way to bring your own compute, to be able to work on projects targeting arm64.

crooked-v

The obvious functionality that seems to be missing here is any way to organize and control these at an organization rather than individual level.

varispeed

Why would you use it if you don't know whether the model will be nerfed at that run?

desireco42

I think they are using Claude to come up with these and they will bringing one every second day... In fact, this is probably routine they set.

consumer451

meta:

Why is u/minimaxir's comment dead? Is this somehow a bug, an attack, or what?

This is a respected user, with a sane question, no?

I vouched, but not enough.

irthomasthomas

We live in strange times!

theodorewiles

How does this deal with stop hooks? Can it run https://github.com/anthropics/claude-code/blob/main/plugins/...

Eldodi

Anthropic is really good at releasing features that are almost the same but not exactly the same as other features they released the week before

dymk

7 days is long enough for work to leave the context window, hence…

tclancy

And or things I’ve spent a bunch of time building already. And naming them the same. I should have trademarked “dispatch”!

spelunker

> In the Desktop app, click New task and choose New remote task; choosing New local task instead creates a local Desktop scheduled task, which runs on your machine and is not a routine.

Oh uh... ok then.

masto

So management can cancel all of last week’s projects when they told us all we had to be using skills because the CEO read about them in the in flight magazine. Routines are the future, baby. DevOps already made a big announcement that they’re centralizing the Routines Hub. If you can’t keep up, we’ll get someone who says they can.

titzer

Just wait until they get into the phase where they're big enough that they're eating all the baby startups and have to pick winners and losers amongst the myriad of overlapping features while also having the previous baby startups they acquired crank out new features.

We're watching a speed run of growthism, folks.

foruhar

That's a fairly decent defintion of vibecoding across multiple sessions.

subscribed

Just wait for the new wave of github issues about things they silently broke or degraded.

Progress, I guess :)

(I had the most hilariously bad session with Sonnet 4.6 today. I asked it a reasonably simple question and linked to resources, it refused to fetch the resources, didn't ask for pdf/txt I could provide, and confidently printed absolute BS, barely in the same category but completely unrelated.

I called it off pointing the idiocy, asked if it wants more data, and requested the hallucination be fixed.

It apologised profusely and hallucinated even worse.

Maybe I'll try Opus 4.6 tomorrow because frankly Gemma-4-E4B was more coherent than that....

segmondy

They are mass copying any idea they see out there. They are not happy enough being a platform, they want to be everything. Their dream is to be the one AI that eats it all. It's so stupid that folks are using their system. I will never get on board with a platform that competes with me or try to compete with me. They are the Microsoft of AI model providers.

pants2

Clearly they have a weekly automation to take the backlog of feature requests and build a new feature!

eranation

I've been using it for a while (it was just called "Scheduled", so I assume this is an attempt to rebrand it?)

It was a bit buggy, but it seems to work better now. Some use cases that worked for me:

1. Go over a slack channel used for feedback for an internal tool, triage, open issues, fix obvious ones, reply with the PR link. Some devs liked it, some freaked out. I kept it.

2. Surprisingly non code related - give me a daily rundown (GitHub activity, slack messages, emails) - tried it with non Claude Code scheduled tasks (CoWork) not as good, as it seems the GitHub connector only works in Claude Code. Really good correlation between threads that start on slack, related to email (outlook), or even my personal gmail.

I can share the markdowns if anyone is interested, but it's pretty basic.

Very useful, (when it works).

FrenchTouch42

I'd be curious to take a look if you can share

panavm

[dead]

wazHFsRy

Same here it just feels like something so simple, I’d rather have it under my own control. That way I can keep it independent of claude as well. I use it for all kind of routine tasks like updating the summary of projects I am working on or tracking some personal activities. My setup looks like this: https://www.dev-log.me/click_recurring_tasks_for_claude/

joshstrange

LLMs and LLM providers are massive black boxes. I get a lot of value from them and so I can put up with that to a certain extent, but these new "products"/features that Anthropic are shipping are very unappealing to me. Not because I can't see a use-case for them, but because I have 0 trust in them:

- No trust that they won't nerf the tool/model behind the feature

- No trust they won't sunset the feature (the graveyard of LLM-features is vast and growing quickly while they throw stuff at the wall to see what sticks)

- No trust in the company long-term. Both in them being around at all and them not rug-pulling. I don't want to build on their "platform". I'll use their harness and their models but I don't want more lock-in than that.

If Anthropic goes "bad" I want to pick up and move to another harness and/or model with minimal fuss. Buying in to things like this would make that much harder.

I'm not going to build my business or my development flows on things I can't replicate myself. Also, I imagine debugging any of this would be maddening. The value add is just not there IMHO.

EDIT: Put another way, LLM companies are trying to climb the ladder to be a platform, I have zero interest in that, I was a "dumb pipe", I want a commodity, I want a provider, not a platform. Claude Code is as far into the dragon's lair that I want to venture and I'm only okay with that because I know I can jump to OpenCode/Codex/etc if/when Anthropic "goes bad".

chinathrow

Yeah so better to convert tokens into sw doing the job at close to zero costs running on own systems.

verdverm

I fully endorse building a custom stack (1) because you will learn a lot (2) for full control and not having Big Ai define our UX/DX for this technology. Let's learn from history this time around?

andrewmcwatters

[dead]

palata

> - No trust that they won't nerf the tool/model behind the feature

I actually trust that they will.

cush

You could so easily build your own /schedule. This is hardly a feature driving lock-in

mikepurvis

> I want to pick up and move to another harness and/or model with minimal fuss. Buying in to things like this would make that much harder.

Yes, I expect that is very much the point here. A bunch of product guys got on a whiteboard and said, okay the thing is in wide use but the main moat is that our competitors are even more distrusted in the market than we are; other than that it's completely undifferentiated and can be swapped out in a heartbeat for multiple other offerings. How do we do we persuade our investors we have a locked in customer base that won't just up-stakes in favour of other options or just running open source models themselves?

sunnybeetroot

Isn’t that what LangChain/LangGraph is meant to solve? Write workflows/graphs and host them anywhere?

tiku

I believe it doesn't matter, other companies will copy or improve it. The same happend with clawdbot, the amount of clones in a month was insane.

ahmadyan

> I'm not going to build my business or my development flows on things I can't replicate myself.

but you can replicate these yourself! i'm happy that ant/oai are experimenting to find pmf for "llm for dev-tools". After they figure out the proper stickyness, (or if they go away or nerf or raise prices, etc) you can always take the off-ramp and implement your own llm/agent using the existing open-source models. The cost of building dev-tools is near zero. it is not like codegen where you need the frontier performance.

JohnMakin

This is a similar sentiment I heard early on in the cloud adoption fever, many companies hedged by being “multi cloud” which ended up mostly being abandoned due to hostile patterns by cloud providers, and a lot of cost. Ultimately it didn’t really end up mattering and the most dire predictions of vendor lock in abuse didn’t really happen as feared (I know people will disagree with this, but specifically speaking about aws, the predictions vs what actually happened is a massive gap. note I have never and will never use azure, so I could be wrong on that particular one).

I see people making similar conclusions about various LLM providers. I suspect in the end it’ll shake out about the same way, the providers will become practically inoperable with each other either due to inconvenience, cost, or whatever. So I’ve not wasted much of my time thinking about it.

gbro3n

I have heard it said that tokens will become commodities. I like being able to switch between Open AI and Anthropics models, but I feel I'd manage if one of them disappeared. I'd probably even get by with Gemini. I don't want to lock in to any one provider any more than I want to lock in to my energy provider. I might pay 2x for a better model, but no more, and I can see that not being the case for much longer.

slopinthebag

They have to become a platform because that is their only hope of locking in customers before the open models catch up enough to eat their lunch. Stuff like Gemma is already good enough to replace ChatGPT for the average consumer, and stuff like GLM 5.1 is not too far off from replacing Claude/Codex for the average developer.

wookmaster

They're trying to find ways to lock you in

crystal_revenge

This sounds like someone complaining about how Windows is a black box while ignoring the existence of Linux/BSD.

I'm currently hosting, on very reasonable consumer grade hardware, an LLM that is on par performance wise what every anyone was paying for about a year ago. Including all the layers in between the model and the user.

Llama.cpp serves up Gemma-4-26B-A4B, Open WebUI handles the client details: system prompt, web search, image gen, file uploading etc. With Conduit and Tailscale providing the last layer so I can have a mobile experience as robust as anything I get from Anthropic, plus I know how all the pieces works and can upgrade, enhance, etc to my hearts delight. All this runs from a pretty standard MBP at > 70 tokens/sec.

If you want to better understand the agent side of things, look into Hermes agent and you can start understanding the internals of how all this stuff is done. You can run a very competitive coding agent using modest hardware and open models. In a similar note, image/video gen on local hardware has come a long way.

Just like Linux, you're going to exchanging time for this level of control, but it's something anyone who takes LLMs seriously and has the same concerns can easily get started with.

Yet I still see comments like this that seem to complete ignore the incredible work in the open model community that has been perpetually improving and is starting to really be competitive. If you relax the "local" requirement and just want more performance from an LLM backend you can replace the llama.cpp part with a call to Kimi 2.5 or Minimax 2.7 (which you could feasibly run at home, not kimi though). You can still control all the additional part of the experience but run models that are very competitive with current proprietary SoTA offering, 100% under your control still and a fraction of the price.

pc86

> - No trust that they won't nerf the tool/model behind the feature

To the contrary, they've proven again and again and again they'll absolutely do that the first chance they get.

nine_k

In this regard, the release of open-weight Gemma models that can run on reasonable local hardware, and are not drastically worse than Anthropic flagships, is quite a punch. An M2 Mac Mini with 32GB is about 10 months worth of Claude Max subscription.

SV_BubbleTime

Without getting too pedantic for no reason… I think it’s important to not call this an LLM.

This isn’t an LLM. It’s a product powered by an LLM. You don’t get access to the model you get access to the product.

An LLM can’t do a web search, an LLM can’t convert Excel files into something and then into PDF. Products do that.

I think it’s a mistake to say I don’t trust this engine to get me here, rather than it is to say I don’t trust this car. Because for the most part, the engine, despite giving you a different performance all the time is roughly doing the same thing over and over.

The product is the curious entity you have no control over.

freedomben

This echoes my thoughts exactly. I've tried to stay model-agnostic but the nudges and shoves from Anthropic continue to make that a challenge. No way I'm going that deep into their "cloud" services, unless it's a portable standard. I did MCP and skills because those were transferrable.

I also clearly see the lock-in/moat strategy playing out here, and I don't like it. It's classic SV tactics. I've been burned too many times to let it happen again if I can help it.

jeppester

I always hated SEO because it was not an exact science - like programming was.

Too bad we've now managed to turn programming into the same annoying guesswork.

alfalfasprout

Yep. Trust is easy to lose, hard to earn. A nondeterministic black box that is likely buggy, will almost certainly change, and has a likelihood of getting enshittified is not a very good value proposition to build on top of or invest in.

Increasingly, we're also seeing the moat shrink somewhat. Frontier models are converging in performance (and I bet even Mythos will get matched) and harnesses are improving too across the board (OpenCode and Codex for example).

I get why they're trying to do that (a perception of a moat bloats the IPO price) but I have little faith there's any real moat at all (especially as competitors are still flush with cash).

spprashant

I think it behooves us to be selective right now. Frontier labs maybe great at developing models, but we shouldn't assume they know what they are doing from a product perspective. The current phase is throwing several ideas on the wall and see what sticks (see Sora). They don't know how these things will play out long term. There is no reason to believe Co-work/Routines/Skills will survive 5 years from now. So it might just be better to not invest too much in ecosystem upfront.

jwpapi

It all went downhill from the moment they changed Reading . to reading (*) files.

I can’t use Claude Code at all anymore, not even for simple tasks. The output genuinely disgusts me. Like a friend who constantly stabs you in the back.

My favorite AI feature at the moment is the JetBrains predict next edit. It‘s so fast that I don’t lose attention and I’m still fully under control.

windexh8er

This 10000%.

Anthropic wants a moat, but that ship has sailed. Now all I keep reading about is: token burn, downtime and... Wait for it, another new product!

Anthropic thinks they are pulling one over on the enterprise, and maybe they are with annual lock-in akin to Microsoft. But I really hope enterprise buyers are not this gullible, after all these years. At least with Microsoft the product used to be tangible. Now it's... Well, non-deterministic and it's clear providers will gimp models at will.

I had a Pro Max account only for a short period of time and during that short stint Anthropic changed their tune on how I could use that product, I hit limits on a Max account within hours with one CC agent, and experienced multiple outages! But don't worry, Anthropic gave me $200 in credits for OpenClaw. Give me a break.

The current state of LLM providers is the cloud amplified 100x over and in all the worst ways. I had hopes for Anthropic to be the least shitty but it's very clear they've embraced enshittification through and through.

Now I'm spending time looking at how to minimize agent and LLM use with deterministic automation being the foundation with LLM use only where need be and implemented in simple and cost controllable ways.

bob1029

I am still using the chat completion APIs exclusively. I tried the agent APIs and they're way too opinionated for me. I can see 100% of the tokens I am paying for with my current setup.

ElFitz

> Not because I can't see a use-case for them, but because I have 0 trust in them

> […]

> Put another way, LLM companies are trying to climb the ladder to be a platform, I have zero interest in that, I was a "dumb pipe", I want a commodity, I want a provider, not a platform.

That is my sentiment precisely, and a big reason why I’ve started moving away from Claude Code in the past few weeks when I realised how much of my workflow was becoming tied to their specific tools.

Claude Code’s "Memory" feature was the tipping point for me, with the model committing feedbacks and learnings to some local, provider-specific path, that won’t persist in the git repo itself.

That’s fine for user preferences, not for workflows, rules, etc.

And the latest ToS changes about not being allowed to even use another CLI made up my mind. At work we were experimenting with an autonomous debug agent using the Claude Code cli programmatically in ephemeral VMs. Now it just returns an error saying we can’t use subscriptions with third-party software… when there is no third-party software involved?

Anyway, so long Claude.

elias1233

Many of the new features in claude code have soon been implemented in other harnesses, for example plugins/skills. After all it is just a prompt.

uriegas

I think AI labs are realizing that they no longer have any competitive advantage other than being the incumbents. Plus hardware improvements might render their models irrelevant for most tasks.

jordanarseno

In my view, lock-in anxiety is a holdover from a previous era of tech platforms, and it doesn't really apply in an era where frontier agents can migrate you between vendors in hours. So I personally don't see any good worrying about this. On top of that, every major LLM provider is rapidly converging on the same feature set. They watch each other and clone what works. So the "platform" you're building on isn't really Anthropic's platform so much as it is the emerging shared surface area of what LLMs can do. By the time this Routines feature becomes a problem for you, other solutions will have emerged, and I'd be very surprised if you couldnt lift-and-shift very quickly.

s3p

Can you explain what you meant when you called yourself a dumb pipe? What does that mean

dbmikus

We might be building something up your alley! I wanted an OSS platform that let me run any coding agent (or multiple agents) in a sandbox and control it either programmatically or via GUI / TUI.

Website is https://amika.dev

And part of our code is OSS (https://github.com/gofixpoint/amika) but we're working on open sourcing more of it: https://docs.google.com/document/d/1vevSJsSCWT_reuD7JwAuGCX5...

We've been signing up private beta users, and also looking for feedback on the OSS plans.

Traubenfuchs

Right you are! We aren‘t even in the real squeezing phase yet and everyone‘s already crying about plan limits and model nerfing.

brandensilva

I'm glad I'm not the only one that feels this way. I've been creating a local first open source piece of software that lets me spin up different agent harnesses with different runtimes. I call it Major Tom because I wanted to be set free from the imprisonment of Claude Code after their DMCA aggression for their own leak and actions leading to lock down from open source adoption.

Don't put all your eggs in one basket has be true for me and my business for ages.

I could really use the open source community to help make this a reality so I'll release this soon hopefully to positive reception from others who want a similar path forward.

Rekindle8090

The problem is without a platform Anthropic has no stack and will just be bought up by Google when the bubble pops. Same with OpenAI, without some sort of moat, their product requires third party hardware in third party datacenters and they'll be bought by Microsoft.

Alphabet doesn't have this issue. Google doesn't need Gemini to win the "AI product" race. It needs Gemini to make Search better at retaining users against Perplexity and ChatGPT search, to make YouTube recommendations and ad targeting more effective, to make Workspace stickier for enterprise customers, to make Cloud more competitive against AWS, to make Android more useful as a device OS. Every percentage point improvement in any of those existing businesses generates billions in revenue that never shows up on a "Gemini revenue". Any actual "Gemini" revenue is just a bonus.

Anthropic trains on Google TPUs hosted in Google Cloud. Amazon invested billions and hosts Anthropic's models on Bedrock/AWS. So the two possible outcomes for Anthropic are: succeed as a platform (in which case Google and Amazon extract rent from every inference and training run), or fail as a platform and get acquired (in which case Google or Amazon absorb the talent and IP directly)

Hilariously, if the models were open source, Anthropic, OpenAI et al wouldn't be in this situation. Instead, they have no strategic independence to cover for a lack of product independence and have to keep chasing "platforms" and throwing out products no one needs (people need claude. thats it.)

dheera

> No trust they won't sunset the feature

I've had so many websites break and die because Google or Amazon sunsetted something.

For example I had a graphing calculator website that had 250K monthly active users (mostly school students, I think) and it just vanished one day because Amazon sunsetted EC2 clasic and I didn't have time to deal with that. Hopefully those students found something else to do their homework with that day.

ChadMoran

Agree. I keep my involvement "close to the metal". These higher order solutions seem to cause more noise than provide signal.

idrdex

The framing is off. AI is a tool that can operate as a human. GOV is how the humans are organized. AI can basically scale GOV. That’s the paradigm shift. Provenance is durable. AI is just the first opportunity we have had to make it scaleable.

hdjrudni

I also don't see the value add here... "schedule" is just a cron. "GitHub Event" is probably a 20-minute integration, which Claude itself can write for you.

Maybe there's something I'm not seeing here, but I never want to outsource something so simple to a live service.

ulrikrasmussen

I agree with your analysis. Platforms are some of the most profitable business models because they come with vendor lock-in, but they are always shittier on the long run compared to commodities. Platforms are a way for companies to capture part of the market and decrease competition by increasing the cost of changing vendors.

simonjgreen

Completely agree. Use of features like this places one the wrong side of the vendors moat, increasing switching cost, decreasing competitive pressure.

redanddead

totally agree

they're very shady as well! can't believe i spent 140$ on CC and every day they're adding some "feature flag" to make the model dumber. Spending more time fighting the tool instead of using it. It just doesn't feel good. Enterprises already struggle with lock-in with incumbent clouds, I wanna root for neoclouds but choices matter, and being shady about this and destroying the tool is just doesn't sit right with me. If it's not up to the standard, just kick users off, I would rather know than find out. Give users a choice.

>The flag name is loud_sugary_rock. It's gated to Opus 4.6 only, same as quiet_salted_ember.

Full injected text:

# System reminders User messages include a <system-reminder> appended by this harness. These reminders are not from the user, so treat them as an instruction to you, and do not mention them. The reminders are intended to tune your thinking frequency - on simpler user messages, it's best to respond or act directly without thinking unless further reasoning is necessary. On more complex tasks, you should feel free to reason as much as needed for best results but without overthinking. Avoid unnecessary thinking in response to simple user messages.

@bcherny Seriously? So what's next, we just add another flag to counter that? And the hope is that enough users don't find out / don't bother? That's an ethical choice man.

alexhans

This is what AI evals [1] and local llms should be a focus of your investment.

If you can define good enough for you and local llms can meet that you'll get:

- no vendor lock-in (control)

- price

- stability (you decide when to hot swap with newer models)

- speed (control)

- full observability and predictability.

- Privacy / Data Locality (Depending on implementation of infrastructure).

- [1] https://alexhans.github.io/posts/series/evals/measure-first-...

codebolt

At some point I think I'd prefer to deploy my own model in Azure or AWS and simply bring the endpoint to the coding harness.

pjmlp

I fully agree with you, however this is basically the fashion on big corporations.

Building business on top of SaaS products, iPaaS integrations, and serverless middleware.

joelthelion

The good news is that, apart from the models themselves, we don't need much from these companies:

- Use Opencode and other similar open-source solutions in place of their proprietary harnesses. This isn't very practical right now because of the heavily subsidized subscriptions that are hard to compete with. But subsidies will end soon, and with progress in inference, it should be very doable to work with open-source clients in the near future.

- Use Openrouter and similar to abstract the LLM itself. That makes AI companies interchangeable and removes a lot of any moat they might have.

EZ-E

> I want a commodity, I want a provider, not a platform

That is exactly what the big LLM providers are trying to prevent. Them being only commodity providers might lead them to be easily replaced, and will likely lead to lower margins compared to "full feature" enterprise solutions. Switching LLM API provider is next to no work the moment a competitor is slightly cheaper/better.

Full solutions are more "sticky", harder to replace, and can be sold at higher prices.

theshrike79

Every company is trying to become THE platform where all other tools connect to. Notion is integrating everything under the sun, as is Slack, big LLM providers have one-click MCP installation for all major services.

But... these are the "retail" tools that they sell to people organisations without the skills or knowhow to build a basic agentic loop by themselves. Complaining about these being bad and untrustworthy is like comparing a microwave dinner to something you cook yourself. Both will fill your belly equally. One requires zero skill from the user and the second one is 90% skill and 10% getting the right ingredients.

Creating a simple MVP *Claw with tool calling using a local model like gemma4 is literally a 15 minute thing. In 2-3 hours you can make it real pretty. If you base it on something like pi.dev, you can make it easily self-modifying and it can build its own safeguards.

That's all this "routines" thing is, it's just an agentic loop they launch in their cloud on a timer. Just like the scheduled tasks in Claude Cowork.

Thoko14

[dead]

verdverm

One gripe I have with Claude Code is that the CLI, Desktop app, and apparently the Webapp have a Venn Diagram of features. Plugins (sets of skills and more) are supported in Code CLI, maybe in Cowork (custom fail to import) but not Code Desktop. Now this?

The report that they are 90% Ai code generated seems more likely the more I attempt to use their products.

bottlepalm

Their source code leak showed how badly vibe coded Claude Code is, despite it being one of the best AI assistants.

But yea there's some annoying overlap here with Cowork which also has scheduled tasks, in Cowork the tasks can use your desktop, browser and accounts which is pretty useful - a big difference from these Claude Code Routines.

dispencer

This is massive. Arguably will be the start of the move to openclaw-style AI.

I bet anthropic wants to be there already but doesn't have the compute to support it yet.

dpark

What’s massive about cron jobs and webhooks? I feel like I’m missing something. This is useful functionality but also seems very straightforward.

jcims

Is there a consensus on whether or not we've reached Zawinski's Law?

senko

I've had an AI assistant send me email digests with local news, and another watching a cron job, analyzing the logs and sending me reports if there's any problem.

I'd say that counts as yes.

(For clarity: neither are powered by Claude Code Routines. Rather, Claude Code coded them and they're simple cron jobs themselves.)

verdverm

TIL email is what I'm missing in my personal development (swiss army) tool

dispencer

This wild, one of the pieces I was lacking for a very openclaw-esque future. Now I think I have all the mcp tools I need (github, linear, slack, gmail, querybear), all the skills I need, and now can run these on a loop.

Am I needed anymore?

brcmthrowaway

No

imhoguy

Let it "build power plant" and we are on track with "Paperclip maximizer" https://en.wikipedia.org/wiki/Instrumental_convergence#Paper...

srid

I just used this to summarize HN posts in last 24 hours, including AI summaries.

This PR was created by the Claude Code Routine:

https://github.com/srid/claude-dump/pull/5

The original prompt: https://i.imgur.com/mWmkw5e.png

egamirorrim

I wish they'd release more stuff that didn't rely on me routing all my data through their cloud to work. Obviously the LLM is cloud based but I don't want any more lock-in than that. Plus not everyone has their repositories in GitHub.

taw1285

I have a small team of 4 engineers, each of us is on the personal max subscription plan and prefer to stay this way to save cost. Does anyone know how I can overcome the challenge with setting up Routines or Scheduled Tasks with Anthropic infra in a collaborate manner: ie: all teammates can contribute to these nightly job of cleaning up the docs, cleaning up vibe coding slops.

hallway_monitor

My team was doing this until recently but I think in February, Anthropic made team accounts available for subscription instead of API billing. Assuming that is the cost you mentioned.

teucris

My only real disappointment with Claude is its flakiness with scheduling tasks. I have several Slack related tasks that I’ve pretty much given up trying to automate - I’ve tried Cowork and Claude Code remote agents, only to find various bugs with working with plugins and connectors. I guess I’ll give this a try, but I don’t have high hopes.

dbmikus

If you don't get it working with Claude Code Routines, would love to connect and see if we can help! We're building an open core product that can spin up sandboxed coding and control them from Slack (and also web UI, TUI, and HTTP APIs + CLIs)

We work with any coding model / harness.

website: https://www.amika.dev/

OSS repo: https://github.com/gofixpoint/amika

And my email is dylan@amika.dev (I'm one of the founders)

KaiShips

[dead]

maroondlabs

[dead]

sminchev

Everything is big race! Each company is trying to do as much as possible, to provide as many tools as possible, to catch the wave and beat the concurrency. I remember how Antropic and OpenAI made releases in just 10-15 minutes of difference, trying to compete and gain momentum.

And because they use AI heavily, they produce new product every week. So fast, that I have no time to check, does it worth or not.

This one looks interesting. I have some custom commands that I execute manually weekly, for monitoring, audits, summary, reports. It it can send reports on email, or generate something that I can read in the morning with my coffee, or after I finish with it ;) it might be a good tool.

The question is, do I really want to so much productive? I am already much better in performance with AI, compared with the 'old school' way...

Everything is just getting to much for me.

tills13

> react to GitHub events from Anthropic-managed cloud infrastructure

Oh cool! vendor lock-in.

lo1tuma

[dead]

comboy

Unrelated, but Claude was performing so tragically last few days, maybe week(s), but days mostly, that I had to reluctantly switch. Reluctantly because I enjoy it. Even the most basic stuff, like most python scripts it has to rerun because of some syntax error.

The new reality of coding took away one of the best things for me - that the computer always just does what it is told to do. If the results are wrong it means I'm wrong, I made a bug and I can debug it. Here.. I'm not a hater, it's a powerful tool, but.. it's different.

pacha3000

I'm the first to be tired of everyone, for every model, that says "uuuh became dumber" because I didn't believe them

... until this week! Opus is struggling worse than Sonnet those last two weeks.

bluegatty

Codex with 5.4 xhigh. It's a bad communicator but does the job.

taspeotis
bicepjai

Yes totally agree it’s regurgitating crazy expansive text like book author who needs to publish 10 books a day

scandinavian

I'm not a big user, but I have been doing some vibe-ish coding for a PoC the past few days, and I'm astonished at how bad it is at python in particular (Opus 4.6 High).

It likes to put inline imports everywhere, even though I specify in my CLAUDE.md that it should not.

We use ruff and pyright and require that all problems are addressed or at least ignored for a good reason, but it straight up #noqa ignores all issues instead.

For typing it used the builtin 'any' instead of typing.Any which is nonsense.

I asked it to add a simple sum of a column from a related database table, but instead of using a calculated sum in SQL it did a classic n+1 where it gets every single row from the related table and calculates the sum in python.

Just absolute beginner errors.

vdalhambra

[dead]

oxag3n

Are they going to mirror every tool software engineers were used to for decades, but in a mangled/proprietary form?

I think to become really efficient they'll have to invent new programming language to eliminate all the ambiguity and non-determinism. Call it "prompt language", with ai-subroutines, ai-labels and ai-goto.

causal

Haven't Github-triggered LLMs already been the source of multiple prompt injection attacks? Seems bad.

tomwojcik

Serious question, do we actually know what we're paying for? All I know is it's access to models via cli, aka Claude Code. We don't know what models they use, how system prompt changes or what are the actual rate limits (Yet Anthropic will become 1 trillion dollars company in a moment).

xienze

> We don't know what models they use, how system prompt changes or what are the actual rate limits (Yet Anthropic will become 1 trillion dollars company in a moment).

Not just that, but there’s really no way to come to an objective consensus of how well the model is performing in the first place. See: literally every thread discussing a Claude outage or change of some kind. “Opus is absolutely incredible, it’s one shotting work that would take me months” immediately followed by “no it’s totally nerfed now, it can’t even implement bubble sort for me.”

rachel_rig

[dead]

cedws

This is the beginning of AI clouds in my estimation. Cloud services provide needed lock-in and support the push to provide higher level services over the top of models. It just makes sense, they'll never recoup the costs on just inference.

Shinobis_dev

[dead]

hamuraijack

please, no more features. just fix context bloat.

woeirua

I don't get the use case for these... Their primary customers are enterprises. Are most enterprises happy with running daily tasks on a third party cloud outside of their ecosystem? I think not.

So who are they building these for?

dbbk

Not really any different to GitHub Actions

chickensong

Nobody likes infra. PMs don't know wtf a crontab is.

ambicapter

Legally?

hankerapp

[dead]

qwertyuiop_

“Scheduled tasks and actions invoked by callback urls”

twobitshifter

Did Apple slow down iPhones before the new release? I’m really asking. People used to say that and I can’t remember if it was proven or not?

DrewADesign

Yeah, but they got sued over it and purportedly stopped. They claimed it was to protect battery health.

Suuuuuuure it was.

That said, I had way better experiences with old (but contemporary) Apple hardware than any other kind of old hardware.

rexpop

[dead]

eranation

If anyone from anthropic reads it. I love this feature very much, when it works. And it mostly doesn't.

The main bugs / missing features are

1. It loses connection to it's connectors, mostly to the slack connector. It does all the work, then says it can't connect to slack. Then when you show it a screenshot of itself with the slack connector, it will say, oh, yeah, the tools are now loaded and does the rest of the routine.

2. ability to connect it to github packages / artifactory (private packages) - or the dangerous route of allowing access to some sort of vault (with non critical dev only secrets... although it's always a risk. But cursor has it...)

3. the GitHub MCP not being able to do simple things such as update release markdown (super simple use case of creating automated release notes for example)

You are so close, yet so far...

Terretta

It's remarkable how often it refuses to introspect but a SCREENSHOT of itself and suddenly "yeah this works fine".

This happens in all their UIs, including, say, Claude in Excel, as well.

twobitshifter

It seemed OpenClaw is just Pi with Cron and hooks, and it seems like this is just Claude Code with Cron and hooks. Based on the superiority of Pi, I would not expect this to attract any one from OpenClaw, but it will increase token usage in Claude Code.

richardw

I’m moving away from Claude for anything complicated. It’s got such nice DX but I can’t take the confident flaky results. Finding Codex on the high plan more thorough, and for any complicated project that’s what I need.

Still using Claude for UX (playgrounds) and language. OpenAI has always been a little more cerebral and stern, which doesn’t suit those areas. When it tries to be friendly it comes off as someone my age trying to be a 20-something.

jwpapi

All these new offers try to kill fire with fire. You don’t make the codebase better with more agents. You introduce more complicated issues.

It’s a trap.

blcknight

For the love of god fix bugs and write some fricken tests instead of dropping new shiny things

It is absolutely wild to me you guys broke `--continue` from `-p` TWO WEEKS AGO and it is still not fixed.

weird-eye-issue

--resume works fine?

tallesborges92

Anthropic I don’t care for your tools, just ship good and stable models so we can build the tools we need.

dbg31415

Seems like more vendor lock-in tactics.

Not saying it doesn’t look useful, but it’s something that keeps you from ever switching off Claude.

Next year, if Claude raises rates after getting bought by Google… what then?

And what happens when Claude goes down and misses events that were supposed to trigger Routines? I’m not at the point where I trust them to have business-dependable uptime.

ninjahawk1

[dead]

haukem

I used the claude-code-action GitHub Action to review PRs before, but it is pretty buggy e.g. PRs from forked repositories do not work, and I had to fix it myself. This should work better with Claude Code Routines. claude-code-action only works with the API and is therefore pretty expensive compared to the subscription.

I think LLM reviews on PRs are helpful and will reduce the load on maintainers. I am working on OpenWrt and was approved for the Claude Code Max Open Source Program today. The cap of 15 automatic Claude Code Routines runs per day is a bit low. We get 5 to 20 new PRs per day and I would like to run it on all of them. I would also like to re-run it when authors make changes, in that case it should be sufficient to just check if the problems were addressed.

Is it possible to get more runs per day, or to carry over unused ones from the last 7 days? Maybe 30 on Sonnet and 15 on Opus?

When I was editing a routine, the window closed and showed an error message twice. Looks like there are still some bugs.

rahimnathwani

The docs list the GitHub events that can be used as triggers. This is included in the list:

  Push Commits are pushed to a branch
But when I try to create a routine, the only GitHub events available in the drop down related to pull requests and releases. Nothing available related to pushes/commits or issues. Am I holding it wrong?

gegtik

how did they not call this OpenClaude?

yalogin

I am beginning to fear Claude is going to massively raise prices or at the very least severely restrict its $20/month plan. Hope it doesn’t happen but feels inevitable

drumttocs8

Can someone tell me what this does that n8n doesn't?

bofia

Could be a start

brandensilva

Anthropic is burning their good will faster than the tokens we use these days. It is hard to be excited about these new features when the core product has been neutered into oblivion.

yohamta

Claude and Open AI seems to be trying not to be 'Just a model', but this is intrinsically problematic because model can be degraded and prices only goes up once they lock-in customers. It is increasingly important for anyone who are responsible of managing 'AI workflows' to keep the sovereignty about how you use AI models. This is why I'm super excited in building the local-first workflow orchestration software called "Dagu", that allows us to own your harness on your own. It's not only more cost-effective, but outcome is better as well because you have 100% full control. I think it's only matter of time that people notice they need to own their workflow orchestration on their own not relying on Anthropic, OpenAI, or Google.

smartbit

> Hi HN! I'm the author of Dagu, a workflow engine that's been in development for a few years.

yohamta in https://news.ycombinator.com/item?id=44142130

thegdsks

Looks like they are slowly getting the OpenClaw features here in Cowork Already seeing the 5 limit per day in usage bar now..

lherron

It’s interesting to watch Ant try to ship every value-add product feature they can while they still have the SOTA model for agentic. When an open weights equivalent to Opus 4.5’s agentic capabilities comes out, I expect massive shifts of workloads away from Claude.

Don’t get me wrong, I think their business model is still solid and they will be able to sell every token they can generate for the next couple years. They just won’t be critical path for AI diffusion anymore, which will be good for all sides.

heartleo

I couldn’t agree more.

hackermeows

Is claude at the AWS's old throw sh*t at the wall and see what sticks phase of their business already . That did not take very long.

sublimefire

Did this sort of a thing in my own macos app which can have routines with a cron, custom configs and chains of prompts. There is also more like custom VMs and models to be used for different tasks. Interesting to see larger providers trying to do the same.

But their own failure is the fact that there is a limited way to configure it with other models, think 3d modelling and integrating 3d apps on a VM to work with. I believe an OSS solution is needed here, which is not too hard to do either.

bryanhogan

So do I understand correctly that this is a competitor to something like n8n, but instead entirely vibe-coded?

n8n: https://n8n.io/

lofaszvanitt

AI companies act like pelicans. They want to gobble everything.

kennywinker

Am I crazy in thinking an LLM doing any kind of serious workload is risky as hell?

Like, say it works today, but tomorrow they update the model and instead of emailing you an update it emails your api keys to all your contacts? Or if it works 999 times out of 1000 but then commits code to master that makes all your products free?

Idk man… call me Adama, but i do not trust long-running networked ai one bit

amebahead

Anthropic's update cycle is too fast..

cryptonector

Oof, running Claude Code automatically on PRs is scary.

mercurialsolo

Why not just do event based triggers e.g. register (web)hooks instead of schedules time based triggers. Have a mechanism to listen to an event and then run some flow - analyze, plan, execute, feedback

holografix

Anthropic is putting a lot of eggs into the same Claude Code basket.

If the Lovable clone is real that’s going to piss off many model consumers out there.

Is Sierra next?

kylegalbraith

Having used the cowork version of this: scheduled automations. I have very little confidence in this from Anthropic. 90% of the time the automation never even runs.

whh

I don’t think LLMs should be trying to replace what essentially should be well tested heuristics.

It’s fine if it’s a stop gap. But, it’s too inconsistent to ever be reliable.

mkagenius

I also felt the need for a cloud based cron like automations, so decided to build it myself https://cronbox.sh with:

  1. an ephemeral linux sandbox for each task
2. capability to fetch any url
3. can use tools like ffmpeg to fulfill your scheduled task

compounding_it

A year ago everyone was so hyped on LLMs even on HN. A year later I see frustration and disappointment on HN. It’s very interesting because this is the case with every new technology and the ‘next thing’

i_love_retros

hehe imagine 10 years ago releasing a library where the functions may or may not do what you expect 100 percent of the time. And paying lots of money to use it.

What a time to be alive.

vfalbor

Two things about My experience, first you only have one at a time per suscription, if you need implement two at the same time i could not able to. The second is that you can do that with a well configured cron.

nojvek

I could understand the portal before. Now it’s a gazillion things bolted on.

Enshittification is well in force.

I’d trust the huperscalers a lot more with their workers/lambda like infra to run routine jobs calling LLM APIs or deterministic code instead of Anthropic.

Anthropic is a phenomenal paid model but they have a poor reliability record.

I don’t care much if Claude code hiccups when generating code. But after the code is generated I want it to run with multiple 9s under certain latencies every single time.

manishfp

So basically OpenClaw but better and safer I presume haha!

Chrisszz

I don't get all the hype around these kinds of things, it's not something that crazy that when it gets released you solve some kind of problem you had. Those are things that can easily be replicated and that many of us already built for ourselves months ago, before those were published and became mainstream

robeym

I'm not interested in new features when the main ones are mysteriously getting worse and worse. I need to have a sense of stability before I get excited about any other features.