Hacker Newsnew | past | comments | ask | show | jobs | submit | solidasparagus's commentslogin

What do you mean? Costs spiked with the introduction of the 1M context window I believe due to larger average cached input tokens, which dominate cost.

Nah, there's apparently a few caching bugs, one --resume and some noisy tool use. I have a little app that monitors and resets the context window at 70% usage based on 200k tokens and I'm about to run out of weekly allowance after just a couple days. Never happened before

I used them for repeated problems or workflows I encounter when running with the default. If I find myself needing to repeat myself about a certain thing a lot, I put it into claude.md. When that gets too big or I want to have detailed token-heavy instructions that are only occasionally needed, I create a skill.

I also import skills or groups of skills like Superpowers (https://github.com/obra/superpowers) when I want to try out someone else's approach to claude code for a while.


I don't really care if other people want to be on or off the AI train (no hate to the gp poster), but if you are on the train and you read the above comment, it's hard not to think that this person might be holding it wrong.

Using sonnet 4 or even just not knowing which model they are using is a sign of someone not really taking this tech all that seriously. More or less anyone who is seriously trying to adopt this technology knows they are using Opus 4.6 and probably even knows when they stopped using Opus 4. Also, the idea that you wouldn't review the code it generated is, perhaps not uncommon, but I think a minority opinion among people who are using the tools effectively. Also a rename falls squarely in the realm of operations that will reliably work in my experience.

This is why these conversations are so fruitless online - someone describes their experience with an anecdote that is (IMO) a fairly inaccurate representation of what the technology can do today. If this is their experience, I think it's very possible they are holding it wrong.

Again, I don't mean any hate towards the original poster, everyone can have their own approach to AI.


Yeah, I'm definitely guilty of not being motivated to use these tools. I find them annoying and boring. But my company's screaming that we should be using them, so I have been trying to find ways to integrate it into my work. As I mentioned, it's mostly not been going very well. I'm just using the tool the company put in front of me and told me to use, I don't know or really care what it is.


The whole point of "AI" in the first place is that it just vibes and doesn't need an instruction manual!

If "learn to hold it not wrong" is your message, then the AI bubble will be popping very soon.


How is that the point of AI. The point is that it can chug through things that would take humans hours in a matter of seconds. You still have to work with it. But it reduces huge tasks into very small ones


No, the point of AI is to fire your employees and replace them with "agents".

This implies that the managers managing your "agents" can be literal assclowns hired for pennies.


> Here’s what’s genuinely interesting.

That's my current AI detector smell.

> He discontinued the blood exchange after data showed “no benefits.” A suspicious person might note that a vampire would say exactly this after the media got too interested.

I don't think it's the media (clearly the younger generations are media friendly), it's probably pressure from the older vamps.


I felt the same way and came to the comments to see if anyone else smelled it. It's either AI-assisted writing or people are genuinely starting to write like how ChatGPT sounds.

First, the structure of this satirical post is headings and bullet points. Fine, whatever, a lot of people write this way.

Then there's the exhausting litany of super short sentence fragments.

> He published this. Openly. In a book. As a priest.

This is how airport novels and LinkedIn "thought leadership" clickbait is written, so ok, fine, I'll let it pass.

Then I started to notice a lot of: "It's not X. It's Y" or "this isn't just A. It's B."

> Feeding isn’t nutrition. It’s dialysis.

Before LLMs, people weren't writing this way. At the risk of sounding like a curmudgeon: it's insulting to read, like the reader is a 5-year-old.

When several of these smells pile up, I close the tab immediately and try to forget about it. This one was so egregious that I had to read the whole thing and then come to the comments to rant a bit.


> it's insulting to read, like the reader is a 5-year-old.

It's not ELI5. It's ELY5.


The author has a bunch of AI stuff in their bio, so I assume this is partially or fully generated unless otherwise disclaimed.

But hey, maybe someone can get an AI to read it.


Yeah, that does sound pretty AI-ish / marketing-bloggy. It’s not wrong, but it has a few classic “AI vibes”. If you want, I can........oh no!!!!!!

NO CARRIER


Oh my gosh I’m so sorry about that. Let me correct my earlier post and assert with certainty that this post was not at all written by an AI.


> This is a critical narrative shift.

also this, having done a number of ai conspiracy for funsies, that's always the mid point

however I don't mind ai slop, if it's creative and well thought out (and editorialized, as this seem to be)


It had me at "The Twist".


> You know what else is far-seeing? A creature that has been alive for centuries.

Well, hello there!


I'm not sure.

The cost of replacement-level software drops a lot with agentic coding. And maintenance tasks are similarly much smaller time syncs. When you combine that with the long-standing benefits of inhouse software (customizable to your exact problem, tweakable, often cleaner code because the feature-set can be a lot smaller), I think a lot of previously obvious dependencies become viable to write in house.

It's going to vary a lot by the dependency and scope - obvious owning your own react is a lot different than owning your own leftpad, but to me it feels like there's no way that agentic coding doesn't shift the calculus somewhat. Particularly when agentic coding make a lot of nice-to-have mini-features trivial to add so the developer experience gap between a maintained library and a homegrown solution is smaller than it used to be.


I pay $200 a month and don't get any included access to this? Ridiculous


That is the point. They raised prices and want you to pay more for quicker answers.

No different to paying a knowledge worker but this time, you are paying more to get them to respond quicker to your questions.


The API price is 6x that of normal Opus, so look forward to a new $1200/mo subscription that gives you the same amount of usage if you need the extra speed.


I always wondered this, is this true/does the math come out to be really that bad? 6x?

Is the writing on the wall for $100-$200/mo users that, it's basically known-subsidized for now and $400/mo+ is coming sooner than we think?

Are they getting us all hooked and then going to raise it in the future, or will inference prices go down to offset?


The writing has been on the wall since day 1. They wouldn't be marketing a subscription being sold at a loss as hard as they are if the intention wasn't to lock you in and then increase the price later.

What I expect to happen is that they'll slowly decrease the usage limits on the existing subscriptions over time, and introduce new, more expensive subscription tiers with more usage. There's a reason why AI subscriptions generally don't tell you exactly what the limits are, they're intended to be "flexible" to allow for this.


Well, you can burn your $50 bonus on it


..But it says "Available to all Claude Code users on subscription plans (Pro/Max/Team/Enterprise) and Claude Console."

Is this wrong?


It's explicitly called out as excluded in the blue info bubble they have there.

> Fast mode usage is billed directly to extra usage, even if you have remaining usage on your plan. This means fast mode tokens do not count against your plan’s included usage and are charged at the fast mode rate from the first token.

https://code.claude.com/docs/en/fast-mode#requirements


I think this is just worded in a misleading way. It’s available to all users, but it’s not included as part of the plan.


People want convenience more than they want security. No one wants permission grants to go away in minutes or hours. Every time the agent is stopped by permissions grant check, the average user experience is a little worse.


I agree that UX is the hard part. The point isn’t “pop-up permission dialog every minute.” It’s “remove standing power.” You can make short lived authority feel smoother with scoped permits, pre-approved workflows, clear revocation semantics, and defaults that renew narrowly. The non-negotiable part is that authority can be pulled instantly and cannot silently widen. Convenience matters, but “always-on admin” is convenience paid for with failure.


Why do you think python without access to the library ecosystem is a good approach? I think you will end up with small tool call subgraphs (i.e. more round trips) or having to generate substantially more utility code.


The first sentence of the blog post has a link to the product he is building - https://www.monarch.com/


with the AI


Nice post! Can you share a bit more about what variety of tasks you've used agents for? Agents can mean so many different things depending on who you're talking to. A lot of the examples seem like read-only/analysis tasks. Did you also work on tasks where agent took actions and changed state? If yes, did you find any differences in the patterns that worked for those agents?


Sure! So there are both read-only and write-only agents that I'm working on. Basically there's a main agent (main LLM) that is responsible for the overall flow (currently testing GPT-5 Mini for this) and then there are the sub-agents, like I mentioned, that are defined as tools.

Hopefully this isn't against the terms here, but I posted a screenshot here of how I'm trying to build this into the changelog editor to allow users to basically go:

https://x.com/ImSh4yy/status/1951012330487079342

1. What tickets did we recently close? 2. Nice, write a changelog entry for that. 3. Add me as author, tags, and title. 4. Schedule this changelog for monday morning.

Of course, this sounds very trivial on the surface, but it starts to get more complex when you think about how to do find and replace in the text, how to fetch tickets and analyze them, how to write the changelog entry, etc.

Hope this helps.


Neat idea!


Thank you :)


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: