I recently retired my Wordpress blog and replaced it with a static-site generator. My requirements were straight-forward and I ended up having Codex build it for me.
It was the last thing using MySQL, PHP, and Wordpress on my site. 3 big things to not have to keep up-to-date and secured. I can check in markdown to my repo, it builds the site, and Nginx serves it. So fast, and secure.
That has me quite tempted. In general, I stay under the Plus limits, but I do watch my consumption. I could use `/fast` mode all of the time, with extra high reasoning, and use gpt-5.4-pro for especially complex tasks. It wasn't worth 10x the price to me before, but 5x is approachable.
That is disappointing. I find myself mostly using Codex these days. ChatGPT still for one off questions/prompts, but usually complex problems require context, like files and tools I already have on my system, and Codex is way better for that.
MTR is a must have tool for any time you need tracert. I personally have used WinMTR for many years. An invaluable tool when you have connectivity issues to diagnose.
It'd be fun to explore creating a Gemma 4 LLM API server app so you could use your iPhone's processing for agentic coding on a laptop. I don't know how useful it would be, but it'd be fun.
I want local models to succeed, but today the gap vs cloud models still seems continually too large. Even with a $2k GPU or a $4k MBP, the quality and speed tradeoff usually isn’t sensible.
Credit to Google for releasing Gemma 4, though. I’d love to see local models reach the point where a 32 GB machine can handle high quality agentic coding at a practical speed.
Fwiw, the real reason we don't have 100+ GB GPUs is because Nvidia likes to segment their markets. They could sell the consumer cards with 200gb gddr RAM on it, they just know that'd eat into their enterprise offering which is quiet literally all their profit margin (which I may add is gargantuan as of 2025)
If there is one thing that that agents/LLMs have highlighted, it is how much credit those test writers do deserve. Teams that were already following a TDD-style approach seem to be able to realize value from agents most easily because of their tests.
The tests are what enable: building a brand new JS runtime that works, rewriting a complex piece of code in a different language (e.g. Golang instead of TypeScript) that is more performant for that task, or even migrating off of an old stack (.NET WebForms) to something newer.
It is pretty incredible to me that in the pre-LLM/agent coding world, creating a new high-quality JS engine or browser seemed like it would likely never happen again. But now, any large company could build one if they wanted to. In a single digit number of months no less.
There's many JS implementations out there. Quality kind depends on what you need, and there's some engines more or less complete in which quirks are supported.
And for example, v8 doesn't make much sense in embedded contexts
There are definitely plenty of other JS engines, but they aren't always up to date on newer JS features. I'm pretty sure this is the 3rd JS engine to fully support the Temporal API (even JSC hasn't shipped it yet).
Some of the expanded scope that I’ve done almost for free is usually around UX polish and accessibility. I even completely redid the —help for a few CLI tools I have when I would never have invested over an hour on each before agents.
I agree that the efficiency and quality are very hard to measure. I’m extremely confident that when used well, agents are a huge gain for both though. When used poorly, it is just slop you can make really fast.
It was the last thing using MySQL, PHP, and Wordpress on my site. 3 big things to not have to keep up-to-date and secured. I can check in markdown to my repo, it builds the site, and Nginx serves it. So fast, and secure.
reply