I'm sorry to hear that you must have died right after we visited you in May of 2024 [1]. But I'm glad that you've figured things out and that you are now alive again! I imagine that's a much better state of affairs for you.
Here's to many more years of adventures and quilt making.
> With Airdrop you have trivially easy, "just works" sharing with people in proximity.
Funny enough, I encounter so many problems trying to share things via AirDrop with friends, family, and even my own Apple devices that I just tell everyone to install LocalSend and I find that things work better.
I’m not sure why that is, because AirDrop used to work pretty well for me. But it’s been an exercise in frustration more often than not for me.
(Obviously, LocalSend works only as long as everyone is on the same network.)
I've found it very often falls back to sending over your internet connection even if your cell reception sucks. No idea why. People on a previous HN thread talked about solutions
I’ve been on the Claude Code train for a while but decided to try Codex last week after they announced the $100 USD Pro plan.
I’ve been pretty happy with it! One thing I immediately like more than Claude is that Codex seems much more transparent about what it’s thinking and what it wants to do next. I find it much easier to interrupt or jump in the middle if things are going to wrong direction.
Claude Code has been slowly turning into this mysterious black box, wiping out terminal context any time it compacts a conversation (which I think is their hacky way of dealing with terminal flickering issues — which is still happening, 14 months later), going out of the way to hide thought output, and then of course the whole performance issues thing.
Excited to try 4.7 out, but man, Codex (as a harness at least) is a stark contrast to Claude Code.
Do this -- take your coworker's PRs that they've clearly written in Claude Code, and have Codex/GPT 5.4 review them.
Or have Codex review your own Claude Code work.
It then becomes clear just how "sloppy" CC is.
I wouldn't mind having Opus around in my back pocket to yeet out whole net new greenfield features. But I can't trust it to produce well-engineered things to my standards. Not that anybody should trust an LLM to that level, but there's matters of degree here.
I've been using Claude and Codex in tandem ($100 CC, $20 Codex), and have made heavy use of claude-co-commands [0] to make them talk. Outside of the last 1-2 weeks (which we now have confirmation YET AGAIN that Claude shits the fucking bed in the run-up to a new model release), I usually will put Claude on max + /plan to gin up a fever dream to implement. When the plan is presented, I tell it to /co-validate with Codex, which tends to fill in many implementation gaps. Claude then codes the amended plan and commits, then I have a Codex skill that reviews the commit for gaps, missed edge cases, incorrect implementation, missed optimizations, etc, and fix them. This had been working quite well up until the beginning of the month, Claude more or less got CTE, and after a week of that I swapped to $100 Codex, $20 CC plans. Now I'm using co-validation a lot less and just driving primarily via Codex. When Claude works, it provides some good collaborative insights and counter-points, but Codex at the very least is consistently predictable (for text-oriented, data-oriented stuff -- I don't use either for designing or implementing frontend / UI / etc).
You should not get dependent on one black box. Companies will exploit that dependency.
My version of this is having CC Pro, Cursor Pro, and OpenCode (with $10 to Codex/GLM 5.1) --> total $50. My work doesn't stop if one of these is having overloaded servers, etc. And it's definitely useful to have them cross-checking each other's plans and work.
This more or less mimics a flow that I had fairly good results from -- but I'm unwilling to pay for both right now unless I had a client or employer willing to foot the bill.
Claude Code as "author" and a $20 Codex as reviewer/planner/tester has worked for me to squeeze better value out of the CC plan. But with the new $100 codex plan, and with the way Anthropic seemed to nerf their own $100 plan, I'm not doing this anymore.
It cuts both ways. What I usually do these days is to let codex write code, then use claude code /simplify, have both codex and claude code review the PR, then finally manually review and fixup things myself. It's still ~2x faster than doing everything by myself.
100%. On days when I'm sleep deprived (once or twice a week), I fallback to this flow. On regular days, I tend to write more code the old school way and use things things for review.
> One thing I immediately like more than Claude is that Codex seems much more transparent about what it’s thinking and what it wants to do next. I find it much easier to interrupt or jump in the middle if things are going to wrong direction.
I've finally started experimenting recently with Claude's --dangerously-skip-permissions and Codex's --dangerously-bypass-approvals-and-sandbox through external sandboxing tools. (For now just nono¹, which I really like so far, and soon via containerization or virtual machines.)
When I am using Claude or Codex without external sandboxing tools and just using the TUI, I spend a lot of time approving individual commands. When I was working that way, I found Codex's tendency to stop and ask me whether/how it should proceed extremely annoying. I found myself shouting at my monitor, "Yes, duh, go do the thing!".
But when I run these tools without having them ask me for permission for individual commands or edits, I sometimes find Claude has run away from me a little and made the wrong changes or tried to debug something in a bone-headed way that I would have redirected with an interruption if it has stopped to ask me for permissions. I think maybe Codex's tendency to stop and check in may be more valuable if you're relying on sandboxing (external or built-in) so that you can avoid individual permissions prompts.
there is an official codex plugin for claude. I just have them do adversarial reviews/implementations. etc with each other. adds a bit of time to the workflow but once you have the permissions sorted it'll just engage codex when necessary
I’d be interested where you’re getting your data. SteamDB shows an accelerating trend of game releases over time, though comparing January 2026 to January 2025 directly shows a marginal gain [0].
This chart from a16z (scroll down to “App Store, Engage”) plots monthly iOS App Store releases each month and shows significant growth [1].
> After basically zero growth for the past three years, new app releases surged 60% yoy in December (and 24% on a trailing twelve month basis).
It’s completely anecdotal evidence but my own personal experience shows various sub-Reddit’s just flooded with AI assisted projects now, so much so that various pages have started to implement bans or limits of AI related posts (r/selfhosted just did this).
As far as _amazing software_ goes, that’s all a bit subjective. But there is definitely an increase happening.
I got the numbers swapped. Turns out there was an increase of about 40 games between last January and this. Which is exactly what you wouldn’t expect if the 5-10x claims are true.
Also the accelerating trend dates back to 2018 if you remove the early COVID dip. Which is exactly my point. You can look at the graph and there is no noticeable impact correlated to any major AI advancements.
The iOS data is interesting. But it’s an outlier because the Play Store and Steam show nothing similar. And the iOS App Store is weird because they’ve had numerous periods of negative growth follow by huge positive growth over the years. My guess is that it probably has more to do with all of the VC money flowing into AI startups and all the small teams following the hype building wrappers and post training existing models. If you look at a random sample of the iOS new apps that looks likely.
Seriously go to the App Store,
search AI and scroll until you get bored. There are literally thousands of AI API wrappers.
Oh my, this brings me back! One of my fondest gaming memories involves a massive Civilization 3 PBEM match between a number of Civilization fan sites, where we all had private forums and ran these virtual nations against each other. This was way back in 2002 or 2003!
I believe Civfanatics was in it (run by “Chieftess” if I recall), Apolyton (which I was a member of — elected in as Minister of Public Works and had to come up with a plan to clear our pesky jungles) and a number of other sites.
It was such an awesome time. Real diplomacy and trade negotiations between the fan sites while waiting to play our turns. Man, it was fun.
I was also there at Civfanatics watching from the sidelines. Fond memories indeed, and some of those same people laid the foundations for this project.
I didn’t do that stuff but I remember…was it Kryten? Making a multi unit graphic utility, I used it to make and publish some multi units. Fun times. CivFanatics was great.
Point of clarification: llama.cpp is MIT-licensed. Using it downstream (commercially or otherwise) is exactly what that license allows, so calling it a rip-off is misleading.
I've been using GLM 4.7 alongside Opus 4.5 and I can't believe how bad it is. Seriously.
I spent 20 minutes yesterday trying to get GLM 4.7 to understand that a simple modal on a web page (vanilla JS and HTML!) wasn't displaying when a certain button was clicked. I hooked it up to Chrome MCP in Open Code as well.
It constantly told me that it fixed the problem. In frustration, I opened Claude Code and just typed "Why won't the button with ID 'edit' work???!"
It fixed the problem in one shot. This isn't even a hard problem (and I could have just fixed it myself but I guess sunk cost fallacy).
I've used a bunch of the SOTA models (via my work's Windsurf subscription) for HTML/CSS/JS stuff over the past few months. Mind you, I am not a web developer, these are just internal and personal projects.
My experience is that all of the models seem to do a decent job of writing a whole application from scratch, up to a certain point of complexity. But as soon as you ask them for non-trivial modifications and bugfixes, they _usually_ go deep into rationalized rabbit holes into nowhere.
I burned through a lot of credits to try them all and Gemini tended to work the best for the things I was doing. But as always, YMMV.
Amazingly, just yesterday, I had Opus 4.5 crap itself extensively on a fairly simple problem -- it was trying to override a column with an aggregation function while also using it in a group-by without referring to the original column by its full qualified name prefixed with the table -- and in typical Claude fashion it assembled an entire abstraction layer to try and hide the problem under, before finally giving up, deleting the column, and smugly informing me I didn't need it anyway.
That evening, for kicks, I brought the problem to GLM 4.7 Flash (Flash!) and it one-shot the right solution.
It's not apples to apples, because when it comes down to it LLMs are statistical token extruders, and it's a lot easier to extrude the likely tokens from an isolated query than from a whole workspace that's already been messed up somewhat by said LLM. That, and data is not the plural of anecdote. But still, I'm easily amused, and this amused me. (I haven't otherwise pushed GLM 4.7 much and I don't have a strong opinion about about it.)
But seriously, given the consistent pattern of knitting ever larger carpets to sweep errors under that Claude seems to exhibit over and over instead of identifying and addressing root causes, I'm curious what the codebases of people who use it a lot look like.
This has been my consistent experience with every model prior to Opus 4.5, and every single open model I've given a go.
Hopefully we will get there in another 6 months when Opus is distilled into new open models, but I've always been shocked at some of the claims around open models, when I've been entirely unable to replicate them.
Hell, even Opus 4.5 shits the bed with semi-regularity on anything that's not completely greenfield for my usage, once I'm giving it tasks beyond some unseen complexity boundary.
I work on a React based web app in my Day Job and have genuinely enjoyed it.
That said, it always feels like so much boilerplate to get up and running for a greenfield project (and things like NextJS or even TanStack Start add a lot of things that might be overkill for a simple web app).
For some vibe coded side projects with Claude, I’ve been working with just using handlebars templates with Express and it has been pretty glorious!
I don’t think I’d recommend building a complex web app this way, but for some mild JS interactivity, form submission, etc, handlebars works.
Bonus: I find it much easier to get 100 across the board on Lighthouse scores this way.
For what it’s worth, I’ve used Vue and don’t like it (stuff like custom directives that are nonstandard html, not as explicit about event handling and data flow, etc).
I’ve seen a lot of buzz (particularly on HN) about Svelte but have lacked the motivation to try it.
I wish more people would criticise Svelte but most people just don't care because it's irrelevant. It's like complaining about Backbone or something, not worth the effort.
I wanted to like Vue but when I tried it I didn't. For writing ordinary business applications I wish React reified lists more than it does, but what I like about React is that I know how to draw absolutely anything with it, including 3-d worlds, see
I've tried both and disliked both. Already before I did I was wary of patronizing-feeling statements like these, though. The Vue community in particular, or at least the vocal pockets of it I've come across online, seem like a web dev counterpart to the Rust evangelism strike force.
> it always feels like so much boilerplate to get up and running for a greenfield project
This is why I love Bun + React. The setup is so easy I can do it off memory. I always hated boilerplate, templates etc. so this is a huge selling point for me.
Been going at it since 2003! It's a blog but links to all my open source work, and as of late, I talk a lot of various projects I work on and random rabbit holes I fall into.
I'm sorry to hear that you must have died right after we visited you in May of 2024 [1]. But I'm glad that you've figured things out and that you are now alive again! I imagine that's a much better state of affairs for you.
Here's to many more years of adventures and quilt making.
Cheers!
[1] https://daveschumaker.net/adventures-in-topology-the-cuckoos...
reply