Hacker Newsnew | past | comments | ask | show | jobs | submit | systemerror's commentslogin

this is incredible, I have been thinking about this exact project for quite some time but a little wary of which approach I might take. Thanks for posting this.

what workflows do you implement in Nanoclaw that wouldn't be straightforward to build in Claude?


Straightforward is ambiguous. To replicate NanoClaw would probably only take about a day of work and testing and refining in Claude Code, but that's a day I didn't have to spend to get NanoClaw.


yes but then what do you use nanoclaw for, that's its a better fit for than claude code.


Also interested if this supports strudel REPL or tidalcycles. This would be a really awesome device to use for livecoding sets if it does.


That’s a tall order for a 240 MHz CPU.


Why? That's a decent amount of power for running one thing. It's also dual core! You can emulate early consoles on this microcontroller. Someone made a 3D game for the similar (but less performant) RP2040 running at 250 MHz -> https://github.com/bernhardstrobl/Pico3D


Specifically because both Strudel and Tidal Cycles seem to struggle on moderately complex compositions on my ThinkPad X270.


Oh interesting. Thanks!


Don't give money to amazon that is better spent on an amazingly efficient postal service. Amazon is subsidized by imaginary money until they put all their competition out of business(including USPS).


My honest question is: If you pull shenanigans like this, isn't it actually making Amazon burn through said imaginary money, thus hastening its demise? The cost of delivering a potato has to be on the order of at least a couple dollars.


I don't think Amazon is losing money. It's really just that efficient.

E.g. an Amazon van rolls through my street multiple times a day. What is the marginal cost of them stopping at my house and dropping off a potato?


At your house it might be fractions of a cent.

At my house, it's a 140 mile round trip between the fulfillment center ("are you feeling fulfilled yet?") and the drop off location.

OTOH, there's likely more of "you" than there are of "me" ...


Assuming it's the US we're talking about, the federal minimum wage is $7.25, which means that if every worker involved is paid at the minimum wage, you incur a cent of labour costs every 4.97 person-seconds. AFAICT, most Amazon workers are paid substantially higher than the federal minimum wage. And that's just labour costs.

While Amazon is efficient, "fractions of a cent" is probably the wrong order of magnitude for even the most efficient order.


That's not even mentioning their additional overheads, like the cost of fuel for their idling van as they drop off your potato.


You might be 140 miles round trip to the nearest fulfilment centre, but you're almost certainly closer to your nearest neighbours who regularly buy stuff from Amazon, so the van is probably coming pretty close to you any way.


Amazon will close your account before you can impact their bottom line.


I think they let you (not YOU necessarily, but the proverbial you.) get away with stuff because they know your habits and you probably make more money for them than you realize.

I can almost guarantee that everyone mentioned in that blog post is a habitual Amazon user. They're all renewing Prime each year at full price and making a ton of regular purchases. The family has even turned on the FOMO by making Prime a family social network with social pressure to stay. I see it as a self-own, personally.

Edit: I'm taking part of this to the root of the thread


Can you explain? Amazon is wildly profitable, and while AWS is far higher margin than their retail businesses, everything I can find suggests their retail segment also has a healthy operating margin.


If you put all of the money Amazon as a whole has taken since it was founded in 1994 in a stack on the left, and all of the money Amazon as a whole has spent since then in a stack on the right, the stack on the left is slightly larger, but this has only been true for a couple of years now.

It's the difference in 1990s billionaires and 2020s billionaires. Bill Gates was so rich because he owned a lot of Microsoft shares and received profits from those shares as dividends. Jeff Bezos is so rich because he owns a lot of Amazon shares and people keep being willing to pay more and more for those shares so his notional net worth increases (AMZN has never paid a dividend).


None of which supports the argument of the person I replied to that what you buy from them today is somehow "subsidised by imaginary money"


> his notional net worth increases (AMZN has never paid a dividend).

But that’s exactly the loophole: you can borrow for very cheap against this notional equity without incurring a cent in taxes (since divodends are never paid out)


It is worth pointing out that many executives cash out the old fashioned way: by selling stock. That includes Bezos.

https://www.secform4.com/insider-trading/1043298.htm


Can you share numbers? What are Amazon’s margins?


Across the board they're about 11%, I believe, though retail seems to be about 5% despite being the bulk of revenue. AWS has far higher margins.


I hate USPS, and will not be doing anything to benefit them until they offer a way to limit my deliveries to once a month, and opt out of anything that has "or current resident"

At the very least they should charge more for bulk mail, not give out discounts.


In Canada, you can place a red dot (or write no unsolicited mail) on your mailbox and they will withhold delivering anything not directly addressed to you.

I was shocked when I moved to SF and found out there was no way to opt out of unaddressed mail (or "current resident").


In Finland AFAICT there's no bulk postal rate. Instead, paper spam is delivered thru mail slots by private services that hit all the buildings in the neighborhood and drop collections of paper spam. So, many people post a note on their door opting out from this stuff. (Ei mainoksia = No ads.) It must be saving absolutely huge amounts of paper.


Fun fact: in the U.S. it’s illegal to put anything but mail (delivered by USPS) in a mailbox or mail slot. USPS wants a monopoly on that paper spam.


But when you have to push a rotten fish thru someone's mail slot, the law be damned.


Unfortunately bulk mail is the only thing paying the bills. That and being a last mile delivery service for Amazon.


Which is a totally valid reason to hate USPS.

The USPS is a government-run spam delivery service that there is no way to opt out of. Those of us who do banking and other administrative tasks online would be better off if the government shut it down completely, or better yet subsidized it slightly so it doesn't have to deliver spam to survive.

But as it is, I don't see any good reason to have any more respect for USPS than I do for any other spammer.


Doesn't seem like USPS is the spammer. They're Gmail. People send spam and USPS/Gmail delivers it.


No, it’s completely different. Gmail actively tries to prevent spam. If they catch you sending it, you will be banned, and they let individual users block whoever they want. A huge part of their product is automated spam filtering.

On the other hand, spam delivery is the business model of USPS. They actively and intentionally market and sell their services to spammers, and not surprisingly, give normal users no way to opt out.


There are mail forwarding services[0] that let you automatically filter out junk mail. Yes, they cost money, but at least you can accomplish your goal of opting out (or in) from receiving postal mail from certain senders.

  [0] E.g. https://www.usglobalmail.com/virtual-mailbox/


You will still get mail to your address though...


If your goal is to receive zero mail at your address, simply have all your email forwarded to this service and configure it to opt-out of all notifications.

Alternately, you can forward your mail to an entirely different address. (This is just a Gedankenexperiment; please don't spam strangers by forwarding your mail to them.)


A lot of mail is marked as do not forward


Such mail is returned to the sender, which means you will never receive it, so problem solved.


It will still land in your mailbox at home, how would you never receive it?


And junk/bulk rate mail is never forwarded.


Yes, exactly. I wish the post office were subsidized and acted in the interests of the public. But it is not, and does not.


I cannot tell if this is satire, but if it is, bravo.


Why would this be satire? I'm guessing you don't pay that much attention to humans, or you have a very curated social group you're around.


I think the argument that could be made here is that, given evolution, one could say "feature, not a bug." in the context of humans.


>given evolution, one could say "feature, not a bug."

The issue with evolution is huge portions of it just happen to exist and not kill the host before they breed. It could be a massive bug that if corrected could cause the host to breed and spread their genes far further, but evolution itself can't reach there.


You ask why this would be satire?

Well let's take a look at this:

>The best thing about a good deep conversation is when the other person gets you: you explain a complicated situation you find yourself in, and find some resonance in their replies.

>That, at least, is what happens when chatting with the recent large models.

The first sentence says a good conversation is between two people. The author then pulls the rug out and says "Psych. A good conversation is when I use LLMs."

The author points out humans have decades of memories but is surprised that when they tell someone they are wrong they don't immediately agree and sycophantically mirror the author's point of view.

The author thinks it's weird they don't know when the next eclipse is. They should know this info intuitively.

The author claims humans have a habit of being wrong even in issues of religion but models have no such flaw. If only humans embraced evidence based religious opinions like LLMs.

The author wonders why they bothered writing this article instead of asking ChatGPT to write it.

Did you ask an LLM if this is satire?

I did and Opus said it wasn't satire.

This was clearly a hallucination so I informed it it was incorrect and it changed it's opinion to agree with me so clearly I known what I'm talking about.

I'll spare you the entire output but among other things after I corrected it it said:

The "repeating the same mistakes" section is even better once you see it. The complaint is essentially: "I told someone they were wrong, and they didn't immediately capitulate. Surely pointing out their error should rewire their brain instantly?" The author presents this as a human deficiency rather than recognizing that disagreement isn't a bug.


I really can't tell to be honest, so if it's satire it's very good. If it's not then I don't really know what it's saying, humans are bad at stuff too? Many humans are not well educated and are not great conversationalists? Does this somehow make LLMs better in our perspective somehow?


If it were satire, what do you think it would be satirizing?

> I don't really know what it's saying

It's saying that complaints about deficiencies in LLMs, about a fundamental lack of LLM intelligence, about how LLMs are just statistical machines and not really thinking, about how LLMs are incapable of learning from past experiences, about how LLMs lack any coherent epistemology ignore how very deficient humans are in many same exact ways.

> Does this somehow make LLMs better in our perspective somehow?

Better is a relative measure not an absolute one, so possibly, because views of LLMs are inherently formed in relation to views of the human brains they're modeling.


>about how LLMs are just statistical machines and not really thinking,

I don't think the article said anything about statistics?

This seems to be a sort of Rorchasch test but looking at it again:

>This does not bode well for my interest in meeting new people

It really does seem to me the article is making fun of people who think this sort of article is on point.

There's a genre of satire where the joke is that it makes you ask "Who the heck is the sort of person who would write this?"

It could fit in that genre but of course I could be wrong.


> I don't think the article said anything about statistics?

I don't think I said or implied that it did. It's merely one of the many positions that people commonly (and defensively) take for why LLMs aren't and/or can't be intelligent like humans, ignoring that humans exhibit exactly the same patterns.


> If it were satire, what do you think it would be satirizing?

Think of the most terminally online drama you've ever witnessed: the hysterics people work themselves into over what (to outside observers) seems utterly inane and forgettable, the multi-page Tumblr or 4chan posts that become the sacred texts of the "discourse", and the outsized importance people ascribe to it, as if some meme, album cover, or Qanon drop is the modern incantation of the shot heard around the world.

The people wrapped up in this stuff tend to self-select into their own communities because if you're not involved with or amenable to caring about it, why should they spend time talking to someone who will just nod, go "huh, that's wild", and proceed to steer the conversation elsewhere? In their eyes, you may even be a weirdo for not caring about this stuff.

So when I read:

> I’ve got a lot of interests and on any given day, I may be excited to discuss various topics, from kernels to music to cultures and religions. I know I can put together a prompt to give any of today’s leading models and am essentially guaranteed a fresh perspective on the topic of interest. But let me pose the same prompt to people and more often then not the reply will be a polite nod accompanied by clear signs of their thinking something else entirely, or maybe just a summary of the prompt itself, or vague general statements about how things should be. In fact, so rare it is to find someone who knows what I mean that it feels like a magic moment. With the proliferation of genuinely good models—well educated, as it were—finding a conversational partner with a good foundation of shared knowledge has become trivial with AI. This does not bode well for my interest in meeting new people.

I'm imagining the more academic equivalent of someone who got wrapped up in Tiktok drama or Q nuttery but couldn't find a community of kindred souls and, frustrated with the perceived intellectual mediocrity surrounding themself, has embraced LLMs for connection instead. And that's just hilarious. If Silicon Valley was still being produced, I'm sure this would have been made into an episode at some point.

The bits about not generalizing and engaging in fallacious reasoning are also quite amusing since, while yes, the average person likely would benefit from taking (and paying attention in) a couple introductory philosophy classes, expecting all humans to behave logically and introspectively is fantastical thinking.


> expecting all humans to behave logically and introspectively is fantastical thinking

Yes, that is exactly the point of OP's post, that humans are on average quite bad at behaving logically and introspectively and exhibit the very same behaviors that we righteously fault AI for doing. And then OP provides a list of faulty human behaviors that are the same faulty behaviors people give as demonstrating that AI lacks true intelligence.

Meanwhile, AI continues to improve and the human species does not.

And the conclusion is that the fact that the rise of AI has made human faulty behaviors more apparent may creepingly tear at the social fabric.

Read the first paragraph again. It sets the framing through which the rest of the post is understood (as first paragraphs tend to do).

I find this exchange to be a funny example of the truth of OP's list, where the part which sticks with you is some finer detail of one of the examples while the thesis statement itself, the very explanation of the overarching point of the post, seems to have fallen outside of the context window.


The point is that if you have benchmarks for intelligence, which humans would also fail, then you have to concede that either humans are not intelligent, or that the benchmarks are too strict, or aren't a measure for intelligence at all.


The thing is, LLMs would fail that test every time, but humans would pass it most of the time (hopefully). Just because humans are fallible doesn't make LLMs intelligent.

We really haven't got a grip on what intelligence actually is, but it seems that humans and LLMs aren't really in the same ballpark, or even the same league.


>haven't got a grip on what intelligence actually is

Because intelligence isn't a thing, it's a bunch of different things that some intelligent things have more or less (or none of).

This is why measures of intelligence always fail because we try to binary it which doesn't work. Intelligence is spikey. Intelligence scales from very small and dumb to very smart. But even the things that are very smart on a lot of things still do very dumb things. We also measure human intelligence as a function of all humans and LLM intelligence on a particular model.

So yea, this is why nothing seems to make sense.


Why does air-gapped environment require rolling your own CI/CD solution? Plenty of examples of air-gapped Jenkins and/or Argo Workflows. Was this just an educational exercise?


Jenkins sucks but is insanely reliable

Argo Workflows does not live up to what they advertise, it is so much more complex to setup and then build workflows for. Helm + Argo is pain (both use the same template delimiters...)


Jenkins, like many tools with extreme flexibility, sucks as much as you make it suck. You can pretty easily turn Jenkins into a monstrosity that makes everyone afraid to ever try to update it. On the other hand, you can also have a pretty solid setup that is easy to work on. The trouble is that the tool itself doesn't guide you much to the good path, so unless you've seen a pleasant Jenkins instance before you're likely to have a worse time than necessary.


Are you sure, because last time I used Jenkins it actively sucked. The interface was a total mess and it doesn't surface results in any useful way.


What particular issues do you have with it? My company uses it at scale (dozens of different instances, hundreds of workers, thousands of pipelines) to support thousands of applications and we are reasonably happy with it. DSL is incredibly helpful at scale. IAC is incredibly helpful at scale. It requires a good amount of upkeep, but all things underpinning large amounts of infrastructure require a good amount of upkeep.


We've minimized our usage of the DSL, there is no way for devs to debug it without pushing commits, and it means you have to implement much of your CI logic twice (once for local dev, once for ci system).

IMO, ci should be running the same commands humans would run (or could if it is production settings). Thus our Jenkins pipelines became a bunch of DSL boilerplate wrapped around make commands. The other nice thing about this is that it prepares you for easier migrations to a new ci system


Jenkins has pros and cons.

It's one of the few CI tools where you can test your pipeline without committing it. You also have controls such as only pulling the pipeline from trunk, again, something that wasn't always available elsewhere.

However, it can also be a complete footgun if you're not fairly savvy. Pipeline security isn't something every developer groks.


When was the last time you used Jenkins? I don't get the hate. Not only from you, but lots of people on the internet. What makes Jenkins stand out IMO is the community and the core maintainers, they are perhaps moving slow, but they are moving in the right directions. The interface looks really nice now, they've done a lot of ux improvements lately.


I haven't used Jenkins in a few years, so some of this might change, but in working with it I saw that Jenkins has a few fundamental flaws that I don't see them as working to change:

1. There is no central database to coordinate things. Rather it tries to manage serialization of important bits to/from XML for a lot of things, for a lot of concurrent processes. If you ever think you can manage concurrency better than MySQL/Postgres, you should examine your assumptions.

2. In part because of the dance-of-the-XMLs, when a lot of things are running at the same time Jenkins starts to come to a crawl, so you are limited on the number of worker nodes. At my last company that used Jenkins they instituted rules to keep below 100 worker nodes (and usually less than that) per Jenkins. This lead to fleets of Jenkins servers (and even a Jenkins server to build Jenkins servers as a service), and lots of wasted time for worker nodes.

3. "Everything is a plugin" sounds great, but it winds up with lots of plugins that don't necessarily work with each other, often in subtle ways. In the community this wound up with blessed sets of plugins that most people used, and then you gambled with a few others you felt you needed. Part of this problem is the choice of XMLs-as-database, but it goes farther than that.

4. The way the server/client protocol works is to ship serialized Java processes to the client, which then runs it, and reserializes the process to ship back at the end. This is rather than having something like RPC. This winds up being very fragile (e.g.: communications breaks were a constant problem), makes troubleshooting a pain, and prevents you from doing things like restarting the node in the middle of a job (so you usually have Jenkins work on a Launchpad, and have a separate device-under-test).

Some of these could be worked on, but there seemed to be no desire in the community to make the large changes that would be required. In fact there seemed to be pride in all of these decisions, as if they were bold ideas that somehow made things better.


both the old & new interfaces to Jenkins are riddled with bugs, work seems to be maintenance mode, across the plugin ecosystem too

If you are talking about Jenkins-X, that is a different story, it's basically a rewrite to Kubernetes. I haven't talked to anyone actually using it, if you go k8s, you are far more likely to go argo


It seems like a simple CI/CD in an airgapped environment might be simpler to implement than to (1) learn and (2) onboard an off-the-shelf solution when your airgapped requirement limits your ability to leverage the off-the-shelf ecosystem.


This was more like an educational exercise


Since you're exercising, you can take it to the next level where you don't specify the next step but the inputs to each task, allowing you to infer the DAG and implement caching...


You can do this with cue/flow, but have not turned it into a full CI system. The building blocks are there


Never heard of cue/flow will definitely check it out


cue/flow is a Go package and what powers the scripting / custom tools part of CUE. It's built on top of the language and a graph resolver.

I have another custom flow implementation that I find more ergonomic: https://hofstadter.io/getting-started/task-engine/


The big issue with LLMs is that they’re usually right — like 90% of the time — but that last 10% is tough to fix. A 10% failure rate might sound small, but at scale, it's significant — especially when it includes false positives. You end up either having to live with some bad results, build something to automatically catch mistakes, or have a person double-check everything if you want to bring that error rate down.


Depending on the use case, a 10% failure rate can be quite acceptable. This is of course for non-critical applications, like e.g. top-of-funnel sales automation. In practice, for simple uses like labeling data at scale, I'm actually reaching 95-99% accuracy in my startup.


yes, the entire design relies on a human to check everything. basically it presents what it thinks should be done, and why. the human then agrees or does not. much work is put into streamlining this but ultimately its still human controlled


At the risk of being obvious, this seems set up for failure in the same way expecting a human to catch an automated car's mistakes is. Although I assume mistakes here probably don't matter very much.


This reminds me the issue with the old windows access control system.

If those prompts pop up constantly asking for elevated privileges, this is actually worse because it trains people to just reflexively allow elevation.


yes, mistakes are not a huge problem. they will become evident farther down the process and they happen now with the human only system. worst case is the LLM fails and they just have to do the manual work that they are doing now


I hate reading documentation too.


Ah a fellow tab-autocomplete power user.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: