Hacker Newsnew | past | comments | ask | show | jobs | submit | nextos's commentslogin

Lobsters and filter posts with vibecoding tag.

Progress with RL is very interesting, but it's still too inefficient. Current models do OK on simple boring linear code. But they output complete nonsense when presented with some compact but mildly complex code, e.g. a NumPyro model with some nesting and einsums.

For this reason, to be truly useful, model outputs need to be verifiable. Formal verification with languages like Dafny , F*, or Isabelle might offer some solutions [1]. Otherwise, a gigantic software artifact such as a compiler is going to have a critical correctness bugs with far-fetched consequences if deployed in production.

Right now, I think treating a LLM like something different than a very useful information retrieval system with excellent semantic capabilities is not something I am comfortable with.

[1] https://risemsr.github.io/blog/2026-02-04-nik-agentic-pop


Human-written compilers have bugs too! It takes decades of use to iron them out, and we’re introducing new ones all the time.

Snap and Flatpak do both sandboxing and package management.

You can use the underlying sandboxing with bwrap. A good alternative is firejail. They are quite easy to use.

I prefer to centralize package management to my distro, but I value their sandboxing efforts.

Personally, I think it's time to take sandboxing seriously. Supply chain attacks keep happening. Defense is depth is the way.


I really like everything Uri Alon (last author) publishes, but these types of studies have a history of inflating genetic contributions to phenotypes. Decoupling genetics from environment is not easy as they are both highly correlated.

In fact, the article discussion states: "Limitations of this study include reliance on assumptions of the twin design, such as the equal environment assumption". My take on this is that the main result of the article is probably true, but the 50% figure is likely to be inflated.


I hit the jackpot with the ultrasound technician who spoke passionately about what she believed about lifestyle risk for cardiovascular conditions and she believed quite strongly that heart disease runs in families more because lifestyle runs in families than because of genetics. She's not at the top of the medical totem pole but I can say she inspired me to take responsibility for my health than the specialist who I talked to about the results.

Actually the opposite is true.

If the environment was significantly more varied in health impact between twin comparisons than expected, then the correlations they found under estimate the genetic component.

Noise weakens correlation. Removing noise strengthens correlations.

Some randomness is part of the signal being studied, and some is undesired measurement noise to be controlled for. And it is only the latter that is beneficial to be carefully removed or otherwise controlled for.


I prefer X11 as well, but it has some security issues. Notably, all applications can read your input at any time. It's really hard to sandbox.

Wayland brought some irritations, including increased latency, and an architecture that requires rethinking all window managers. A rewrite is not enough. Very annoying.


I will never understand why "the computer can tell what input it is receiving" has turned into an accepted threat model.

I understand that we have built a computer where our primary interface depends on running untrusted code from random remote locations, but it is absolutely incredible to me that the response to that is to fundamentally cripple basic functionality instead of fixing the actual problem.

We have chosen to live in a world where the software we run cannot be trusted to run on our computers, and we'd rather break our computers than make another choice. Absolutely baffling state of affairs.


Defense in depth. One compromised application may do a lot of harm if it has access to your keyboard inputs. Supply chain attacks are not that uncommon. While you can trust software developers, you cannot completely trust their builds.

[dead]


I agree. I think fixing the keylogging issue should be possible without dumping the entire architecture. Perhaps the new X11 fork https://x11libre.net will achieve that? At least, it's encouraging to hear it's getting maintained.

Regarding (recent) supply chain attacks, Linux needs to take supply integrity and sandboxing more seriously. The tools to do so are there (e.g. Nix and firejail/bwrap) and, unlike Wayland, they play well with existing software.


I have doors between rooms in my house, despite its being inhabited by members of the same family who trust each other.

And when someone violates that trust, do you then tear the house down and build one with only external doors, requiring inhabitants to circle in the yard to move between rooms? The point of the Wayland security model is that the inhabitants of the house do not trust each other, and the architecture of the house must change to accommodate that.

I'm not impressed with the analogy. I am not confused about the goals of Wayland's security model. I am dismayed at the poor judgment elsewhere in computing that has led to its necessity.


Deep SSMs, including the entire S4 to Mamba saga, are a very interesting alternative to transformers. In some of my genomics use cases, Mamba has been easier to train and scale over large context windows, compared to transformers.

True, but Google Translate was already "AI". They previously used LSTMs. And before LSTMs, it was ML-like statistical translation.

Maybe formal methods have a chance of becoming mainstream now [1]?

This would increase the rigor of software engineering and put it on par with civil engineering.

Some niches like real-time embedded systems are already pretty much the same.

[1] https://martin.kleppmann.com/2025/12/08/ai-formal-verificati...


I doubt it, I feel like it might improve shops that already care and are already creating with rigor. I don't think it'll raise the bar for the avg shop. However, perhaps that's just be being cynical. By real time embedded is the same do you mean the same in the sense that they are just as poor in quality?

> [...] the same in the sense that they are just as poor in quality?

I mean some real-time software for critical embedded systems has an incredible level of rigor, making heavy use of static analysis, model checking, and theorem proving.


Noted, perhaps I'll investigate as a possible next career step. Thanks!

Medicaid holds previous addresses, household details, previous diagnoses, ethnicity, etc.

It is quite trivial to infer if someone is likely to have emigrated to the US due to obvious gaps in records or in their relatives' ones.

This is what Palantir does, essentially. Simple inference and information fusion from different sources.


Ironically, the Danish Government is a heavy user of Palantir systems, including creepy predictive policing solutions.

I would be keen to know if citizen data is being handled correctly, following GDPR/LED.

Given previous Danish client-state-like cooperation with NSA to spy on other EU countries, I can imagine the answer.


> including creepy predictive policing solutions.

Minority Report coming right up.


Which predictive policing solution from palantir are they using?


Their local Palantir implementation is called POL-INTEL. This thesis presents a good critical overview [1].

[1] https://en.itu.dk/-/media/EN/Research/PhD-Programme/PhD-defe...


GDPR has carveouts for governments and law enforcement so they can do whatever for those purposes.


The framework is the one I referred to (EU LED). In Denmark, LED is implemented in the Danish Law Enforcement Act.

However, LED has some purpose limitations, which critics argue the Danish Law Enforcement Act has bypassed. Some are trying to challenge it.


I love how powerful the GDPR marketing was that it made people forget that there are massive exceptions for prevention of crime and for the government


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: