Hacker News new | past | comments | ask | show | jobs | submit | mmoskal's comments login

Mobile comment view is messed up on Safari.

The top level view seems to leave a bit too much margins on mobile.


Thanks for the feedback, i need to give mobile some more attention. Will fix ASAP

Assuming $30k GPU with 3yr deprecation, it's additionally $1.14/h. Much more than energy.

I guess the short way to say it is that "undecidable" doesn't mean "it can't ever be decided", just not always.

And of course all programs of practical significance are finite state machines (since there is only a finite number of atoms in the universe).


Isn't the point closer to, humans simply go "hey that seems to be taking a little long?" when a program doesn't halt, so why couldn't a machine? Basically a fairly obvious constraint on the solution space is "completes in less then N wall-clock time".

You can definitely detect a portion of halting machines this way, but it's probably a relatively small portion because the Busy Beaver numbers grow inconceivably quickly: the longest-running machines that halt can go practically forever, you'd need more time than the universe has negentropy left to detect them.

One sentence from The Economist seems to explain more than TFA: "Microsoft reported a 31% increase in its indirect (Scope 3) emissions last year from building more data centres (including the carbon found in construction materials) as well as from semiconductors, servers and racks."

So no, it's not about lack of renewable electricity.

https://www.economist.com/the-world-this-week/2024/05/16/bus...


Projects change license for new code going forward. The old code remains available under the previous license (and sometimes new). Here, they are able to change the conditions for existing weights.

From model card:

Falcon2-11B was trained on 1024 A100 40GB GPUs for the majority of the training, using a 3D parallelism strategy (TP=8, PP=1, DP=128) combined with ZeRO and Flash-Attention 2.

Doesn't say how long though.


It does say how long on Huggingface:

> The model training took roughly two months.


The post seems to be about execution speed though. However, even there it's definitely not #1 factor as witnessed by popularity of CPython...

Survivor bias? maybe they just don't catch the ones that only had two suitcases of cash.

This. + Hastiness and overconfidence. When you do it over a long stretch of time, you try to stretch the limits occasionally

From the fraud cases I've worked on, greed does factor in and you can often see a point in time within the data where it seems like they realize they are 'getting away with it'.

TFA says you can teach it new facts, but it's very slow and makes the model hallucinate more.


Yes. It's speculative decoding but instead of generating just a few sequential tokens with the draft model they generate a whole tree of some sort of optimal shape with hundreds of possible sequences.

It ends up being somewhat faster than regular speculative decoding in normal setting (GPU only). If you are doing CPU offloading it's massively faster.

Edit typo


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: