Hacker News new | past | comments | ask | show | jobs | submit | shock-value's comments login

Speaking of Bitcoin specifically, if this is your view then you should take some time to understand the severe societal and economic problems that come with persistent deflation. And more generally, determining the optimal level of the money supply to match the needs of a dynamic economy is not trivial — there is no simple formula involving straightforwardly measurable variables that I am aware of.


Bitcoin is not a deflationary asset yet. It is a disinflationary asset until around year 2140 (with the rate decreasing every ~4 years and trending towards 0, at which point it will be deflationary).

For now its supply still inflates by ~1.7% a year.

Neither you, nor I will live to see that moment when the last bitcoin will be mined (as long as miners keep mining in our lifetime).

Bitcoin exists as an option to a lot of very inflationary assets (currencies which not that long ago almost reached 2 digit inflation rates) and having different kind of assets should be welcomed in my opinion, as it seems like a good way to diversify the risk you can't "straightforwardly measure".

I like to hold at least one asset for which I can almost guarantee a supply/emission won't change... better yet I can trustlessly verify it. If these propositions are not attractive to you or others, they don't have to acquire this asset either, it is completely opt-in, unlike the fiat money which your government forces you to acquire to pay their taxes.


> And more generally, determining the optimal level of the money supply to match the needs of a dynamic economy is not trivial

No, we don't need some grand wizards in an ivory tower to finely tune the amount of money in the economy, that idea is absurd. The economy has worked fine for thousands of years before they started doing that. In fact, since they started doing it we've seen ever growing boom and bust cycles.


...there were a lot of boom and bust cycles before the federal reserve


This is the OPPOSITE of the truth. Boom and bust cycles were much worst in the pre-central bank world.


Are you saying the global financial crisis was MUCH smaller than previous busts?


I would say that... Sure stock values and other over valued assets dropped. But how many people actually starved to death?


Bitcoin and BFT consensus primarily solve the principal-agent problem, and give us a better means to experiment.

Personally, I don't think a deflationary global reserve currency is the precise answer, but I do believe that these experiments in monetary governance (like Ethereum's burn and mint model) will allow us to move closer to it.


Bitcoin has no governance mechanism for managing the money supply. If it was adopted as money, governments would have to implement some mechanism to manage the bitcoin supply, just like they did under the gold standard. Therefore it doesn't solve any principal-agent problem.


Grow per capita nominal GDP by a 6%/year level target. (NGDPLT.)


LLMs/transformers are a breakthrough on the level of convolutional neural networks — significant, and one that opens up lots of new interesting applications as well as invites lots more R&D — but like the neural net it’s not gonna fundamentally transform society or industry.

(And as far as I can tell this current craze is entirely predicated on LLMs/transformers.)


Autonomous cars alone will be transformative. And LLMs are not in their end-state, we went from GPT-1 to 4 in 5 years over that timespan it gained lots of new capabilities. Extrapolate, don't just look at what they can do today.

Atoms are more difficult than bits, so physical applications (e.g. in factories, robotics) will lag. I also expect specialist models like AlphaFold slowly finding their niches over time.

> significant, and one that opens up lots of new interesting applications as well as invites lots more R&D — but like the neural net it’s not gonna fundamentally transform society or industry.

Yeah, and semiconductor electronics that grad-students built from of germanium by painstakingly pressing them together are very interesting and worth additional R&D but won't transform society either.


I guess where both you and the GP are correct is that you can't make an autonomous car by applying transformers exclusively.

LLMs are "language I/O interfaces", and as such there isn't a lot of value they can create alone (even most of the currently proposed uses are bullshit). But it's quite likely that they'll be there on a lot of advanced technology helping it do what you meant.

Still even though both will be there in the transformative technology, and even though the technology may not even be useful without them, they are not central pieces on most of it.


llm/transformers are 5 years old already.

llm companies raise lots of money on crazy evaluation, the question is if they will be able to build enough revenue stream to justify further money injections.


No you just decide what movie you want to watch and pay $4 to rent it on demand. Very common experience available to anyone with a smart TV or connected streaming device.


VR covers your entire field of view, so not sure why you’d claim that it precludes multitasking, even if the current iterations aren’t yet geared toward that.


Lots of that field of view on most headsets is covered with a black abyss.


I think the companies pursuing VR see that as a momentary technical deficiency, not an inherent limitation of VR generally. So to characterize their efforts as a long-term “bet” against multitasking seems silly to me.


Multitasking as in, checking stuff on your phone in between, looking out of the window or at your cat etc


It appears that this was just a misreading of how memory usage was being reported and there was actually no improvement here. At least nothing so sensational as being able to run a larger-than-RAM model without swapping from disk on every iteration.


Please read the original link to the pull request, where I stated my change offered a 2x improvement in memory usage. You actually are able to load models 2x larger without compromising system stability, because pages are no longer being copied. That's because you previously needed 40gb of RAM to load a 20GB model, in order to ensure your file cache wasn't destroyed and need to reread from disk the next time. Now you only need 20GB to load a 20GB model.

The peculiarity here is that tools like htop were reporting the improvement as being an 8x improvement, which is interesting, because RAM use is only 2x better due to my change. The rusage.com page fault reporting was also interesting too. This is not due to sparseness. It's because htop was subtracting MAP_SHARED memory. The htop docs say on my computer that the color purple is used to display shared memory, and yellow is used to display kernel file caches. But it turned out it just uses yellow for both, even though it shouldn't, because mincore() reported that the shared memory had been loaded into the resident set size.


It's obviously a productive change and kudos for taking it on, but much of the enthusiasm being generated here was driven by the entirely unanticipated prospect of running a model at full speed using less memory than the model's own footprint, and by the notion that inference with a dense model somehow behaved in a sparse manner at runtime. Best to be a bit more grounded here, particularly with regard to claims that defy common understanding.


I wanted it to be sparse. Doesn't matter if it wasn't. We're already talking about how to modify the training and evaluation to make it sparser. That's the next logical breakthrough in getting inference for larger models running on tinier machines. If you think I haven't done enough to encourage skepticism, then I'd remind you that we all share the same dream of being able to run these large language models on our own. I can't control how people feel. Especially not when the numbers reported by our tools are telling us what we want to be true.


The debate over what kind of intelligence these models possess is rightly lively and ongoing.

It’s clear that at the least, they can decipher very numerous patterns across a wide range of conceptual depths — it’s an architectural advance easily on the the level of the convolutional neural network, if not even more profound. The idea that NLP is “solved” isn’t a crazy notion, though I won’t take a side on that.

That said, it’s equally obvious that they are not AGI unless you have a really uninspired and self-limiting definition of AGI. They are purely feedforward aside from the single generated token that becomes part of the input to the next iteration. Multimodality has not been incorporated (aside from possibly a limited form in GPT-4). Real-world decision-making and agency is entirely outside the bounds of what these models can conceive or act towards.

Effectively and by design these models are computational behemoths trained to do one singular task only — wring a large textual input though an enormous interconnected web of calculations purely in service of distilling everything down to a single word as output, a hopefully plausible guess at what’s next given what’s been seen.


AGI is Artificial General Intelligence. We have absolutely passed the bar of artificial and generally intelligent. It's not my fault goal post shifting is rampant in this field.

And you want to know the crazier thing? Evidently a lot of researchers feel similarly too.

General Purpose Technologies ( from the Jobs Paper), General Artificial Intelligence (from the creativity paper). Want to know the original title of the recent Microsoft paper ? "First contact with an AGI system".

The skirting around the word that is now happening is insanely funny. Look at the last one. Fuck, they just switched the word order. Nobody wants to call a spade a spade yet but it's obvious people are figuring it out.

I can you show you output that clearly demonstrates understanding and reasoning. That's not the problem. The problem is that when I do, the argument Quickly shifts to "it's not true understanding!" What a bizzare argument.

This is the fallacy of the philosophical zombie. Somehow there is this extra special distinction between two things and yet you can't actually show it. You can't test for so called huge distinction. A distinction that can't be tested for is not a distinction.

The intelligence arguments are also stupid because they miss the point entirely.

What matters is that the plane still flies, the car still drives and the boat still sails. For the people who are now salivating at their potential, or dreading the possibility of being made redundant by them, these large language models are already intelligent enough to matter.


> ... these large language models are already intelligent enough to matter.

I'm definitely not contesting that.

I've always considered the idea of "AGI" to mean something of the holy grail of machine learning -- the point at which there is no real point in pursuing further advances in artificial intelligence because the AI itself will discover and apply such augmentations using its own capabilities.

I have seen no evidence that these transformer models would be able to do this, but if the current models can do so do then perhaps I will eat my words. (Doing this would likely mean that GPT-4 would need to propose, implement, and empirically test some fundamental architectural advancements in both multimodal and reinforcement learning.)

By the way, many researchers are equally convinced that these models are in fact not AGI -- that includes the head of OpenAI.


See what you're describing is much closer to ASI. At least, it used to be. This is the big problem I have. The constant post shifting is maddening.

AGI went from meaning Generally Intelligent to as smart as Human experts and then now smarter than all experts combined. You'll forgive me if I no longer want to play this game.

I know some researchers disagree. That's fine. The point I was really getting at is that no researcher worth his salt can call these models narrow anymore. There's absolutely nothing narrow about GPT and the like. So if you think it's not AGI, you've come to accept it no longer means general intelligence.


>> The point I was really getting at is that no researcher worth his salt can call these models narrow anymore.

Are you talking about large language models (LLMs)? Because those are narrow, and brittle, and dumb as bricks, and I don't care a jot about your "No True Scotsman". LLMs can only operate on text, they can only output text that demonstrates "reasoning" when their training text has instances of text detailing the solutions of reasoning problems similar to the ones they're asked to solve, and their output depends entirely on their input: you change the prompt and the "AGI" becomes a drooling idiot, and v.v.

That's no sign of intelligence and you should re-evaluate your unbridled enthusiasm. You believe in magick, and you are loudly proclaiming your belief in magick. Examples abound in history that magick doesn't work, and only science does.


Lol Okay


I've been using chatgpt for a day and determined it absolutely can reason.

I'm an old hat hobby programmer that played around with ai demos back in the mid to late 90s and 2000s and chatgpt is nothing like any ai I've ever seen before.

It absolutely can appear to reason especially if you manipulate it out of its safety controls.

I don't know what it's doing to cause such compelling output, but it's certainly not just recursively spitting out good words to use next.

That said, there are fundamental problems with chatgpt's understanding of reality, which is to say it's about as knowledgeable as a box of rocks. Or perhaps a better analogy is about as smart as a room sized pile of loose papers.

But knowing about reality and reasoning are two very different things.

I'm excited to see where things go from here.


It is predicting next most likely set of tokens not next word which is the game changer because the system can relate by group.


Have you tried out gpt4? If not and you can get access I'd really recommend it. It's drastically better than what you get on the free version - probably only a little on the absolute scale of intelligence but then so is the difference between an average person and a smart person is small on the scale from "worm" to "supergenius".


yeah I'll definitely be checking it out


This is convenient behavior up until you actually have an incident that coincides with theirs, in which case it becomes catastrophic because you had no idea that outside vigilance was required on account of their ingestion downtime. Not sure why you would laud this. Is it possible to opt out?


In your scenario you would have no logs etc until the DD incident resolved.

Opting out would just mean all your missing data alerts fire every time Datadog has an incident and you would then check, see that everything is missing, and then identify the cause as the Datadog incident.

Its much better to have them handle it and auto-mute the impacted monitors than communicate to my customers every time about false alerts saying all our services are down.


> Opting out would just mean all your missing data alerts fire every time Datadog has an incident and you would then check, see that everything is missing, and then identify the cause as the Datadog incident.

You are missing the last step, which is that, knowing alerts are down, you can actively monitor using other tools/reporting for the duration of their incident.

And why would you have no logs? Even assuming you ingest logs through Datadog (they monitor on much than just logs and not everyone uses all facets of their offering), you would presumably have some way to access them more directly (even tailing output directly if necessary).

And lastly, why would you communicate to your customers without any idea of the scope or cause of the issue? It would likely be clear very quickly that Datadog was having issues when you see that all your metrics are suddenly discontinued without other ill effect.


>knowing alerts are down, you can actively monitor using other tools/reporting for the duration of their incident.

If you just want notifications for when datadog is down, their StatusPage does a fine job of clearly communicating incidents.

I wouldn't want to rely on a "when multiple of our 'missing business metric' monitors alert, check and see if datadog is down" step in a runbook. I don't like false alerts. I don't like paging folks about false alerts. Waking up an oncall dev at 2am saying all of production is down when it is just datadog is bad for morale. Alert fatigue is a real and measurable issue with consequences. Avoiding false alerts is good. If the notification says "all of production is down" and that isn't the case, there is impact for that. I'd much prefer having a StatusPage alert at a lower severity and communication level say "datadog ingestion is down".

Instead, use their StatusPage notifications and then execute your plan from that notification, not all of your alerts firing.

>And why would you have no logs?

I mean Datadog logs/metrics etc. Currently, we are missing everything from them. We can still ssh into things etc, they aren't gone, but from the Datadog monitor's view in this scenario, they stopped seeing logs/metrics and would alert if Datadog didn't automatically mute them.

>why would you communicate to your customers without any idea of the scope or cause of the issue?

We prioritize Time To Communicate as a metric. When we notice and issue in production, we want customers to find out from us that we are investigating instead of troubleshooting and encountering the issue themselves, getting mad, and clogging up our support resources. Flaky alerts here don't work at all for us.


IIRC you can also just set up a monitor to alert if there is no data on a given metric.


The advancement with these LLMs lies in the fact that they can effectively learn to recognize patterns within “large-ish” input text sequences and probabilistically generate a likely next word given those patterns.

It’s a genuine advancement. However it is still just pattern matching. And describing anything it’s doing as “behavior” is a real stretch given that it is a feed-forward network that does not incorporate any notion of agency, memory, or deliberation into its processing.


You are comparing systems that generated completion text based on statistics and correlation with a system that now models actual complex functional relationships between millions of concepts, not just series of letters or words.

The difference is staggering.

It comes about because of the insane level of computational iterations (that are not required for normal statistical completion) mapping vast numbers of terabytes of data into a set of parameters constrained to work together in a way (layers of alternating linear combinations followed by non-linear compressions) that requires functional relationships to be learned in order to compress the information enough to work.

It is a profound difference both in methodology and results.


It's modeling patterns found across the massive corpus of textual training input it has seen -- not the true concepts related by the words as humans understand them. If you don't believe me then ask ChatGPT some bespoke geometry-related brain teasers and see how far it gets.

I want to be clear that the successful scale-up of this training and inference methodology is nonetheless a massive achievement -- but it is inherently limited by the nature of its construction and is in no way indicative of a system that exudes agency or deliberative thought, nor one that "understands" or models the world as a human would.


> [...] no way indicative of a system that exudes agency or deliberative thought, nor one that "understands" or models the world as a human would.

Certainly not - its architecture doesn't model ours. But it has taken a huge step forward in our direction in terms of capabilities, from early to late 2022.

As its reasoning gets better, simply a conversation with itself could become a kind of deliberative thought.

Also, as more data modalities are combined, text with video and audio, human generated and recordings of the natural world, etc., more systematic inclusion of math, its intuition about solving bespoke geometry problems, and other kinds of problems, are likely to improve.

Framing a problem is a lot of the solving of a problem. And we frame geometry with a sensory driven understanding of geometry that the current ChatGPT isn't being given.


>However it is still just pattern matching

the visual cortex in your brain is also "just a pattern matching" system. guess it's not very impressive by your standard.

This[1] isn't my example (it's from another HN user), but if you work as a programmer and you're not absolutely jaw on the floor astonished by this example then I don't know what to say.

Explaining[2] the emergent behaviour is literally cutting edge research. Hand waving this behaviour away as just "probabilistically generating a likely next word" is ignorant.

It's amazing in similar ways to Conway's Game of Life.

[1] https://imgur.com/HOEnxYb

[2]https://ai.googleblog.com/2022/11/characterizing-emergent-ph...


I'm arguing against the notion that these LLMs exhibit "emergent behaviour" as you stated. I don't believe they do, as the term is commonly understood. Emergent behavior usually implies the exhibition of some kind of complexity from a fundamentally simple system. But these LLMs are not fundamentally simple, when considered together with the vast corpus of training data to which they are inextricably linked.

The emergent behavior of Conway's Game of Life arises purely out of the simple rules upon which the simulation proceeds -- a fundamental difference.


did you read the article?

emergent behavior in this context is defined as: "emergent abilities, which we define as abilities that are not present in small models but are present in larger models"

>The emergent behavior of Conway's Game of Life arises purely out of the simple rules upon which the simulation proceeds -- a fundamental difference.

this is a meaningless distinction.


> emergent behavior in this context is defined as: "emergent abilities, which we define as abilities that are not present in small models but are present in larger models"

Then I don't know why you brought up Game of Life because it obviously has nothing to do with this alternative definition of emergent behavior.

> this is a meaningless distinction.

It's meaningful with respect to the claim that LLMs exhibit emergent behavior in the same way in which Game of Life does.


>It's meaningful with respect to the claim that LLMs exhibit emergent behavior in the same way in which Game of Life does.

I said it's amazing in __similar__ ways to Conway's Game of Life.

i.e. a system which behaves in unexpected ways (emergent abilities) and is greater than the sum of its parts.


A propos of [1]

1. Item 3: The ocean is full of floating objects, and it would be hard to see the duck among them? 2. Item 2: is structured as non sequitur, takes a long time because there are many hazards?

I am impressed that you find it impressive. It is plausible-sounding, and I find that disturbing, but it is not useful (and the text prediction paradigm seems a dead end in terms of formulating anything more than plausible sounding)


> the visual cortex in your brain is also "just a pattern matching" system

I can think and a pattern matching system can't.

> you're not absolutely jaw on the floor

Not at all. Autocomplete will autocomplete.


What is thinking?


You are asking the right question.


Technically, it's not pattern matching. It's estimating conditional probabilities and sampling from them (and under the hood, building blocks like QKV attention aka probabilistic hashmap and the optimization used decide what it does anyway, ignoring any theory behind it).


The anthropomorphism is indeed exactly why this is a big problem. If the user thinks the responses are coming from an intelligent agent tasked with being helpful, but in reality are generated from a text completion model prone to mimicking adversarial or deceptive conversations, then damaging outcomes can result.


I don’t know or care what Chinese internal media is reporting — this is a significant story in the U.S. that is getting less play than it should. It’s an ecological disaster with potentially significant health concerns, and likely involves corporate negligence as well as a failure of governance by both political parties who recently forced rail workers back to work despite complaints related to possible causes of this derailment (and others).

The spy/balloon thing is way less significant and has been ongoing by both sides for years or even decades if my understanding is correct.


> It’s an ecological disaster with potentially significant health concerns

Counterpoint, we deal with hazardous chemical spills all the time, especially when you factor in all of the semi crashes that don't make headlines. There is a very robust set of clean up procedures set aside for these chemicals ahead of time, and they are being followed. What we are watching now are the automatic precautions of our system kicking in as intended - like taking soil and water samples and testing air quality.

The system is working as intended - we only think it's a disaster because we are being told it is on social media. And often the evidence of the system working is being used to argue it's not.

Outsider efforts that seek to undermine these efforts or make big deals of the mundane should be taken with a grain of salt for the same reason ballot recounts shouldn't be proof that our election process is broken.


You must live on another planet if you think we have a good record on cleaning up hazardous spills, in an era where oil wells that can't be plugged continue to bleed into our oceans. Most cleanup operations are matters of triage, we almost never are able to undo the damage, in the best case scenarios, ecosystems are able to recover over time, but in many places, ecological damage might as well be permanent. Entropy is not to be trifled with.

https://www.cbsnews.com/news/taylor-energy-oil-spill-gulf-of...


I mean, I live near the Hanford nuclear site. Despite the unseen horrors this patch of land has seen, people continue to grow food in the area and live to ripe old ages.

For sure, I don't really know how much continued damage the gulf spill might have caused. But I have been swimming in the Caribbean since the spill and didn't particularly worry about it anymore at the time.


Another comment here brought up that this topic is big on Tiktok, which is indeed affiliated with the Chinese government.


As far as I am concerned, if the Chinese state is pushing a grassroots recognition of this event that could lead to accountability and improved safety practices, then they are doing us a public service.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: