This isn't even the worst example, since it does at least have the correct info buried amongst tons of Ai generated garbage, but I can't use this for reference, since it tells me 4 different drill sizes. I've had to switch back to a paper copy of the machinist's handbook, since I can't trust the internet to give me accurate information anymore. 10 years ago, I could easily search for the clearance hole for a 10-24 fastener, now I get AI junk that I can't trust.
How have we regressed to the point that I'm better off using a paper book than online charts for things that don't change?
Half of web in Russia is blocked. Literally, powers that be think of Russian tech companies as of their servants and nothing more. Yandex basically sold their main asset, domain name to other entity.
This, I am a terrible note taker. For years a huge part my knowledge and skills relied on "if I found that information once, I'll find it again". My brain compressed the information by memorizing the path to retrieve it again.
Now that does not work anymore. You know some information is out there, you found it once when google worked, now it's lost in the noise.
I'm learning to take notes again and organize them so I can search them easily.
For queries like that I now turn to Gemini / ChatGPT first. Of course, this is only a good idea if I have some way of sanity checking the answer. If I doubt the answer I get back I try Google search instead.
I really like Kagi's approach to this, which is to give a list of references. There's still no guarantee that the answer is correct, but you can at least check the references :).
You can ask a model to provided an analysis of its answer including a probability that it is correct as part of the prompt, helps with doublechecking a lot.
They're consistent to the model, particularly if you ask the model to rationalize its rating. You will get plenty of hallucinated answers that the model can recognize as hallucinations and give a low rating to in the same response.
Models can get caught by what they start to say early. So if they model goes down a path that seems like a likely answer early on, and that ends up being a false lead or dead end, they will end up making up something plausible sounding to try and finish that line of thought even if it's wrong. This is why chain of thought and other "pre-answer" techniques improve results.
Because of the way transformers work, they have very good hindsight, so they can realize that they've just said things that are incorrect much more often than they can avoid saying incorrect things.
Does that extra information come from a separate process than the LLM network? If not then, assuming the same output is not guaranteed from the same input as per usual, then all bets are off correct?
Sorry for the late reply, but if you read this, there is research that shows that prompting a LLM to take variety of perspectives on a problem (IIRC it was demonstrated with code) then finding the most common ground answer improved benchmark scores significantly. So, for example if you ask it to provide a brief review and likelihood of the answer, and repeat that process from several different perspectives, you can get some very solid data.
> How have we regressed to the point that I'm better off using a paper book than online charts for things that don't change?
because products that require iteration lend themself to subscription models which in turn mean a recurring revenue which is deemed superior to onetime payments for a 'finished product'.
This isn't even the worst example, since it does at least have the correct info buried amongst tons of Ai generated garbage, but I can't use this for reference, since it tells me 4 different drill sizes. I've had to switch back to a paper copy of the machinist's handbook, since I can't trust the internet to give me accurate information anymore. 10 years ago, I could easily search for the clearance hole for a 10-24 fastener, now I get AI junk that I can't trust.
How have we regressed to the point that I'm better off using a paper book than online charts for things that don't change?