All OCR is untrustworthy. But sometimes, OCR is useful. (And I've heard it said that all LLM output is a hallucination; the good outputs are just hallucinations that fit.)
A few months ago a warehouse manager sent us a list of serial numbers and the model numbers of some gear they were using -- with both fields being alphanumeric.
This list was hand-written on notebook paper, in pencil. It was photographed with a digital camera under bad lighting, and that photograph was then emailed.
The writing was barely legible. It was hard to parse. It was awful. It made my boss's brain hurt trying to work with it, and then he gave it to me and it made my brain hurt too.
If I had to read this person's writing every day I would have gotten used to it eventually, but in all likelihood I'll never read something this person has written ever again. I didn't want to train myself for that and I didn't have enough of a sampleset to train with, anyway.
And if it were part of a high-school assignment it would have been sent back with a note at the top that said "Unreadable -- try again."
But it wasn't a high school student, and I wasn't their teacher. They were a paying customer and this list was worth real money to us.
I shoved it into ChatGPT and it produced output that was neatly formatted into a table just as I specified with my minimal instruction ("Read this. Make a table.").
The quality was sufficient to allow us to fairly quickly compare the original scribbles to the OCR output, make some manual corrections that we humans knew how to do (like "6" was sometimes transposed with "G"), and get a result that worked for what we needed to accomplish without additional pain.
0/10. I'm glad it worked and I hope I never have to do that again, but will repeat if I must.
US is the same country that allows a practical monopoly of NVIDIA on GPUs and Intel on CPU (or at least an oligopoly), and then pretend "foreigners are out to get us". It gets one to know one.
On the other hand, I find this a bit concerning too? The USA is starting to look a bit more like China. There is now only “one world view” for us. Given the friend group between the people who run X and Meta and it might leave us in a precarious situation?
Banning TikTok only treats the symptom , the real disease is that people are way to susceptible to propaganda and misinformation.
Did you miss the part where China is a foreign adversary? They don't play nice. If you try to play nice with someone who wants to kill you then you get killed.
But I’m not saying China are nice guys, I’m saying we’re now left with the same thing, just ran by the US government. You might think that’s a good thing. I don’t.
Personally I think all closed source social media should be outlawed and all algorithms used should be audited by a third party.
End users just won't care about the algorithm. Try talking to a niece or nephew, especially one who makes money on the platform about The Algorithm and you'll get blank stares, or, at best, a "yeah I know, but...".
If you've had better luck, let me know (actually).
As for "being China", every country has protections on what goes in or out of the country including media. A lot of countries won't let you own a newspaper or news broadcast channel, so this is the next extension of that sort of idea.
It's the same idea as not allowing a company from the USSR to run a news channel during the Cold War, although obviously the lines are fuzzier and still being discovered with apps and algorithms.
We have nukes. If they try to kill us, everyone dies. "Foreign adversary" just means they're big enough to get a seat at the table in a multipolar world.
Dominant market leaders aren't inherently bad for the world. That's why anti-trust laws are narrow. Only when they are so ingrained and conspire to be anti-competitive (usually via lobbying gov policy to create barriers to entry) that they harm the ability for competition to replace them. NVIDIA constantly and perpetually have companies at their throats looking to take their market, which means they better deliver to customers.
It's the same country that allows a practical monopoly of NVIDIA on GPUs and Intel on CPU (or at least an oligopoly), and then pretend "foreigners are out to get us". It gets one to know one.
> They predict what they should say based on what they read.
There's so much anthropomorphization in the air in these debates, that I worry even this statement might get misinterpreted.
The text generator has no ego, no goals, and is not doing a self-insert character. The generator extends text documents based on what it has been initialized with from other text documents.
It just happens to be that we humans have purposely set up a situation where the document looks like one where someone's talking with a computer, and the text it inserts fits that kind of document.
Yep. As I said in another post, they’re human simulators. It looks and sounds enough like a human that it’s tricking people into believing in this illusion of intelligence or intent. I have to imagine the very smart people at OpenAI and Anthropic understand this, and I think a lot of these reports about apparent sentience are being released to push the hype wave and generate investment before the truth becomes apparent to everyone.
Am I the only one who's getting annoyed of seeing LLMs be marketed as competent search engines? That's not what they've been designed for, and they have been repeatedly bad at that.
If everything you do starts by asking an LLM, then you start with superficial research, because frankly it was never anything better than a fancy google search.