Hacker News new | past | comments | ask | show | jobs | submit login

wow, the site is so slow. getting "This resource could not be found" errors on reloads.

anyway, here's the text:

> Today was the first day that I could definitively say that #GPT4 has saved me a significant amount of tedious work. As part of my responsibilities as chair of the ICM Structure Committee, I needed to gather various statistics on the speakers at the previous ICM (for instance, how many speakers there were for each section, taking into account that some speakers were jointly assigned to multiple sections). The raw data (involving about 200 speakers) was not available to me in spreadsheet form, but instead in a number of tables in web pages and PDFs. In the past I would have resigned myself to the tedious taks of first manually entering the data into a spreadsheet and then looking up various spreadsheet functions to work out how to calculate exactly what I needed; but both tasks were easily accomplished in a few minutes by GPT4, and the process was even somewhat enjoyable (with the only tedious aspect being the cut-and-paste between the raw data, GPT4, and the spreadsheet).

> Am now looking forward to native integration of AI into the various software tools that I use, so that even the cut-and-paste step can be omitted. (Just being able to resolve >90% of LaTeX compilation issues automatically would be wonderful...)

Ironically, Tao's post convinces me that AI, though amazing, isn't really the solution. Better UX and data quality is. Why was the data so disjoint to begin with? Why is Latex so hard to work with?

In this case GPT-4 is used to solve a problem that shouldn't have even been one to begin with. The administrators of the ICM could've simply exported the raw data as a Google Sheet (for example) and his problem could've been trivially solved even without GPT-4.




> Ironically, Tao's post convinces me that AI, though amazing, isn't really the solution. Better UX and data quality is.

incredible what mental hoops people will jump through to disqualify AI!

the reality is that a lot of things have a terrible UI with unorganized data. that's why this tool is so amazing - because it doesn't matter anymore.


To reinforce this point, we have known for a very long time that better UX and data quality has innumerable benefits. Remember back when everybody was excited about web 2.0 and all the amazing mashups which would soon become possible?

If you are producing data then exposing it in a nice programmable format is an extra cost and generally provides you no benefit. It usually hurts you, if people stop visiting your site and see fewer of your ads!

This is "really" a problem of incentives. It is usually not possible to capture any of the positive externalities of exposing your data. So maybe we could convince everybody in the world to switch to using different browsers with a native micropayment system; that might incentivize everybody to release all data as clean machine-readable tuples.

What I'm saying is, the phrase "Better UX and data quality" ignores just how hard that solution really is. It turns out training an LLM over most of the internet is _easier_ than global coordination.


i.e. semantic web was dead on arrival.


But chatgpt can crawl a semantic web and use it without us knowing.

I have asked a langchain bot about wikidata ids for specific places, links to the page, to read it and then to answer facts about places and got very good results instead of made up numbers.

Wikidata links to FIPS codes, OSM ids, GeoNames and that gives us an opening to link against the cool datasets from Flickr, Foursquare and others who have created gazetteers.

To me, Semantic Web was dead on arrival because of its UX, but now a semi-smart agent can help us get past the UX problems and jump from plain text to json output.


I think it's a great tool but I think a generous interpretation of the OP is that it's solving problems that aren't necessary in the context of a better overall system/process.

A woodworking example is that a planer is great tool that helps you make nice flat surfaces. But, to a certain extent, it's a downstream fix that wouldn't be necessary if a carpenter was using a better overall process. I.e., if their upstream process for cutting/ripping wood made nice flat surfaces to begin with, the awesomeness of the planer becomes moot. (Apologies to the legitimate woodworkers if this analogy is off).

Where tools like GPT becomes invaluable is when you have no control over those upstream processes but still need to get the job done. But leveraging a tool for a downstream fix when upstream fixes are possible is usually a less-good approach to creating good systems.


You mean it "only" solves problems in the real world and not in the ideal world?


That's not what I mean, unless you assume that you have no control over other elements of the process.

To torture the woodworking analogy, your assumption is that the carpenter has no control over ripping the boards. In some instances that may very well be the case, but there will also the instances where the carpenter does have influence over creating the boards, or even wholesale control over ripping them. In those cases, using a planer to fix poorly ripped boards may not be the best approach.


How often do you actually have control over the entire process of anything? Even website development, unless you’re going to handle raw TCP sockets, you’re going to build on top of someone else’s tools one way or another. And in a business world, you almost always have to deal with other teams, other people, other priorities. Even when you run a company, not all of your employees and partners can always do things exactly the way you want. Having a flexible tool that works in the real world on real world data on top of real world processes is incredibly valuable.


I think you're engaging in some dichotomous thinking. I'm not making the claim you'd have to have "control over the entire process". What I'm cautioning against is just looking to the tail end of a process and assuming that's where you need to add leverage.

Even in your examples, yes, you have to work with other teams. "Control" doesn't mean you have dictatorial control over those teams. But it does sometimes mean you have build relationships, leverage what you can, and explain the value to those that do have some modicum of control. The idea that we just throw our hands up and jump to workarounds is often an excuse for taking the short-term easy at the expense of a better long-term solution.


I like your planer analogy, though it is indeed off for woodworkers, but in a way that’s rather nuanced, so it doesn’t inhibit getting the idea across to normal people who don’t know a lot about intricacies or working with wood.

There are a couple of reasons why it would be hard to change the upstream process to not necessitate planers. The main one is that logs are typically ripped into boards when the wood is still green, and in the process of drying, boards change shape and dimensions: they bow, cup, warp, and shrink, and you might still need a planer to bring them back to flatness and to desired final thickness.


I guess the only way to change it would be to rip them when they are green... dry them to a decently low moisture content and then plane them again. it's still a bit off since wood is never static and the differential in moisture between the shop and your shop can also change the wood. you really do need to dimension it after it has stabilized in your shop for a spell.


Nope.

Cost of implementing better process for all carpenters is significantly higher, than all carpenters still using bad process + _one_ AI being able to clean it for carpenter, plumber, translator, developer (you name it, you got it).

Not even entering laziness/corposlowness gardens etc.


This is an misunderstanding of how processes typically work for a couple of reasons. For one, it assumes the "cost of implementing better process for all carpenters is significantly higher". This may be the case for Tao, where he has limited control over the inputs, but probably not the case for the woodworking analogy, for a variety of reasons. You are essentially advocating for "rework" to fix problems which is considered a unnecessary waste in process design.

"Corposlowness" is just another name for "bad processes". It supports the claim rather than negates it. Using AI to overcome bad bureaucracy makes it a workaround, not an idealized process. What often happens when implementing workarounds rather than good processes is that the workaround can create bloat and waste of its own and overtime, not really fix the problem. Like hiring more administrators for a large organization, they can take on a life of their own, eventually becoming divorced from the problem they were intended to solve.

Again, I'm not saying that AI is misapplied in Tao's case. I'm just cautioning that it's not a panacea for bad processes. In many ways, it can be misused as a band-aid for bad processes, just like creating excess inventory is a band-aid for bad quality control.


you have better explained my own point, haha


i still don't get it. i only understand libraries of congress and cars.


> incredible what mental hoops people will jump through to disqualify AI!

There are high peaks and troughs in AI buzz right now.

Yes, on the one hand you've the but-can-it-dance crowd.

On the other hand, Terence Tao on GPT-4. I mean, I'm not weird for really expecting the story here either be about GPT4 helping Terence Tao on some difficult newfangled proof, or Tao talking about the math behind large language models. Instead this boils down to

GPT4 even does the work of some of the smartest mathematicians in the world[1]

1: by parsing some web pages and PDFs for their meetings


> GPT4 even does the work of some of the smartest mathematicians in the world[1]

> 1: by parsing some web pages and PDFs for their meetings

Like that old joke about the guy who impressed people by claiming he had helped a brilliant mathematician solve a problem that had stumped him. And the punchline is something like "yeah, and it only took me a few minutes, all I had to do was replace his timing belt."


The insight I think is that if smart mathematicians can use it to do drudge work they then have more time available to do smart maths.


I wonder why he doesn't have a secretary. Isn't he like one of the top mathematicians in the world? Why the heck are they having him do drudge work?


Profs don't get secretaries. There's usually one admin assistant for the entire dept. and they're busy with applications for admissions, scholarships, grants, etc.


Or grad students, for that matter. The notion that Tao should have to monkey around with Excel in person is just bonkers.


He'd be a seriously bad professor if he wasted the time of the people he was supposed to be teaching with copying badly formatted data for a talk.


mathematicians do not do math all day. there is probably a point of diminishing returns where more doing more math is not useful.


The first time you copy and paste your disorganized data into ChatGPT and then into a spreadsheet, it's fun because you know how much longer it took before.

The hundredth time you do it you're going to be like "why is this so f'in annoying still."

Today's interface to language models is subpar for a lot of applications. Lots of room to improve that. A tool can be both amazing and still be just another step on the road to something truly seamless - just like how now it's "tedious" to use the computer for it instead of mailing/faxing paper forms around and filling out tables by pen and pencil.


> the reality is that a lot of things have a terrible UI with unorganized data. that's why this tool is so amazing - because it doesn't matter anymore.

how naive. how do you know it's right? ah, you have to manually do the calculation anyway to confirm, this is what Tao ended up saying in a reply asking as much.

AI is great, but it's not a silver bullet, since its correctness can never be 100% under the current LLM framework.


How naive to think regularly structured spreadsheets are the answer to this problem. There is a famous joke about "all your budgeting spreadsheets being useless when you discover you have a formula error in one cell that propagates throughout the whole spreadsheet".

You have to do those same confirmation calculations anyway when you use a spreadsheet. In my experience the utility of something like what ChatGPT can do is still unparalleled.


I'm convinced people like you are in for such a rude awakening.

I don't understand how you can hold this position with AI considering it's only the beginning.


Ensuring correctness by a human is existential to the work itself, not a mere annoyance. Yes AI can do it, but the more work you give to the AI, the less control you will have on the quality of the work.

This AI future you're wanting is an Idiocracy-like world where nobody knows how anything works and everything is in decay.


> considering it's only the beginning.

Some of us are just burned out on the hype cycles and prefer not to count our chicks before they hatch.


I think we're all in for a rude awakening, regarding what societal changes AI actual makes, but that's neither here nor there.

You're making the same mistake you're accusing them of making - assuming to know the future at the beginning. You're assuming that fixing these issues will prove to be trivial or at least inevitable. Sure, recent progress has been swift, but if you recall, it was damn-near stagnant for decades. Some were even claiming we were in the middle of an "AI Winter" and could not see the spring!

Based on all currently-available evidence, the current techniques that we utilize for generative AI are unreliable, in terms of accuracy of derived facts. It will require either a different or complementary approach to iron that out, or we're going to have to start seeing some _very interesting_ emergent properties from scaling higher. This stuff could show up tomorrow, or it might never show up at all! But the _current_ LLM framework does not look like it can do what we're looking for here, not reliably, certainly not 100% reliably.


ChatGPT is impressive, but I do not think it is a silver bullet, even for data cleansing and processing tasks. I would use your words and say people who think (that ChatGPT solves the problem of data cleansing once and for all) so are the ones who will be in for a rude awakening.


I can't wait for 3d printers to take over manufacturing or self driving cars to self drive. Still waiting for the metaverse. Or cryto to take over banking. We are not even at those points for AI. You might see the future.. let's get a 3d printer in every home first.


What’s the use case of a 3D printer in every home? Why is that even relevant in comparaison of how useful AI can be? The use cases are already there, today.


Ask those people who rode the hype wave 10 years ago. They had grand plans to make Star Trek replicators happen a few centuries before they were predicted.

I suspect the satire whooshed over your head.


Can’t wait for the internet, web, cell phones, or social networks to change society… AI is closer to these than the others.


AI only has to be as reliable as a human. The task is ultimately trivial and clearly in the capability of GPT-4. It is easy to statistically verify if GPT-4 has higher or equal correctness than a human. Of course Tao did not do this here, but it will be done.


Right but at the very least we can strive for consistency and efficiency, if only to save cost and energy. ChatGPT may save time, but not energy or money, assuming your idea is to just totally rely on LLM's for data processing (which is why it "doesn't matter" for you anymore).

If somehow magically from the beginning of computing we had a natural language interface to a computer's operations, we would still have arrived at particular standards/specs for data formats. There would still be something like xml/json/csv. Indeed, I'd wager there would still at a certain point be some kind of high level programming (or otherwise formal) language adopted, to answer the particular clumsiness of natural language itself [1].

Putting aside any issues of reliability, its simply not sustainable (economically, environmentally) to put all our work into this stuff. Even if it does shine with one off stuff like this.

1. https://www.cs.utexas.edu/~EWD/transcriptions/EWD06xx/EWD667...


And to the extent that it does matter it has already been shown that in many areas the best UI is simple human readable text.


AI is the better UX! I've been using ChatGPT to do all sorts of things--proofreading, generating "diffs" on unstructured documents, getting text web-ready by substituting HTML-entities--that I already have tools for but use so infrequently that I waste time looking up exactly how I need to specify the request. Using ChatGPT, I can just use English. It's amazing. And frustrating, because now I want this built in to everything, and it's not ... yet.


And AI can be used to make improvements in existing UI, like it's just so damn stupid.

If you don't like this tool, don't use it! If you like, it, use it!


GPT = new UI more conversational and easier transformation to commands

I think a lot of people miss that due to being shown in search instances and companion apps.


> a lot of things have a terrible UI with unorganized data. that's why this tool is so amazing - because it doesn't matter anymore.

Do you imagine a future where machines output data that can be barely read by humans, but can only managed through the help of AI? Honest question.


This is an advertisement for something like the semantic web, not AI or language models.


And let me know when the semantic web becomes successful after 2+ decades trying. The semantic web was always pretty much doomed to failure because it imposed a large cost on the content creator, who themselves get little benefit out of the structure.


The bigger reason why it was a failure is not because it imposes a cost in terms of work on the content creator, it imposes a lack of control on the part of the party that considers themselves the 'owner' of the data. If every webpage that right now is plastered with ads and various junk to get you to subscribe/form a relationship/amplify the content would just be the structured data and a default way to render that data would be present in every browser then most content monetization strategies would fail. Follow the money, not the effort.


Nail in the head. But can you imagine what it would have been if hakia would have been a thing instead of the SEO-spam, ad-infested Web that Google and Co. gave us?


That's a hard question. I really have no idea, but I would have loved a browser that takes in structured data and displays it in a way that I control any day over the junk that the web has become.


> becomes successful after 2+ decades trying

Rich pro-ai argument.


Fair point. Still, the semantic web is dead because we already solved the problem with a better solution, which is open APIs.

The idea that everything would work great if only all of our data was structured and easily parseable everywhere just leads me to ask "Do you not interact with humans on a regular basis?"


Actually I know a guy who've been working on semantic web for like two decades and his project is finally gaining some traction right now precisely because he can now leverage AI to sort it out and turbocharge his application.


I'm very skeptical. It seems to go against the grain.


What do you mean? "Semantic web" has been around since XML was touted as the data silver bullet, promising structured and semantically organised data 25 years ago. Countless startups tried : XML databases, digital asset management, countless o-so solutions... Until today.


Not to mention elimination of pointless work.

I needed to gather various statistics on the speakers at the previous ICM

Why did this work need to be done? Are even the people who say they want this data actually going to use it for something productive? Is there something revelational in this data?


He was the chairman of the ICM structure committee. Having stats on how previous ICMs were organized is obviously useful to him. It's not "people who want this data" who ordered him to gather it (not that anyone could order Tao anything), he wanted the data to use it himself. He had the choice between sifting manually through dozens of webpages or automating the task.


> Ironically, Tao's post convinces me that AI, though amazing, isn't really the solution. Better UX and data quality is. Why was the data so disjoint to begin with?

Because the world is not a database. His source were formats meant for Human consumption, not machines. This will never change, people are lazy, greedy, or fear "data leaks", so we will never have a machine-first-format that everyone will use.

I mean, that guy is using latex, others have markdown, or org-mode, or Microsoft Word, those all are not meant for easy scripting. This is why AI will be such a dealbreaker, because it will close this gap, make human-formats to machine-formats when neccessary.


I find the "human consumption" here funny, because it's the "human" part which led the author to turn to machines which can do it better.


Well, yes, this is kinda the whole point of the argument. That's what LLMs bring to the table - machines that can better deal with data and documents meant for human consumption.


> Ironically, Tao's post convinces me that AI, though amazing, isn't really the solution. Better UX and data quality is.

Umm, and I want a pony? The world doesn't come in perfectly structured data formats. I was pretty amazed when I pasted in some of my medical lab test results and ChatGPT was accurately able to parse everything out and summarize my data. It worked extremely well.


> The world doesn't come in perfectly structured data formats.

In Tao's case, the data was organized already, simply not given to him.

In any case, Tao, per the post, had to manually calculate to confirm that GPT-4 was correct in any case (he implies as much in a comment asking if he checked for correctness).


It doesn't matter what the data format was originally in - what mattered is that Tao didn't have access to that format.

I'm sure the company that did my medical labs also has my data in a structured format somewhere. So what, should I call them up and demand they give me my data in a spreadsheet? I can go down that useless path, or I can paste my data on ChatGPT and get results in 15 seconds.


> I'm sure the company that did my medical labs also has my data in a structured format somewhere. So what, should I call them up and demand they give me my data in a spreadsheet? I can go down that useless path, or I can paste my data on ChatGPT and get results in 15 seconds.

Yes, go ahead and send OpenAI all of your HIPAA data.


Lol, the good ol' "HIPAA bogeyman" strikes again.

It's my data. The entire point of HIPAA is that I own the data and I can send it to whomever I want if I decide to. I get value out of it, others may not choose to do it, that's their right. But I'm pretty sure sharing my CBC results is not going to be the death of me.


Checking a table for correctness and filling up a table are not the same thing. I also check for correctness when I manually count entries. It's an extremely error prone task.


How exactly could you check for the correctness without effectively finding out the answer? I mean this discrepancy is fundamentally the problem of P = NP or not.


NP is the class of problems for which it's hard to find the answer but easy to verify. The fact such a class even exists should hint to you that finding the answer and checking correctness can be two different things.


the point of my comment is that whether P = NP to begin with is still to be determined. in the case of ChatGPT, it's easy for it to respond, but how difficult is it for you to verify that it was correct, or not? how does that compare with the difficulty in doing the task to begin with?


Lets take an example from Demo of =GPT3() as a spreadsheet feature - https://twitter.com/shubroski/status/1587136794797244417 /// https://news.ycombinator.com/item?id=33411748

Consider the time it would take to do manual entry for each of those examples compared with the time it takes to verify that the generated content is correct.

"Are these the correct state and zip codes?" is much faster than typing it by hand. You just ask yourself "is MA the correct state code for Massachusetts? Yep, next" rather than "Massachusetts that's... (lookup) MA, type MA; next that one is MI, already a state code ..." and down the list.

I would be willing to content that GPT will do the list faster and with better accuracy than a human doing the same work (that would also need to be checked for correctness).


>I was pretty amazed when I pasted in some of my medical lab test results and ChatGPT was accurately able to parse everything out and summarize my data

I think an important question is how much faith we give in the answer (especially with medical data!). There are lots of examples of great uses but also a number of examples of hallucinations and just plain bad summaries. When the stakes are high, the conviction needs to be couched in the risk of it being wrong. We need to be cognizant of the automation-trust factor that sometimes makes us place unwarranted trust in these systems.


> Ironically, Tao's post convinces me that AI, though amazing, isn't really the solution. Better UX and data quality is. Why was the data so disjoint to begin with? Why is Latex so hard to work with?

> In this case GPT-4 is used to solve a problem that shouldn't have even been one to begin with.

Friction. These problems shouldn’t exist, but they do anyway, and they’re everywhere. Anything human is inevitably going to be imperfect and messy to some greater or lesser degree, introducing friction into dealing with it. Especially as we produce more and more unstructured or semi-structured data, which AI is particularly good at wrangling. If AI can help us cut through that friction significantly faster and more accurately, that’s a win.


Let's just hope that someone doesn't task the AGI with eliminating friction and it realizes that humans are the problem.


Ironically, interacting with ChatGPT still comes with a lot of human-like friction and imperfection.


IME most of that is of the form "I'm sorry, but as a large language model I..." and seems to have been drilled into it with training to dodge tricky subjects.


> Ironically, Tao's post convinces me that AI, though amazing, isn't really the solution. Better UX and data quality is. Why was the data so disjoint to begin with? Why is Latex so hard to work with?

I’m fairly convinced that for many problems (including this), AI is not the best solution but will become the preferred solution. “Best” doesnt get in the way of “good enough” when convenience is at stake (and making a good UX is often very very inconvenient)


I think it's not the case of AI or ChatGPT is the best solution because it can be a local optima. But it's currently the "best" solution available to the average Joe. The fact in this OP case study the average Joe is probably one of the best mathematicians of our time, makes the situation much more interesting and intriguing.

I'm in the middle of reading Steve Jobs' very own words, Make Something Wonderful book. Apparently throughout his whole life he frequently mentioned and discussed a lot about providing and creating easy access to the computer because ultimately people do not want to program but want to use computer instead. He did mention about how Morse code was not popular as much of the later telephone technology because people refused to learn and use the unintuitive Morse code even though it only takes 40 hours for average person to learn the entire Morse code. As a trained communication engineers I can very much relate to this because even though we learnt much of the underlying technology of communication that enable the Morse code or the telephone to work across the Atlantic ocean most of us can't be bothered to learn the Morse code and use it if we can help it because the telephone make it redundant and it's counter intuitive to use.

Granted, now we know that telephone or circuit switching in general is a suboptimal in providing and solving the human communication problems and needs. As of now the Internet, packet switching and the multimedia approaches are probably the best solutions but telephone has served us a potent and still one of the best solutions for communication until very recently.


Solving for every last mile problem in UX is hard though. A lot of AI value is being a flexible, generic interface to software automation.


We can probably even think of AI as a generalized meta-interface to everything - there’s probably no better UI/UX than natural human language that we’ve all been speaking since childhood.


I don’t know that it’s a better UX, but it is a viable UX, and that is very valuable


And here’s an exchange about how he validated the data was correct - which to me seems like the most important part:

> Out of curiosity, are you sure that GPT did it correctly? If yes, is it because you were able to "spot check" it in a few places? Or have you used GPT enough that you trust it with a task like this? Or is this for some kind of internal use where a few small errors is unimportant, and you only need the broad strokes to be correct?

> Yes, yes, and yes. (There were some obvious checksums, for instance by computing the total number of speakers several different ways.)

Still cool, but the circumstances above aren’t always true. In fact, in a lot of meaningful work, none of them probably are.

If GPT is like an assistant whose work you always have to double check in order to be sure of the validity of the results, that seems to be a pretty big caveat to me. In that case it’s probably better to just write a script whose output you know to be deterministic, even if it takes slightly longer (and the bigger your dataset to verify, the more likely it is that it’ll take less time to write a script than to validate results one by one).

The scary part is if/when companies/bureaucracies/governments/etc start using GPT for all sorts of Important Tasks and skip the validation part because they assume the machine will always get it right.


> If GPT is like an assistant whose work you always have to double check in order to be sure of the validity of the results, that seems to be a pretty big caveat to me.

Sounds like every junior developer fresh out of college I've had to work with. Interns are even worse. Even senior devs mess up from time to time. That's why we do code review and other QA. Trust no one.


True, but one obvious difference is that a senior contributor can typically validate the work of a more junior one in less time than it took them to produce, that’s why the overall equation works. And if you decide to keep the more junior person, it’s probably because you feel like they’re improving and becoming more trustworthy over time.

Another key difference is that a deterministic script that has determined to be valid will be valid for all subsequent runs, regardless of input size. That is not the case for ChatGPT, which might be correct on some runs and not others, or do well on small input sizes but start messing up when thousands (or more) of data items are involved.

Again, I still think ChatGPT is really cool, but thinking about all these nuances about the nature of the work that is actually being done when you use ChatGPT vs writing a script vs handing it off to an intern seems crucial to the debate.


HackerNews' hug of death hitting a site with normally 4k active users. Not that surprising...


it's plain text. it can be cached. it's embarrassing.


Feel free to share your know-how https://github.com/mastodon/mastodon


Mastodon's architecture inherently does not prioritize speed of first page load. Any PR would be rejected :D


I thought what and when to cache was one of the hardest problems in computer science.


Something like Mastodon is likely 90% reads, or even more. How often is the OP editing their post? When they post, cache it. Kind of like what I did by including it in my original post. Caching is a hard problem because sometimes invalidation results in serious consequences, therefore knowing when to invalidate becomes potentially intractable. In this case there would be few, if any.


it was cache invalidation, and naming things. Just throwing stuff in cache is easy.


Why would it be difficult to cache static text?


Perhaps they've optimised for the normal levels of traffic they receive, not unanticpated spikes. That's hardly an embarrassing choice.


I'm not talking about the administrator of this instance, I'm talking about how mastodon is fundamentally designed. Even an iPhone could probably serve the (static) text of Tao's post to a million people on a 5G connection. Computers are fast.


I run Mastodon on a $5 VM. If I hit HN homepage, my site will probably be slow as well. I don’t see how that reflects on Mastodon though.


what am I saying is that there's a way to design a site on a $5 VM that serves static text that wouldn't be slow even if it was #1 on HN.

if you disagree with that then we'll have to agree to disagree.


Wow I didn't know Mastodon failed that massively in converting people from Twitter. Then again, maybe it's not that surprising.


That's not how the fediverse works, Mathstodon is just one of many servers in the network (running the Mastodon software).


yes, and unfortunately almost all of the servers are unable to cope with even a moderate amount of traffic

Twitter is a far better user experience in that respect, to the point that I actively avoid clicking mastadon links now because they fail so often


The joys of federation


> Why was the data so disjoint to begin with?

Because real life is messy.

> Why is Latex so hard to work with?

Because people that use latex want to be able to brag they use latex

(just joking for the second one.... unless ?)


> Because real life is messy.

This doesn't make sense - the data inherently is centralized. Tao implied as much.


>Better UX and data quality is. Why was the data so disjoint to begin with? Why is Latex so hard to work with?

Do you mean build better, specific, UX and normalize data for only this task? Why spend that time, even as a developer of whatever improved app/system you're thinking of, when he can just turn to AI?

I'd rather lean on GPT4's ability to generalize highly specific technical work rather than ruminate on each individual app's lackings and how it could be better than using AI if we just... put more effort into it?


> AI, though amazing, isn't really the solution. Better UX and data quality is.

In this case, AI was the tool used to have a better UX. People build something with use-case X in mind, and use the best tool for that job. Piping that into use-case Y requires some duct tape and plumbing. It turns out that AI is great at that sort of repurposing.

It's great if someone using the best tool for their job is nearly the best tool for the flexible infinity of other use cases, which is what flexible enough duct-tape gets us.


> wow, the site is so slow. getting "This resource could not be found" errors on reloads.

Oh dear. Federation at its finest /s

If the instance is too small, it will easily fall over under heavy traffic. If it is too big, it is highly centralized and even then it cannot scale at a maximum of 300,000+ users at the same time.

Eventually, they would all be re-centralizing back to Cloudflare. Once that goes down, all of the biggest Mastodon instances go down at once.

As for GPT-4 in mathematics (LLMs in particular), in goes inline on what I said before. Fundamentally, these LLMs cannot reason transparently, nor can it directly bring up its own mathematical proofs with thorough explanations that allows mathematicians to work with. Only low hanging fruit of summarization of existing text.

If Mr. Tao can see its limitations, surely it puts the hype of unexplainable magical black-box LLMs to rest as being great sophists and bullshit generators.


I would not discount the potential of LLM's to support research level mathematics. There's interesting research going into combining the LEAN theorem prover with LLMs, e.g., this recent paper: https://arxiv.org/abs/2202.01344

"... automatically solving multiple challenging problems drawn from high school olympiads."

This combination of GPT and LEAN is the first thing that has finally got me really interested in the potential of LEAN. And LEAN very directly addresses the "bullshit" issue, since it formally verifies whether or not a proof is correct (it's like a compiler like Rust, but for mathematical proofs).


Better source data is indeed a better solution, but the source data is often not under one's control.


> In this case GPT-4 is used to solve a problem that shouldn't have even been one to begin with.

But the problem DID exist, and problems like that are likely to continue to exist for the foreseeable future.



They should have done it right the first time? That's not the world we live in.

Coworkers, associates, clients, family - everyone sends us funky data. Sometimes it's a PDF of text instead of a text file, or a picture of their computer screen with a paragraph in it, or copy-pasted text with unwanted formatting, etc, etc.

And the point of this article wasn't really about the tedious job of copy/pasting, it was that once the data was ready, instead of hunkering down and having to work through a problem with an unknown number of variables, GPT-4 can drop an answer in seconds flat. Maybe a few more if you have to tweak your prompt a few times.

Part 2 here is that we have a problem to solve, and that's how data is sent and received, and that it all needs to end up in a structural format that's usable by our AI tools.

I am thinking that writing a script that can dump textual data from literally anything I'm looking at so it's AI-ready makes a lot of sense, then we have step 1 taken care of as well for my own personal GPT-solvable situations.


> Ironically, Tao's post convinces me that AI, though amazing, isn't really the solution. Better UX and data quality is. Why was the data so disjoint to begin with? Why is Latex so hard to work with?

AI is just that. A UI/UX for any/your data.


Yeah, I waited a good minute of it trying to load before I gave up.

Anyone that got in can share what it says?


I accessed the message through a mastodon client instead. They are a lot better at federation than the various instance sites IMO, a more consistent interface when there is no option of sending you out to a non home instance page.


I didn’t even know there were mastodon clients, though I should have assumed. Know and good ones for Linux and Mac?


I use Tusky/Android and that's the limit of my knowledge


Check the comment to which you replied (maybe it was added after)


And how does he know the results are correct without verifying each and every entry?


I find this a real concern when it comes to the proposition of "replace every programmer with GPT".

I think the reality is, there are cases where it matters, and cases where it doesn't matter (at least, not very much).

I work on aerospace software systems. I suspect that most software developers would be surprised at the process and rigor involved. That does not mean that humans are flawless; of course not. But there are going to be a lot more hurdles to replacing all of the humans with non-deterministic machines.

There are numerous software components on a modern airplane. How good is good enough? If the overall system works correctly 99% of the time, is that good enough? There are about 16 million flights annually in the United States alone. If 160,000 of them crash, is that good enough?

I mean, like, if Microsoft Word failed to save your document 99% of the time, you might be irked. If Professor Tao's data was only 99% accurate, he probably still has a reasonably good picture of the information he needs.

Regardless of if a human or a machine is writing the code, there are different levels of software criticality. As humans, we don't apply avionics software development process to iOS games because it's not worth the extra time and expense. Likewise, there will be a spectrum of where it makes good enough sense to automate tasks with AI and where it doesn't.

Which is not to dismiss AI! It's amazingly useful technology. But there really are places where above average accuracy is needed, and AI that works great for handling some tasks might not suffice for other tasks. If you truly need all of the results correct, then a nondeterministic neural network is probably not the best path to be on. There are other methods; even other automated methods.


> I work on aerospace software systems. I suspect that most software developers would be surprised at the process and rigor involved.

I did this once, a fuel estimation program for 747's, the degree of understanding of the problem required to create something that would pass review was off the scale compared to any other program that I've ever written. It is also the only time that the customer wanted to get it right rather than that they were looking at what it cost (because the savings would pay for the development many times over this was probably less of an issue anyway). I'd love to be able to always work like that.


Imagine Terrence Tao manually entering data into spreadsheets!!


Wait.... is it possible to tell chat gpt to grab data from a page?


And AI is a way to enable better UX and cleaner data...


Yeah bold prediction I know but I think basically everyone will move back to twitter by the end of the year


> Yeah bold prediction I know but I think basically everyone will move back to twitter by the end of the year

You do know Mastodon is a federation of servers, right? I predict most people will move off this particular one to a faster one.


Sure, but on twitter you don't have to think about which federation you are on




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: