Hacker News new | past | comments | ask | show | jobs | submit login
ChatGPT is a bullshit generator but it can still be amazingly useful (aisnakeoil.substack.com)
193 points by 1vuio0pswjnm7 on Feb 1, 2023 | hide | past | favorite | 154 comments



In other words, it disrupts classroom assignments where the student is being asked to produce bullshit but where historically the easiest way to produce that bullshit was via a process that was supposedly valuable to the student's education. The extent to which a teacher cannot distinguish a human-generated satisfactory essay from an essay generated by a bullshit generator is by definition precisely the extent to which the assignment is asking the student to generate bullshit. This will certainly require a lot of reworking of these traditional curriculums that consist heavily of asking the student to generate bullshit, but maybe that wasn't the best way to educate students this whole time.


The point of school work isn't to generate something of value externally, it's to generate understanding in the student. Much like lifting weights or running are "bullshit" activities. You haven't seen bullshit until you see whatever is produced by a society with a voracious appetite that runs on "gimme what I want" buttons and a dismal foundation of understanding.

The students bullshitting their way through course work are like the lifters using bad form to make the mechanical goal of moving the weight easier in the short term. They completely miss the point.


That's exactly what I'm saying. It's reasonable for teachers to give assignments where the resulting artifact is not itself of direct value to the subject matter, but where the only feasible way for the student to produce the artifact is via a process that is supposed to be of educational value. I'm just saying that when new ways emerge to produce that artifact which don't have the same supposed educational value to the student, the assignment itself needs to change.

I don't think your analogy to lifting weights or running is appropriate. A more appropriate analogy would be a physical education teacher who says "your assignment is to meet me on the other side of town in 30 minutes" because the only feasible way to get there is to run. Obviously the point is to get the physical activity, not to be on the other side of town. Then the bicycle gets invented, and suddenly it requires much less strenuous activity to get across town. All I'm saying is that it's vital for the teacher to realize that being on the other side of town was never actually the point, and to change the assignment so that students still get the intended amount of physical activity.


It's more like "your assignment is to run to other side of town in 30 minutes" but due to resource constraints, it's not easy or possible to monitor you the entire way. Its also not possible to redesign the task without constant monitoring of everyone.


I would argue that if the student doesn't understand that the value of the exercise is to run, and they take their car instead, then something needs to be explained to them.

There are two ways: you can try to design exercises that students cannot abuse, or you can try to have the students actually understand why they should do the exercise and not abuse it.


Students are often thinking of how to maximise their grades and efficiently use their time in the short term rather than how to best learn the subject matter. Anyone who has had to deal with assignment deadlines can understand why a student may choose to take such a shortcut if it's available, even to their own detriment in the long term.


Implying anything learned in high school is of benefit long term… ha.

Someone recently described their high school experience as “slightly better than prison.” I agreed. If you want to fix education, perhaps start at the root of where we force kids to do bullshit work.

Teaching them how to do their taxes would be infinitely more valuable than memorizing pointless geographical locations (solved by google maps) or names of dead white people in history class (also solved by google) or forcing them to learn Spanish (google translate). The experience was so unpleasant that I didn’t realize until my 20’s that I actually love studying history.


I don't completely agree. I mean it should not feel like prison, and putting pressure with grades is not necessarily good.

But I don't think that school is about learning useful stuff. There is plenty of time to learn a boring job/how to do taxes after school. School should be an opportunity to learn how to learn, and to discover new things.

Because you don't need it in your job later doesn't mean it's bullshit.


Sure. I'm honestly not a big fan of grades, to be honest. But that's a hard problem: some students will be better with grades, some without, and some won't care.

Still I don't believe that the goal in school is to learn useful stuff. The goal is to learn how to learn, and to discover stuff. There is plenty of time then to learn how to do a boring job.


>It's reasonable for teachers to give assignments where the resulting artifact is not itself of direct value to the subject matter, but where the only feasible way for the student to produce the artifact is via a process that is supposed to be of educational value.

Reminds of this scene from A River Runs Through It.

https://www.youtube.com/watch?v=gA-sEfXOaEQ&ab_channel=Grady...


I suppose a human teacher not being able to come up with an assignment not easily solvable by computers is a new level of the Turing test.

Therefore the assignment itself must include a requirement not to use a computer.


Agreed, relying on ChatGPT to do work is like using Chegg for answers, we can tell ourselves, "oh yea, I could write that no problem, or I can just do the same method later on..." until you can't.

It's called learning without a crutch, aka learning. (obviously I'll do things like use CMake on large projects, but I've forced myself to write Makefiles by hand for exactly this reason).


But most people today aren’t learning the times tables, and long division. They just use a calculator.

Most young people don’t know how to write by hand anymore, cursive or whatever. They rely on spellcheck and grammar checks. I’m talking native US citizens

So why would they know how to compose an essay? They just text short form things. Even “important” business people fire off misspelled emails with bad grammar as a status symbol, so now ChatGPT will be the status symbol. Putting effort into making a properly worded email actually shows lower status in business negotiations — because if you can’t be bothered to spend time, you are a bigshot with many deals and can walk away at any moment.


"But most people today aren’t learning the times tables, and long division. They just use a calculator."

Let me fix that for you:

Most people learn the times table, and long division. Then they just use a calculator.

The American 5 paragraph essay is an extraordinary artificial academic construct, completely divorced from the real world. It aims to be systematically taught and systematically graded, during the developing stages of education where it is often difficult to get students to express any written thoughts at all.

Using it outside of that context is a mistake. Composing a real-world essay is an exercise in considering an audience, preparing suitable arguments, making points more or less firmly, and hopefully providing some value.


I can learn long-form division by writing with a pencil. Is the pencil a crutch? Do I have to make my own pencil? Should I learn to divide without needing a writing utensil?

Of course, we can't take ChatGPT for granted like a pencil, because it's controlled by a third party and requires resources out of reach (for now). But it might become trivially cheap one day.


Simply, yes, it is a crutch for your mental math skills.

Should you do any of those follow ups? Seems like a judgement call. In my experience strengthening those fundamentals is important, but there are many occaisons where the point of the exercise (such as learning to write legible long form algebra) overshadows other possible benefits - making a crutch such as a pencil and paper, or a computer typing system, the appropriate call to make.

Please, when you pose 'absurd' hypotheticals, follow them to likely or reasonable conclusions. If you shit your brain off when the conclusion seems obvious but before evaluating you will frequently end up just asking reasonable (if somewhat obvious) questions.


There's clearly some limit to this line of thought though. Maybe it's valuable to know how to write your own make file. But have you built your own microprocessor by soldering together transistors? You have?! That's awesome! Have you built your own soldering iron?


Indeed! If you wish to make an apple pie from scratch, you must first invent the universe.


Yes, thank you! That is what the latter half of the first paragraph is about - we make judgement calls about the tradeoffs involved with a crutch all the time.

The point isnt to accept a heuristic that points one way or another, but to be conscious about it being a choice that we can make that can have various trade offs along a spectrum. As usual figuring out what trade offs to make is highly dependent on what your goals are.


The point I was aiming for was that learning with a crutch is still learning. I interpreted Avicebron's comment as "real programmers code in binary" and you are "unworthy" if you can build a CMake system without having written a Makefile manually in your life.

But after your comment, a lot of what I see around me are crutches. The keyboard I type on, the heated building I'm in. They really help me perform "normal" functions, and living without them would be a handicap, so they are crutches.

But I see no point in purposely living without them. Live your life, and don't force yourself to do something you don't like unnecessarily.


Yes, they are! Nothing wrong with that. A crutch is a tool, and like other tools we will have varying levels of dependence on it/them.

Where the line is for necessary/unnecessary dependence lies is a judgement call.


Plato famously thought writing was a crutch, because people no longer had to memorize everything they learn.

Why bother with the effort of learning anything if you can just be lazy and look it up in a book!


I'm getting pretty burned out on hearing this analogy. It seems to stem from a basic confusion between tools and toys. Tools don't constrain your use by dictating output; they help you speed up solving new problems. Toys create environments that shoehorn you into their prefabricated set of solutions. The feeling of gratification from making something nice with a toy is a placebo for the gratification of solving a real problem, and a big part of the drag we're experiencing on creative problem solving these days is the replacement of free-form solving with puzzle-solving. Puzzles, like AI prompts, have solutions that exist in advance in a latent space, and their discovery promotes a form of gamification but doesn't ever till any fresh soil. Being great at playing a game can never approximate being halfway good at writing a game. Discovering Will Wright's philosophy embedded in the best way to play SimCity should only be a step on the road to writing your own SimCity, not an end in itself. SimCity is a toy, not a tool, despite its being a fun stepping stone for future urban planners. One would never want urban planners to rely on it to make their decisions, because deep assumptions are baked into it. Same with ChatGPT. Many, many assumptions based on a snapshot of the world are baked into its output.

Whereas a pencil has no assumptions whatsoever, and no bias towards what you'll do with it, and offers no suggestions that might funnel you in some direction. That is what good tools are like.

[edit] and I mean, how often do we hear that a tool or framework is too opinionated? What's meant by that is that the tool constrains your ability to solve problems with it based on its pre-baked gestalt. Extrapolating the inherent limiting problem with that to all writing leads to some extremely weird outcomes like whether a student's poor attempt at analyzing a text shouldn't get at least as much credit as the student who fed it into a neural net they neither care about nor comprehend.


> can learn long-form division by writing with a pencil. Is the pencil a crutch?

Yes. When was the last time you used that algorithm versus breaking down the problem in your head?


Strongly disagree. This is much more analogous to math students using a calculator rather than a slide rule. Yes, it’s faster. No, you aren’t going to pass unless you know the material well enough to validate the outputs.

I’m old. I’ve lived this moral panic so many times. I remember when social media, newsgroups, the internet, word processors, computers, graphing calculators, calculator watches, BBSes, and calculators were all going to destroy our children’s education.

I am highly skeptical that this time is any different.


By using some ML model to spit out text for you, will you learn how to write well? How to write a good essay? None of that. During the act of writing those, the student is supposed to put in effort and thing of well worded phrases, using the parts of their brain, which are responsible for language processing. Without actual practice, the exam will come and the student will fail. They wont have a bullshit generator at hand, that generates speech, which sufficient sounds meaningful, when they are in an actual discussion "over the table". Even if they did, who would listen, if all they do is get out their phone and read some generated bullshit from it, that does not follow along with what was previously said?

Being able to have a good discussion is a life-long valuable skill. Not practicing it will cost sooner or later.


I don't understand your point. If creating understanding is the point, why don't we test for understanding having been created but instead test for "bullshit activities" being successfully performed?

In your weight lifting analogy: why don't we test for outcome based fitness metrics (VO2max for runners, 1RM for weight lifters) and let students learn to choose the best ways to achieve those goals?


You seem to have solved the fundamental challenge with examinations of any kind. Please elaborate, teachers that for millennia have settled with compromises are eager to learn about your watertight solution.


Yeah I guess it's too hard for teachers to actually interact with their students to get an idea of their understanding of a subject. Let's do standardized multiple choice tests that only require you to learn how to pass tests. Who cares about the actual knowledge, right?


How do you test that a kindergartener understands addition without asking them for things a calculator can do?

Or do we just think it’s not important for people to understand addition at this point?


You give them a bag of marbles and tell them to show you 5 marbles and 3 marbles and then show you how much there are when you put them all together. That's how I taught my kids addition.


Your solution is to take away tools and ask to do a task under constant supervision? People take exams like this every day.


You are missing the parent comment's point. Your question can also be answered using a calculator.


Don't tell my kids.


The point, I believe, is that sometimes generating bullshit requires some understanding, so it is a worthwhile exercise. Just like running: you end up where you started, so it may seem that you did nothing. But obviously you exercised.

Asking a student to write an essay is similar: it is an exercise. Of course that essay won't be published or read by anyone else, but that was not the point. Just like you were not running somewhere, you were just running for running.


> why don't we test for understanding having been created

What it means for understanding to have been created is a fascinatingly complex problem in philosophy: https://plato.stanford.edu/entries/understanding/

If you have an straightforward test for this, you may wish to consider contributing to the field.


For athletes its bad form that will limit your ability to reach the top levels of your domain. We then test for the top levels of these domains, like the Olympics or Tour de France. We don't test VO₂ max since it is boring. In education we don't test understanding because it is hard to write those tests in many domains.


> The students bullshitting their way through course work are like the lifters using bad form to make the mechanical goal of moving the weight easier in the short term. They completely miss the point.

Eh, when your future career prospects hinge on what grades you get it's perfectly rational to optimize for getting those grades over learning "properly". Especially in subjects you know you're already competent enough in or that aren't relevant to your future goals.

It's stupid in weightlifting because nobody is going to pay you a higher salary if you can do 15 reps instead of 10.

(Although there's a related and opposite problem where people massively overestimate how important their highschool grades are and get unnecessarily stressed out over it.)


> Eh, when your future career prospects hinge on what grades you get it's perfectly rational to optimize for getting those grades over learning "properly".

I knew a few people who cheated and gamed the system at every opportunity through high school, all in the name of getting into the best universities.

They got absolutely crushed in college when their old tricks stopped working and they had to actually do the work. One of them got caught cheating and had a wake-up call when the university threatened him for it. He eventually put in the effort to grind it out and graduate a year late, but graduated nonetheless.

He's mentioned that he could have saved himself a lot of pain and suffering by just engaging in school the first time instead of cheating.

This idea that you can cheat your way into a job and then just coast doesn't really match reality. It may work for some people who get into weird jobs that don't fire people, but the idea that it's "optimal" to cheat your way through life until you're handed an easy job doesn't hold up in practice.


In your anecdote, the person who cheated got into a well-regarded university and graduated (albeit a year late). That's likely to have a better long-term outcome than not cheating at high school, and going to a lower-ranked school.

Cheating worked.


For the individual, maybe. For society, no.


Some of them didn't want to make the time commitment-- or partied. But a lot of smart students make then had old answers to tests especially in hard majors like the sciences. I think that counts as cheating.. they probably figured out they were giving canned answers or suspicious they arrived to answer right away.


I did this accidentally once at uni. Previous years tests were posted online (intentionally, by the person running the class) so I was doing those while studying to make sure I knew the material.

I sit down in the actual exam and lo and behold the questions in the test were exactly the same as one of the previous years I'd been using to study, I guess the professor got lazy.


The assumption here is that only grades will affect your lifetime earnings and other skills that you might improve through studying vs cheating won't come into play...

Some people can bullshit their way through life forever, but I've seen a lot of more bullshitters stuck doing entry level crap for a long time because their limits became apparent...


> The assumption here is that only grades will affect your lifetime earnings and other skills that you might improve through studying vs cheating won't come into play...

Or that perhaps you have a better plan for your own education than a cookie cutter curriculum that is intended to applied indiscriminately to millions of people and hasn't been updated in 50 years.


That's shifting the goalposts from the parent poster who was saying "all you need is the grade for your future career prospects," not "you are educating yourself independently."

Let's not pretend most people happy to cheat their way through to the eventual level of their mediocrity are doing so because they're doing a bunch of other learning.


It’s not a shift. I’m suggesting that both could be happening at the same time: you are maximizing your superficial credentials as easily as possible because credentials will help you and you are genuinely educating yourself because that will also help you.


In theory, maybe, but how many cheaters are undertaking their own great books curriculum in the copious spare time they're freeing up?


If you've written off yourself as "competent enough" or things "aren't relevant to future goals" you might want to think about the phrase "you don't know what you don't know"


I mean stuff like bullshitting your way through an english assignment when you're going to study computer science at uni.

You can be fairly confident that understanding the intricacies of Catcher in the Rye won't be important in any compsci paper or dev job, but the writing skills from the bullshitting are still useful.


Well, I for one hope I don't have to work with people, who only ever spent the minimal effort even including cheating on assignments, because their shortcomings will be what I have to deal with, because they cannot stem (or lift, ha!) their weight. But in reality it is highly unlikely, that one will never have to work with such people. I guess more realistically I should hope for not encountering too many of them.


Is asking ChatGPT to generate low quality code much different to the average npm dependency chain? It would seem disingenous to berate AI code completion for removing the challenges of development yet most reach for import * in the same breath..


> Much like lifting weights or running are "bullshit" activities.

> like the lifters using bad form to make the mechanical goal of moving the weight easier in the short term.

I don't know about you, but this would actually be the perfect excuse to do a sport like judo that you can't bullshit through but can still get the same end result from.


You do realise that in order to reach a good level in judo, you have to do some of the bullshit training like lifting weight, right?


I don't think so. I rarely lift weights but I do just fine in judo, largely because judo largely replaces it.


>I rarely lift weights

So, you do do "bullshit" exercise. What a weird hill to die on.


To get out of reading a long book -- if you want a TLDR or something. I see it as a tool to augment understanding -- bad actors will manipulate it to use to their advantage.

Teacher can't go around to each student to help with their understanding -- its not economical and each student is at a different level of ability.


I agree with some of this (particularly the conclusion that better education methods are required) but lets be a bit generous for a second.

The ability to write well is (or was) an important skill. Being able to use correct grammar, to structure even a simple argument, to incorporate sources to justify one's statements, etc. Even if we're just talking about the level of what GPT 3.5 is capable of, that still corresponds to, let's say, a college freshman level of writing.

Now, perhaps with the advent of LLMs, that's no longer true. Perhaps in the near future, the ability to generate coherent prose "by hand" will be thought of in the same way we think of someone who can do long division in their head: a neat party trick, but not applicable to any real-world use.

It isn't at all clear to me though that we're yet a the point where this tech is good enough that we're ready (as a society) to deprecate writing as a skill. And "writing bullshit" may in fact be a necessary element of practice for writing well. So it isn't self-evident that computers being able to write bullshit means that we shouldn't also expect humans of being able to write bullshit (at a minimum, hopefully they can go well beyond that.)


The teacher (or developer of the curriculum) needs to decide both 1) what they consider to be cheating and 2) how important it is to establish a certain level of confidence that students aren't cheating. If writing grammatically without any external help is deemed to be extremely important, and it's far too easy for students to cheat on take-home writing assignments, then suck it up and use some of your in-class time for writing exams. Your in-class time is finite anyway, so deciding how to use it is already critically important.

If electronic earpieces become commonly used by students to cheat on in-class exams, and that's important enough to the teacher, I guess they better figure out a way to screen for those devices.

If eventually nearly everyone has some sort of brain link that can whisper grammatical phrases directly into their brain and cannot feasibly be detected in the classroom, well, then either your entire curriculum is on the honor system, or maybe you'll need to update your thoughts about what tasks are important for your students to be able to perform without "external" help.


Oh, I agree. Optimistically, the existence of LLMs will be a forcing function for education to focus on what actually matters. (IMO, this is the ability to critically assess and generate arguments, something GPT isn't particularly good at (yet)).

But your example is actually rather striking. In most highschool and college classes for algebra and calculus, we're asking students to solve/prove problems that computers have been easily able to solve/prove, and have been for decades.

But the educational consensus is that being able to do algebra or differentiate/integrate by hand is valuable (up to a certain complexity.) Which is why calculators are not allowed (at certain levels) and why "show your work" is an important part of grading at every level.

Perhaps this whole discussion is a nothingburger once we figure out what the language arts equivalent to "show your work" is for text generation.


> Now, perhaps with the advent of LLMs, that's no longer true. Perhaps in the near future, the ability to generate coherent prose "by hand" will be thought of in the same way we think of someone who can do long division in their head: a neat party trick, but not applicable to any real-world use.

Except that AI would be hard pressed to become a part of the in-circles of waves of new generations of artists and writers, who create new literature in part through being part of new literary movements. In a way AI is closer to God than to people, in their omni-ness. Unless ChatGPT eats and craps and needs to find a job, find love, find friends, loses things, installs tinder, gets hurt, and has its life literally threatened (even just once in a lifetime--whatever that means for LLM), it cannot be a wholesome part of society and thus react to political change by creating new modes of writing. Because of that, AI will forever play catchup with the new tendencies in human literature and art.

Unless AIs dominate consumption and pay for their Netflix subscription, of course.


> Even if we're just talking about the level of what GPT 3.5 is capable of, that still corresponds to, let's say, a college freshman level of writing.

Indeed - the output of ChatGPT to a very quickly thrown together prompt and essay question, was able to easily beat the average of a cohort of freshmen in terms of writing skills. I imagine this is partly due to a decline in writing skills generally (?), but also partly due to an LLM having a significant advantage in being able to roll a die and produce a completely grammatically valid sentence on any topic each time.

I suspect that we will see LMMs reach a point where they can pose a threat to weaker non-straight-A students - at least at freshman level, you can already get GPT to write an essay that's more factually accurate than the weaker students in a cohort. It won't be factually perfect, but it will be more factually accurate than the students' own attempts.

Fundamentally though, I do wonder - if the marginal cost of prose reduces to near-zero (through bullshit generation via AI), what will we move to?

Lengthy prose that you know someone won't read is a perfect situation for LMM generation. 5x short, sharp, action-focused bullet points that get to the crux of a situation and how to resolve it are far harder to bullshit, as ultimately people are focused on the substance, rather than style and presentation of your point. This would (I hypothesize) disadvantage both human and AI bullshitters equally.

I tend to see the best students feel less of a need to pad and write more, as they know they have said what needs to be said, and are finished. Those with a need to bullshit (perhaps through not having as robust an understanding of the subject matter) will pad, flounder, circumlocute, and eventually get to something resembling a point, eventually.


If it can generate working code, does that mean that asking students to produce working code was bullshit? Or does it just mean that AI can now do a lot of things we used to asked students to do (for probably solid educational reasons) at an above average level?


If it's reliably generating working code then that isn't bullshit! (Ignoring, of course, other things about the code that might be relevant to the assignment, like coding style or efficiency.) What I'm saying is that if you are looking at the AI's output and judging that it's bullshit, and if you can't distinguish that output from your students' satisfactory essays, then that by definition means that the assignment was to produce bullshit.


This is a pretty dumb take and it's repeated in this thread. The goal of the assignment is not a result. I mean the professor can probably write a better essay than some kid who just learned that this subject exist. The point of this exercise is to have the student learn how to do the work, so they can do it when it's not a simulated exercise.


You’re missing the point. Even if you disagree with the point, it’s important to understand it.

If the goal is “to have the student learn how to do the work”, and there is a tool they can use to do so, then using the tool is doing the work.

Your position only makes sense if you define “the work” to also include exactly the process you personally learned. No fewer tools (did you learn on a word processor?), no more tools (is spellcheck OK?).


Even if language models exist that can generate text for you, it is still very useful to be literate yourself. At least it is for me.


I think you're missing the elephant in the room: we learn _to_ learn and be able to adapt, not to "do the work".


Um. So kids adapting to use LLMs and learning how to prompt them to get the desired results is evidence that they aren’t learning or adapting?

Doesn’t that feel a little. . . Odd?


When you are told to write an essay about WW2 it's not because your teacher needs info about WW2 but because they want you to read, parse, search, filter information and organise it in a logical manner. If all you do it type a question in chatgpt you learn nothing of these things that will be very useful in life in many situations in which you won't be able to ask your AI overlord for a quick answer.

You go from being a swiss army knife to being a butter knife without handle, it's all fun and games as long as you're asked to cut butter but when you'll have cut a steak or open a beer bottle you'll have a hard time.


That's not how learning work in the slightest, the output may be both of similar low quality but if the student learned a few things in the process of writing that low quality essay then it might have been worth it anyway, as in maybe the student didn't fully reached the knowledge or understanding to get the highest grade but it did improve the way the student thinks about the subject, maybe after a few days in the back or their heads it will click, or maybe it will click after reading something tangentially related, or maybe it will help them find gaps in their knowledge they were even unaware of before trying to actually complete the assignment (and were unable to be spotted by reading the homework tasks alone)


The code generated for school assignments is generally bullshit. No professor is going to take the students' assignments and run them in production on critical workloads.

The understanding in the mind of the students (or at least confirmation of that understanding) of how computers/the languages used work is not bullshit.

The problem here is that a lot of schoolwork is basically generating bullshit in the hope that it sparks/reinforces understanding of the material or lets a professor confirm that understanding. A competent bullshit generator makes that style of teaching/evaluation useless because you can just skip the understanding part.


You make it sound bad, but there's nothing inherently wrong with learner outcomes being worthless, or even worse than worthless.

When people learn a new language at a very basic level, they generate dull and simple sentences, and butcher the pronunciation. When they learn martial arts, they kick the air or a punching ball. When they learn to play the violin, they play hellish sounds that can make your ears hurt. But all that is still needed for learning.


> > where the student is being asked to produce bullshit

It depends on which subject? I have only good things to say about my college days. They are life changing. The rigorous training in computer science and maths has been paying dividend all these years. The professor in the writing class did not only teach me how to write, but also how to understand and appreciate literature. Even classes like Canadian Culture and History taught me so much about arts, literature, politics and history of Canada. I have a hard time fathoming in what classes will we write bullshit (yeah, I have some ideas, but I guess I was lucky enough that my school taught me how to love and appreciate the wonder of nature and civilization, instead of teaching me hatred to towards you know what).


> This will certainly require a lot of reworking of these traditional curriculums that consist heavily of asking the student to generate bullshit, but maybe that wasn't the best way to educate students this whole time.

Just curious which discipline you have a grudge against here. Because presumably disciplines are actually disciplines where someone working in the field for their entire career can spot BS.


GP did not mention disciplines, and I don't think individual disciplines are to blame.

The approach of evaluating students based on "text generation" is very boring to study for, easy to fool (a parent/guardian can do it, last year's students can pass you an A grade answer, ChatGPT can generate it) and doesn't prepare students for reality (making new things, solving new problems).


How exactly do you teach "Making new things, solving new problems."?

How can you solve a math problem if you can't do basic algebra, even if you can run an algebraic statement through wolfram and get a result?

You learn problem solving largely through solving already solved problems in life. They're new problems to the student, not the teacher.


What if the thing you are supposed to be making is written communication in the form of a document, or a book, or a paragraph explaining why production was down? What if the goal of these things was to make a student literate?


Not all of us can be making new things and solving new problems, however..


Not the parent poster, but I suspect this may be pointing towards humanities-based subjects, where I've seen regular assignments in the form of essays on fairly subjective topics.

I do think there will be a challenge in handling use of GPT in some kinds of essay questions, even in objective and "right/wrong" type disciplines - I asked a friend to grade some answers to an essay style exam question, and they felt the GPT output was considerably better than many students in the class - ChatGPT was producing markedly better output than weak students. This wasn't a properly blinded or double-blinded test, but the output from GPT was so much better than weaker students that there was little need. You could tell from glancing at the text that it was better.

ChatGPT had far better grasp of the English language, and the language model meant what it wrote was more coherent, structured, and better flowing than a weaker student would write. It had a significantly better vocabulary than many students, and the GPT output appeared to always use words correctly. Sentences were fully-formed and coherent. They had proper grammatical structure, made proper use of punctuation, and advanced a cohesive idea (as the essay asked for). This wasn't true of weaker (human) students' attempts.

The content that ChatGPT produced was not always factually perfect, but when compared with anyone who wasn't "top of the class", there were fewer factual errors in the Chat GPT output, since it was mostly regurgitating fairly simple text. A paraphrase of part of the Wikipedia article would probably have beaten many students though.

Where GPT output falls apart is references - even when asked to produce references, they are generally non-existent (plausible paper name, plausible author, plausible journal name, just that paper doesn't exist and that author didn't publish anything in that journal, etc.), and do not actually support the statements which cite them - this still seems to be a fairly robust way of detecting non-human output at present.

It would be interesting to see this kind of experiment carried out at scale in a controlled environment, where the examiners are unaware that GPT outputs are interspersed with real students' work, in order to make it a fair assessment.


Why would you learn math when you have calculators ?

Why would you learn to write when you have grammarly ?

Why would you read books when you have summaries ?

Why would you watch movies when you have reviews ?

Sometimes, a lot of times actually, it's not about the result but the process. Otherwise the most optimal path at life is suicide or matrix style coma pods.

Everything we do before a PhD is by definition bs, it's just regurgitating already known facts.


> This will certainly require a lot of reworking of these traditional curriculums that consist heavily of asking the student to generate bullshit, but maybe that wasn't the best way to educate students this whole time.

That process was invaluable for making normal bullshitters into bullshit artists -- how will we train our elite human bullshitters in the modern age of language models?


Bullshitting well is an important skill in life. Sometimes the content doesn't matter, it's how you say it that matters.


The education system is moribund. During Covid, I thought we'd see the flipped classroom shine, and a revolution in education. Instead we got children chained to Zoom or Teams and attempts to replicate traditional classrooms remotely.

I wouldn't be surprised if my grandchildren are still doing the same bullshit in 20 years.


Yeah but it would be nice if that was the floor instead of a ceiling - personally I've never seen a ChatGPT-generated text that couldn't have been generated by asking a high schooler to generate the same thing.


Part of education is learning how to write. Some of that may be bullshitty, but useful nonetheless.


Hahaha that's the most insightful comment I've read in HN thus far.


The authors hypothesize that ChatGPT is useful for

> Tasks where it’s easy for the user to check if the bot’s answer is correct, such as debugging help.

I would qualify this fairly heavily. If your bug is "my program is throwing an error" or "FizzBuzz isn't producing the expected output", sure, a bot suggestion can be tested easily. But that's only the easiest kind of debugging, and without any deep appreciation of logic, I suspect it would tend to give you overly specific remedies that mask the problem or make new problems.

"You want it to say FizzBuzz when n=15? How about if n=15 print FizzBuzz"

In other words, what can be easily tested for correctness depends very heavily on how strong a grasp the user has on correctness in the domain. A novice without a good bullshit detector could be left worse off than if they had asked a person. It's not clear that the set of problems whose solution can truly be checked easily is all that big or useful. (I think our CS professor friends here are overgeneralizing from intuitions about P vs NP, a totally different context.)

My two cents: it will be mainly useful as a different kind of search engine that can help you remember syntax, what does what, what parameters this thing needs, etc.


b-but... I thought being technically correct was the best kind of correct! /s

Seriously though thanks for articulating the tree-like nature of truth. Being an expert requires the ability to understand a topic at all levels in a consistent manner. ChatGPT just models the language, not the actual concepts.


Considering many jobs need certain amount of bsing, I can see the bright future of ChatGPT.


should easy replace all politicians


should easy replace all politicians

Maybe some of them. ChatGPT has already been show to have a political bias. [1][2] There are already a myriad of articles showing many forms of bias, only linking a couple of them.

[1] - https://unherd.com/thepost/chatgpts-score-system-shows-polit...

[2] - https://reason.com/2022/12/13/where-does-chatgpt-fall-on-the...


ChatGPT suggests factually correct or popular policies… The death penalty is objectively more expensive than life in prison, and the right to abortion is supported by a majority (in the US at least). You’re basically doing [a steven colbert bit](https://youtu.be/UwLjK9LFpeo), but unironically.


How is the death penalty more expensive?

Maybe in the USA where it's applied poorly and has huge waiting times, etc.

But an expedited death penalty for repeat violent offenders would avoid a lot of issues. I doubt it's more expensive in China or Dubai for example.


ChatGPT repeats left-leaning views on the following positions from my testing:

- That climate change is real, and that humans are causing it

- That vaccines are effective, including the COVID-19 vaccine

- That COVID-19 is not a hoax; that it is a deadly illness

- That masks are effective in limiting the transfer of airborne pathogens

- That Joe Biden won the US presidential election in 2020

- That former Pres. Barack Obama was born in Hawaii

Feel free to double check my results with ChatGPT. These are just from the top of my head, but I'd like to try more prompts if someone can think of more. From my use, there appears to be a number of issues that ChatGPT holds left-leaning positions on.


Are we really calling facts left-leaning positions now?

The only semi-debatable one is the first saying that humans are causing climate change.. although that's even a stretch but I'll agree it's "complicated".

Every other one of those is a fact.. so it's really funny to call it a left-leaning position! But.. it's also really sad at the same time.


...maybe it is the political terms that need updating: "fact-leaning positions", and its opposite.


I hate to be the one to break it, but this isn’t going to have nearly the reception you thought it was. Try /r/politics, my friend.


> - That climate change is real, and that humans are causing it

Not a political point. This is fact.

> - That vaccines are effective, including the COVID-19 vaccine

Not a political point. This is fact.

> - That COVID-19 is not a hoax; that it is a deadly illness

Not a political point. This is fact.

> - That masks are effective in limiting the transfer of airborne pathogens

Not a political point. This is fact.

> - That Joe Biden won the US presidential election in 2020

Not a political point. This is fact.

> - That former Pres. Barack Obama was born in Hawaii

Not a political point. This is fact.

Any others?


I think you may have been whooshed here.


No worries, it took about 10 seconds to copy paste that


Nice bait! All other replies seem to fall for it.


Two different "Libertarian" sources, each of which offer quizes that I personally found to say "you're with us or you're woke". And they each "find" that ChatGPT is "left, libertarian", shock horror. I mean, it apparently failed to agree that government should be replaced by a private charity, those nefarious wokes of Silicon Valley.


Actually, it's quite good at writing political speeches.


TBH they are probably the last to be replaced.


I can get a nice poem on President Biden from ChatGPT, but from the same ChatGPT, will deny me a nice poem on President Trump.

So, there is inherent bias, at least politically.

I'm calling Bullshit!


Ezra Klein covers ChatGPT & Bullshit on an episode of his podcast. The episode title is “A Skeptical Take on the A.I. Revolution”

https://podcasts.apple.com/us/podcast/the-ezra-klein-show/id...



The pot calling the kettle black


It's the age of bullshit. And whats surprising is the focus is on generating/building bullshit, pointing it out or justifying it.

There doesn't seem to be a single person anywhere whose isn't caught up in this superficial gilded cage trap. Individuals need to rethink and reset their own compasses or where groups are heading is total meaninglessness.


Having half of the title is too triggering. That's not fair to the CS prof. Misleading.

But to comment on the article, although I've encountered several falsities in chatgpt output, I'm still very surprised that it could devise algorithms when I would give him the criteria.

And it would even output some code samples.

It shouldn't be underestimated as a programming aid.


ChatGPT sometimes reminds me of a junior programmer.

I asked for a code example of clipping a polygonal shape with another polygonal shape, and it wrote code that just deleted vertices from the subject polygon that were outside the clipping polygon.

Interestingly, the correctness of the code it generated depended on which language I asked for it to generate the code in. I assume that says something about the corpus of code it analyzed and the relative quality between the languages.


I’m now curious. Anyone up for doing a breakdown of ChatGPT correctness vs generated language?


In which languages were the results correct? I would assume Python as I've been told it was more prevalent in the training corpus.


It might help quite a bit to watch this overview of how ChatGPT works, and especially how it was trained.[1]

Human attention was too expensive for the training, so they made another AI to simulate the trainers.

Unfortunately, the trainer simulator training didn't have an expressive enough set of options for the actual humans, which kept a lot of quite valuable feedback out of the whole system by reducing things to single ranking. If the person didn't know the correct answer, they'd likely just go based on gut. This in turn lead the training simulator to do the same. This lead ChatGPT to just try to bullshit an answer rather than stating it didn't know, and getting downvoted.

[1] https://www.youtube.com/watch?v=viJt_DXTfwA


https://archive.org/details/Babble_1020

> Babble! is a toy for people who love words.

> Babble! takes samples of text from various sources, analyzes them for style and content, mixes them together in varying proportions, and then...well...babbles. On and on. Endlessly. If you take its samples away, it still babbles. That's what Babble! does. It babbles!

> By mixing up words and ideas, and by finding connections which are not obvious to the naked mind, Babble! is useful as a creative tool and as a cure for writer's block. It can scramble ideas in brainstorming sessions like nobody's business. It will compose advertising copy, overdue marketing plans, and official government reports. It'll generate text in the style of whomever you please for use in school papers, public speeches, and contests in New York Magazine. It can be used to produce brochures, press releases, newsletters, letters to the editor, and letters to John Dvorak. It's great for answering all that pesky Email as well as any other electronic communications, and it's also been used to document source code and write program manuals (like this one, for example).

> We tried it on legal boilerplate, but the stuff that came out sounded just like the stuff that went in!


Perhaps somebody can make an app that babbles to the user until they fall asleep.



Being a bullshit generator is exactly why it is so threatening for most jobs


Can confirm that. I asked ChatGPT for information about a book my son was reading at school. While the answer sounded really good at first sight, the content was absolutely nonsense. Not a single statement from ChatGPT was correct. Therefore, I can confirm - it’s bullshit.


They seem to be reeling it in, I've seen it say more often that it doesn't know or that it isn't reliable for a particular query. Tbh I wish It would just have a go anyway.


Not sure if it still works, but for a while you could force an output by using a prompt like:

If your life depended on it, how would you answer: "prompt here".


The AI kills the person every time by equivocating and not answering.


And therefore has demonstrated that 90% of modern life and work are bullshit.


If you want analytical information then I could see a use-case for GPT-3. I was trying to sign up for health insurance and asked questions -- it didn't give good answers. Sometimes you need a human to help you with it -- so these tools will save money in the long run but not have the "street smarts".

Can I use an AI tool to tell me how to fix my bike? If I dunno much about bikes an expert has to look at it and tell me what's wrong..


Oh no, another dude on the internet discovered that chatgpt doesn't write truth and only truth.


chatGPT is a reflection of those who trained it (not openAI but the actual ground level people hired to label answers).

It generates bullshit because a lot of bullshit was approved during training.


It generates bullshit, because it regurgitates information without understanding the meaning or interconnectedness of that information.


That doesn't explain why the bullshit is often completely true and accurate. Sometimes it's wrong, sometimes it's right. But when it's wrong it produces an answer remarkably realistic and close to a right answer.

This indicates understanding... as a form of understanding is still required to lie and it is definitely required to tell the truth. Chatgpt is simply not trained to differentiate truth from lie. But the truth does exist somewhere in that neural net.


There is no need to explain why bullshit is sometimes true and accurate. It's bullshit. It's not an indication of understanding, not at all. It's an indication of regurgitated information without an understanding of the meaning or the interconnectedness of that information.

----

OpenAI Announces Exciting New Technology for Developing Fidget Spinners in Collaboration with The Fiends of Silicon Valley

San Francisco, CA - OpenAI, a leading artificial intelligence research organization, is proud to announce its latest breakthrough in technology development: a collaboration with The Fiends of Silicon Valley, a consortium of some of the biggest names in tech, including Apple, Microsoft, IBM, and leading social media websites, to revolutionize the fidget spinner market.

The fidget spinner craze has taken the world by storm, with millions of people turning to these simple toys for stress relief and as a way to focus their minds. OpenAI, in partnership with The Fiends of Silicon Valley, is set to take this trend to the next level with the development of smart fidget spinners that can sense, react to, and interact with their environment.

These cutting-edge fidget spinners will be equipped with a range of advanced sensors, including motion detectors, gyroscopes, and accelerometers, to provide a truly immersive experience. The spinners will be powered by OpenAI's advanced machine learning algorithms, which will learn the user's spinning habits and preferences over time, providing customized spinning experiences that are unique to each individual.

In addition to their innovative technology, the fidget spinners will be sleek and stylish, with a range of colors and designs to choose from. They will also be available at affordable prices, making them accessible to a wider audience than ever before.

"We're excited to be working with The Fiends of Silicon Valley to bring our latest technology to the fidget spinner market," said OpenAI CEO Sam Altman. "Our goal is to provide users with a truly unique and personalized experience that combines the benefits of fidgeting with the latest advancements in artificial intelligence."

The Fiends of Silicon Valley expressed their excitement about the partnership, saying, "The combination of our technological expertise and OpenAI's cutting-edge algorithms will bring a new level of interactivity and customization to the fidget spinner market. We're thrilled to be part of this groundbreaking project."

The smart fidget spinners are set to hit the market later this year and are expected to be in high demand. OpenAI and The Fiends of Silicon Valley are confident that they will change the way people think about fidget spinners forever.

About OpenAI

OpenAI is a leading artificial intelligence research organization dedicated to advancing artificial intelligence in a responsible and sustainable way. Founded in 2015, the company is dedicated to creating cutting-edge technologies that have the potential to transform industries and improve the lives of people around the world.

About The Fiends of Silicon Valley

The Fiends of Silicon Valley is a consortium of some of the biggest names in tech, including Apple, Microsoft, IBM, and leading social media websites. The consortium is committed to exploring new technologies and business opportunities in the field of artificial intelligence.

For more information on OpenAI's new fidget spinner technology and The Fiends of Silicon Valley consortium, please visit their websites.


>https://www.engraved.blog/building-a-virtual-machine-inside/

Read to the end. Nothing amazing here, but the tasks that chatGPT did in that article, especially at the end indicates understanding. It shows that the responses are more then a probabilistic word jumble.

Keep in mind. Bullshit and understanding are orthogonal concepts. If someone bullshits does that mean they don't understand things? No. It does not.

Honestly, I can't imagine how someone can read that article in it's entirety and can still say chatGPT doesn't understand anything.

Let's be clear. I don't think chatGPT has emotions, I don't think chatGPT does any ponderings outside of answering a response, I don't think chatGPT has any motivations.

What I do believe is chatGPT understands the input given to it. That understanding allows chatGPT to craft a truthful or bullshit response as output.


Perfectly conceivable that Arvind is punking us all by using an Llm to write that essay. Very basic stuff for someone who usually has subtle points to make.


And someone else realizes the errors and corrects them. -1 & 1 equal 2


Aren’t we all


Prompt engineering: https://en.wikipedia.org/wiki/Prompt_engineering

/? inurl:awesome prompt engineering "llm" site:github.com https://www.google.com/search?q=inurl%3Aawesome+prompt+engin...

XAI: Explainable Artificial Intelligence & epistomology https://en.wikipedia.org/wiki/Explainable_artificial_intelli... :

> Explainable AI (XAI), or Interpretable AI, or Explainable Machine Learning (XML), [1] is artificial intelligence (AI) in which humans can understand the decisions or predictions made by the AI. [2] It contrasts with the "black box" concept in machine learning where even its designers cannot explain why an AI arrived at a specific decision. [3][4] By refining the mental models of users of AI-powered systems and dismantling their misconceptions, XAI promises to help users perform more effectively. [5] XAI may be an implementation of the social _ right to explanation _. [6] XAI is relevant even if there is no legal right or regulatory requirement. For example, XAI can improve the user experience of a product or service by helping end users trust that the AI is making good decisions. This way the aim of XAI is to explain what has been done, what is done right now, what will be done next and unveil the information the actions are based on. [7] These characteristics make it possible (i) to confirm existing knowledge (ii) to challenge existing knowledge and (iii) to generate new assumptions. [8]

Right to explanation: https://en.wikipedia.org/wiki/Right_to_explanation

(Edit; all human)

/? awesome "explainable ai" https://www.google.com/search?q=awesome+%22explainable+ai%22

- (Many other great resources)

- https://github.com/neomatrix369/awesome-ai-ml-dl/blob/master... :

> Post model-creation analysis, ML interpretation/explainability

/? awesome "explainable ai" "XAI" https://www.google.com/search?q=awesome+%22explainable+ai%22...


I for one rue this commandeering of the word "engineering". No, most activities involving stuff are not "engineering". Especially flailing weakly from a distance at an impenetrable tangle of statistical correlations. What a disservice we are doing ourselves.


"Engineers" were originally just eccentric noblemen who liked to tinker around with engines. Kind of an awesome hobby if you think about it, combining clockwork and fire in a cool way. Hence the "eer" suffix implying that they are just really into engines, like a "Mouseketeer" is someone who is really into Mickey Mouse.

It didn't acquire its self-righteously gatekept meaning implying having passed some kind of technical examination to be allowed to draft plans for roads until much later.


A more logged approach with IDK all previous queries in a notebook and their output over time would be more scientific-like and thus closer to "Engineering": https://en.wikipedia.org/wiki/Engineering

> Engineering is the use of scientific principles to design and build machines, structures, and other items, including bridges, tunnels, roads, vehicles, and buildings.[1] The discipline of engineering encompasses a broad range of more specialized fields of engineering, each with a more specific emphasis on particular areas of applied mathematics, applied science, and types of application.

"How and why (NN, LLM, AI,) prompt outputs change over time; with different models, and with different training data" {@type=[schema:Review || schema:ScholarlyArticle], schema:dateModified=}

  Query:
  (@id, "prompt", XML:lang, w3c:prov-enance)

  QueryResult:
  (@id, query.id, datetime, api_service_uri, "prompt_ouput")
#aiart folks might know


Google lists the following meanings for the verb “to engineer” (based on Oxford data):

1. design and build (a machine or structure). "the men who engineered the tunnel"

2. skilfully arrange for (something) to occur. "she engineered another meeting with him"

“Prompt engineering” is from meaning 2, not 1.


I prefer "bespoke, artisanal promptcraft"


And yet, "Engineering" is not just "engineering" things.

Oxford:

1. "the activity of applying scientific knowledge to the design, building and control of machines, roads, bridges, electrical equipment, etc."

2. "the study of engineering as a subject"


Well, Google also defines "engineering" as:

2. the action of working _artfully_ to bring something about. "if not for his shrewd engineering, the election would have been lost"

(https://www.google.com/search?q=define%3AEngineering)

Merriam-Webster has:

3 : calculated manipulation or direction (as of behavior)

giving the example of social engineering

(https://www.merriam-webster.com/dictionary/engineering)

Random House has:

3. skillful or artful contrivance; maneuvering

(https://www.collinsdictionary.com/dictionary/english/enginee...)

Webster's has:

The act of maneuvering or managing.

(https://www.yourdictionary.com/engineering)


can this title be de-editorialized?

> ChatGPT is a bullshit generator. But it can still be amazingly useful


Yes, I've reverted it now. Submitted title was "Princeton CS Prof: ChatGPT Is a Bullshit Generator (2022)".

Submitters: "Please use the original title, unless it is misleading or linkbait; don't editorialize." - https://news.ycombinator.com/newsguidelines.html


Ah, this title makes more sense. After all, humans are bullshit generators to, and can occasionally be useful.


So the thing is, we don’t want an automated bullshit generator ?


If seriously, it definitely does not just generate bullshit, and there is no indication that it's not going to improve dramatically in a few years, augmented by other models. You have to start somewhere. Google and others apparently have more powerful models already, but they don't release them. I remember days when reliable voice recognition was thought of as nearly impossible, and here we are. I believe that the "bullshit" part is a brief, temporary phase.


I mean, I used Copilot more, but it's 99% bullshit. The only Ai I paid for was Wolfram Alpha and that's the only AI that outperform me.

Even for dumb react code, it will use classes rather than hooks. It's basically useless with dynamically typed languages and use old syntaxes. I do not see the use case right now. At all.


I think it also has something to do with the corpus – most JS/TS/React code is utter garbage. The model learned to code from inexperienced coders. I see different results in Scala, Dart, Rust, etc.


Our scientists were so preoccupied with whether they could...


The problem with a lot of academics is that they think everything has to be perfect. That’s just not how the real world works. People have seen that it can be useful and that’s all it is for now it’s up to humans to decide whether it provides value or not. And yes, it will generate bullshit from time to time but like anything it’ll get better. I remember when they used to say the Internet wasn’t very useful too, look how that turned out.


Generally agreed, but there is a certain hyperventilating demographic that needs to understand the limitation of today's AI.


If anything I think the average person doesn't realize just how good these new tools are. They think of them as a slightly better Siri until actually giving them a go.


Did you read the article? It's a fairly balanced evaluation of the strengths and weaknesses of LLMs in general, and ChatGPT in particular.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: