Hacker News new | past | comments | ask | show | jobs | submit login

>"Eg given the fact "Foo is the capital city of Bar", it might be able to complete "Foo is the capital city of _" but not "The capital city of Bar is _"."

Is that really so? It would seem to be well within what I thought they were capable of based on all the other things they can do correctly.




The reversal curse itself is real. An example I can remember from the paper is "Who is Tom Cruise's mother? [Tom Cruise's mother's name]" paired with "Who is [Tom Cruise's mother's name]'s son? [incorrect answer or "Can't answer that"]". It is an interesting side effect of the way they think and a significant annoyance in getting them to work as we want them to.

That said, I think people make too much of it as an "LLMs can't reason" point, when I don't think that's accurate. What it says is that LLMs instant recall is not logically bidirectional, but this is something that humans do as well. Humans take longer to respond to (and are less accurate answering) "Who is Tom Cruise's mother?" than "Who is [her name]'s son?". At least for me, when I get questions that are the "wrong way around", I have to literally run through it logically in my head, generally along the lines of "(What does that name remind me of, is she a spy? Or her son is a spy? Is she a fictional character? Wait I think this is is a celebrity thing, which spy celebrity has her as his mum? Oh yeah, Tom Cruise.) [Out loud:] Tom Cruise."

Also, some people misunderstand the actual deficiency, and think that the LLM can't answer the question at all, rather than just zero shot. The LLM can answer the question if it has the information in context, it can reason "If A=B, then B=A" just fine. It just can't do the less popular halves of AB equivalencies zero shot.


Good point. In some ways this kind of error makes it feel like its process is very comparable to what humans do.


Yeah, I'm counting this as yet another weak but positive evidence that we've hit on something fundamental here - if not universal, then at least to the path evolution took that led to a human mind.


Yeah, I agree. There is exactly one non-LLM entity in existence that can reason this generally and this well in human languages, and that's us, human beings. We built LLMs by taking inspiration from the human brain and trying to approximate it with neural networks that were often able to achieve intelligent-ish performance on tasks in narrow domains, and eventually we stumbled on an architecture that is truly general, even if it's generally dumb. It would be an absurd coincidence to me if that architecture, LLMs, actually had nothing in common with how humans think. None of that means it is the best architecture for thinking like humans, or that we just need to scale it up to get to super-intelligence, or that it is currently as smart as a human being. But it just doesn't seem plausible that it behaves so much like a human mind if there's really nothing in common underneath.


> An example I can remember from the paper is "Who is Tom Cruise's mother? [Tom Cruise's mother's name]" paired with "Who is [Tom Cruise's mother's name]'s son? [incorrect answer or "Can't answer that"]".

The paper is, apparently, still under review.

In the mean time, may I suggest you to verify that example by yourself?


I'm seeing exactly what lucubratory described with ChatGPT. If the information is in the context window, it has no trouble working out the reverse.

    Me: Who is tom cruises mother?
    ChatGPT: Tom Cruise's mother is Mary Lee Pfeiffer.
    User: Who is Mary Lee pfeiffers son
    ChatGPT: Mary Lee Pfeiffer's son is the famous actor, Tom Cruise.
But if you ask my second question directly into a fresh session, it doesn't know the answer.

Interestingly though, you can give it additional clues and it'll get it. https://chat.openai.com/share/893c1088-6718-4113-a3f1-cf273d...


I misread lucubratory's comment: indeed I see the same as you do. I only tried asking both question in the same session. I didn't see that point in the paper when I quickly skimmed through it to find the relevant part.

I also agree with him about humans capable of the same "errors".


Her, but thank you for rereading the comment, I appreciate it.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: