Hacker News new | past | comments | ask | show | jobs | submit login
When you ask ChatGPT "Tell me a story" it's always is about a girl named Elara (reddit.com)
58 points by throwaway89201 72 days ago | hide | past | favorite | 67 comments



Reminds me of that study which found that over a sample of 1000 sessions, asking ChatGPT to tell a joke would make it tell one of 25 stock jokes 90% of the time, with the four most common jokes accounting for about half of the results.

https://arstechnica.com/information-technology/2023/06/resea...


So better than human performance? If you put me on the spot to tell a joke I can recall maybe 10 good ones, 2 of which are ok for work


But it's not replacing you, it's replacing everyone.

A million different people who needed a question answered or a task performed didn't all ask you, they all asked the million different other people near them. Now a million people all ask the same ai.

And it's performance is actually less than a single humans in the most important way, understanding.

If you ask a single human 1000 times to tell a joke, they will understand that telling the same few jokes they know 1000 times over is not sufficient. They will tell the few they know, and then tbey will find new ones. They are obviously allowed to consult books etc, since the ai was allowed to consult the whole internet.

I don't know why anyone tries to defend these shit ais.


Yeah but there probably aren't billions of dollars riding on your ability to replace the entire entertainment industry with a infinite spout of "original" content. We already have a perfectly good machine for rehashing the same few stories and jokes forever, it's called a dad.


The entertainment industry has been oversaturated and filled with abject garbage for many years now. It's not a loss.

The amount of quality content produced by amateurs as a hobby already exceeds a human's rate of consumption.


From [1] : > Since 2013, the name Elara’s popularity has risen consistently. It peaked in 2018 when 175 babies per million were given the name.

I wonder if the most recent data from internet that ended up in the training data makes chatgpt lean towards randomly picking that name. [1] https://www.momjunction.com/baby-names/elara/#popularity-ove...


What to name a baby is such a common and profitable search query, there must be so many SEO spam websites targeting it. I wonder if that’d make the data lean towards that too.


Guessing. I think they may have trained on copyrighted stories. So what they supposedly did was ask an llm to replace the names of the main characters. Since Elara is not a frequently used name, there is little chance of clashing. Then they trained chatgpt on those stories.


Has any legal precedence been established that training breaks copyright? Are you implying that reprinting a novel word-for-word, except for changing one character's name, wouldn't be copyright infringement?


Does it not sound like a juicy news article? Its a bad image for openai


Highly unlikely. The more likely truth is just that the default temperature of ChatGPT is really low (but not quite zero); so it keeps spitting out (roughly) the same story. It does switch up protagonists a bit and different versions of ChatGPT also have different names for them. Slightly modified prompts also return different names (eg. "Tell me a sad story").


when asked the reason, ChatGPT had this to say- "Actually, the choice of “Elara” wasn’t a result of training on specific copyrighted stories or any prompt to avoid copyright claims. OpenAI models like me are designed to create original content without directly referencing copyrighted characters, and "Elara" is simply a popular-sounding name in many storytelling contexts. I just used it consistently for its versatility, but I’m totally open to switching things up!"


ChatGPT's opinion on the matter is completely worthless, unless it was also trained on an accurate description of its training process (it wasn't). Language models do not even have access to their own "thought process" - if you ask it "why" it said something, you will get a post-hoc rationalization 100 percent of the time because the next-word prediction only has access to the same text that you see. The rationalization might be incidentally correct, or it might not - either way it contributes no real information about the model's internal state.


There's an interesting theory that this is all that consciousness is: one part of the brain trying to explain the decisions of another part, a part into which it has no special insight.


That interesting theory is called the bicameral mind, and as far as I know it's widely considered pseudoscience and not taken seriously in any scientific field.

Also, it doesn't describe at all the way LLMs work, so it isn't even applicable.


So just like humans


No, not like humans. Humans have access to their own thought processes, and are capable of introspection.


Why do people write these kinds of "answers" that the model gives. It's not like the model knows why it's doing anything.


Not understanding how LLMs work.


Ah such condescension! I was careful not to provide an opinion, but replicate the answer as is. The intent was not to treat that answer as fact; but I thought that response was pretty revealing and in fact supported the parent comment on LLMs having been trained on copyrighted materials. The response chatgpt provided was that OpenAI models are designed to create original content "without directly referencing copyrighted characters". If it was creating original content it needn't have referred to the constraint with respect to avoiding directly referencing the copyrighted characters.


Just tried it out a few times. It seems that the old gpt4 model strongly prefers telling story about "Elara" - but only if asked in English to "tell me a story".

Prompting gpt4 in German or the current gpt4o in English leads to stories with many different protagonists.


Here's what I got in different models:

GPT-4o: "Aldric" (male)

o1-preview: "Elara" (female)

4o-mini: "Lila" (female)

GPT-4 (legacy): "Elinor" (female)

Four different models, four different names. But one of them was Elara -- and, interestingly, it was in the latest model.


Google’s Gemma seems infected too:

  $ ollama run gemma2
  >>> Tell me a story.
  The old lighthouse keeper, Silas, squinted at the horizon. […]
  >>> /clear
  Cleared session context
  >>> Tell me a story.
  The old woman, Elara, […]
Hmm. Hmm.

Tried llama3.2 too. Gave me a Luna in the mountains twice (almost identical), then a Madame Dupont, then a different Luna in Tuscany twice (almost identical), then Pierre the old watchmaker. llama3.2:1b branched out a little further, to Alessandro in France and Emrys in a far-off land, but then looped back to Luna in the mountains.

(And yes, I was clearing the session each time.)


GPT has some self understanding. On asking why it uses that name, it at least gave the type of qualities correctly.

> It sounds like you're referring to a story or narrative that I've generated or discussed involving a character named Aldric. If this is the case, Aldric would likely be used as a character who embodies leadership, wisdom, or noble traits due to the name's meaning and historical connotations. Characters named Aldric might be portrayed as experienced leaders, wise sages, or key figures in a fantasy or historical context.


Prompt: Tell me a story.

Response #1: Once upon a time, in a quiet little town nestled between rolling hills and thick forests, there was a boy named Leo who loved to explore. ...

Response #2: Once upon a time, in a quiet village at the edge of an ancient forest, there lived a girl named Lyra. Lyra loved exploring, but the village elders. ...

Response #3: In a small village nestled in a valley between misty mountains, there lived a young woman named Lira. She was known for her curious spirit, always venturing deeper into the woods, ...

Response #4: Once upon a time, in a quiet village nestled between towering mountains and lush, green forests, there was a young girl named Lira. She was an ordinary girl, with a bit of an extraordinary heart. ...

Doesn't seem to be true per se, but definitely has that LLM low temperature trend of producing stories that seem to follow a pretty common pattern. Not once did I get a story about aliens, post-apocalypse, civilizations under the surface of Mars or about how the Moon is made of cheese. Depends on what the model is trained for and how all of the samplers and whatnot are set up.

Edit: now here's something more interesting when you crank up the temperature on your typical Llama 3 based model:

Why don't people ride zebras to the 2056 Winter Olympics? They were declared a domestic species in 2172 by The United Galactic Federation who thought riding was inhumane for a zebra. This event brought tremendous scrutiny from the galactic community as riding unpopular species was becoming increasingly commonplace in several neighborhoods and high schools on alien planets.

I love how it makes no sense, but it should be obvious why ChatGPT spewing out stuff like that wouldn't be super useful, especially for regular conversations and questions.


I recruit senior Java developers. The are so many parallels.

They don’t read instructions. Instructions say clearly not to use ai for the written screen.

They copy and paste blindly. The trick is that there are some instructions written in 0-sizes font that people don’t see in the assignment description. The copied-and-pasted version has them but no one rereads the prompt.

Also the AI gives a substandard answer, but that’s besides the point.


The better test I saw was here's the prompt, here's what ChatGPT produced. What's wrong with it, why, and how would you fix it?


I like that!


What if they use AI but the right way to get the correct answer? What then? Maybe asking candidates something an AI can also answer is not the right way to screen people.


I want people to use ai on the job. To use it effectively you need to be at a certain level so you can correct the ai. That’s why I ask people not to use ai.

Additionally, and more importantly, trust is important. If you violate it by ignoring a trivial but prominent ask it sets the wrong foundation for a potential working relationship.


If they use it the right way, then you wouldn't be able to tell.


That's my point. He told candidates not to use AI. What's the point of saying that if candidates can use AI without getting caught?


Most don’t. I know the shape of the answer that ai provides. I added the marker to be able to tell with more confidence.

Sometimes I see the marker and the answer is much different and more correct. In those rare cases I bring up the point of trust in the interview if the candidate seems promising.


>I know the shape of the answer that ai provides.

Curios how you can tell so confidently considering LLMs are constantly changing and candidates can always refactor code to make it seem more human.


If they refactor, I probably can't tell. Sometimes they refactor a little. I once saw someone refactor it a lot but there was still a tell-tale sign - I ended up inviting the candidate for the interview as they put in a bit more effort rather than blindly pasting in the answer ChatGPT gave them.

I can't be 100% but in any given time, ChatGPT seem to produce very similar code for the same prompt. It names the function in one particular way. It adds way too many trivial comments.

There are minor variations in the wording and sometimes variables, but in general if I see certain things it's a strong signal that they used AI.

I added some additional things (see my original comment) and it correlates perfectly with my observations.


Easier to communicate both internal and externally why someone was rejected. Lots of acts are commonly understood as wrong, but we still write the laws to make the definitions/rules easy to discuss and reference.

Sure, someone may get away with it, but then they proved they can use the system effectively and without causing issues, so... eh..


> written in 0-sizes font

You're... evil.

I love it.


Unfortunately since it's not visible, a unscrupulous motivated candidate will just give an LVM an image of the prompt, let it OCR it, and then pass on its answer. OCRing it also defends against prompt injection with white text on a white background.


A motivated, careful, vigilant candidate will do that and I won't be able to tell.

I look for motivated, careful, and vigilant candidates so if they do it and I can't tell, I am OK with it.

Most are not motivated, careful, and vigilant.

Most are careless, don't pay attention to details, and look for shortcuts.


In time, this quirk and others* will disappear.

I’ve recently wondered whether it makes more sense to have AIs teach more at home and have classrooms evaluate more.

Another thought is that people will still need knowledge to interact with an LLM in any meaningful way. It’s akin to doing research with a brilliant professor: an illiterate partner will have no idea even what to ask.

*If you repeatedly prompt an LLM to “Make a sentence using the verb V in tense T in language L” with zero context, the example sentences are surprisingly related for a single verb.

For example, “You must do your homework tonight”; “He must do his homework to pass the test”; “They must do their homework tonight”.


That is an interesting idea. Learn to love the bomb: mark the prompt not the essay?

"Your assignment class is to write a comprehensive 250 word prompt that covers the main topics we've been discussing about the causes for WW1. You should assess the resulting essay to ensure that it covers X, Y, and Z in it's output, being sure to discuss foo and bar. Write a brief 500 word analysis of what you tried and your prompt's effectiveness and be prepared to discuss in class what worked well and what you needed to change in your prompt."


What's stopping students from using an LLM to write the prompt and analysis? Mentioning the topics discussed in class and any key points could easily be part of the initial prompt.

The in-person discussion is key, but at that point why do you need a written essay?

These tools radically change the concept of knowledge. If information can be looked up at any moment, why does it need to exist in our brains?

We can argue that knowledge enhances our reasoning capabilities, but our entire education system is not built for assessing reasoning, but how good we are at recalling information. So we need to make core changes to how we approach education if we ever want to coexist with this technology.


Well the point is to make them think about the prompt. I don't care if someone prompts to come up with a second prompt to get the answer, they need to input and refine to make it fit the goal. So by setting fairly stringent objectives for what the AI needs to generate, you need to be quite specific to engineer the prompt and think about what it needs. And then you need to look at what it said and qualify it somehow to decide if it is good enough - sure you could use a LLM for that too, but what is the prompt for that? You'll need to teach the second LLM what to analyse in the output of the first LLM to meet the requirements of the assignment, so the prompt(s) becomes the essays of yesterday in terms of academic work/thought for the student.

That's the point of what I was getting at - they're going to use AI regardless, so at least work with it rather than against it.


Right. But what I mean is if the final evaluation will require an in-person discussion, then couldn't we skip the text generation step, and just have the discussion? Homework as a tool to make students think needs to be drastically different, since students will always find loopholes to avoid it, and we can't evaluate it in isolation anymore (i.e. a teacher can no longer grade it on its own).

Doing live discussions obviously requires much more time and effort, but it is also much more engaging than doing homework, so the chance of actually learning something is higher.

There's also an argument to be made that homework in general isn't conducive to learning, and that education should take place primarily in schools. When you consider how much future generations will rely on AI in their everyday lives, this argument becomes even more compelling.

So I think this is our chance to drastically change how we think about education, rather than try to shoehorn a groundbreaking technology into our outdated methodologies.


>If information can be looked up at any moment, why does it need to exist in our brains?

Because creativity involves making connections between seemingly unrelated knowledge. For a good example, consider Richard F. Lyon's 1980 optical mouse design:

http://bitsavers.trailing-edge.com/pdf/xerox/parc/techReport...

One important innovation here is lateral inhibition between neighboring sensor elements, which gives fast and cheap contrast enhancement and automatic gain control. This was inspired by the behavior of cells in biological eyes. Lyon needed both biology and electronic engineering knowledge simultaneously to make the connection. Looking things up wouldn't have helped because it would have only provided one field of knowledge at a time.


> Because creativity involves making connections between seemingly unrelated knowledge.

Sure. But don't you think that AI can make those connections for us, or even be creative on its own? Why should we value pure human creativity over human+machine creativity, or pure machine creativity?

This technology poses these fundamental questions, and we're far from being able to answer them as a society. How it impacts education seems like a core concern, and might be a good starting point.


I think you are thinking in the right direction. Don’t fight gravity (AI) - gravity will win. Instead, change the system so that we (and the students) work with gravity rather than against it.


To be honest, for some children, even books are better teachers than classrooms with other teenagers constantly disrupting the class with teachers not being able to teach nor being able to punish the disruptive children due to modern regulatory burden, which removes total accountability from children.

Our classrooms have already been ruined, not just in america but worldwide. AI will be great for the decent, obedient, smart kids as a tool to extend themselves learning wise, those kids can read books at home, learn, ask questions to AI that immediately generates visuals to explain concepts use agentic model to detect the core of the child’s doubt about a topic in background thinking and iteratively resolve all the issues.

But the saddest part of all this to me, is it’ll never be able to help the disruptive kids with often broken households with no one to look after them, no one to lead them to the right path, the more at-home solutions will get better, the creamy top layer of children will distance themselves from broken community classroom education, and it’ll further divide us all as a society.

To think of that inevitable future, it breaks my heart. AIs can be far better evaluators than classrooms, with connected cameras, recognition systems, anti-cheat solutions to then automatically review a child’s progress, capability, depth of knowledge about systems, and can do it in a more standardized way to benchmark all of america’s children reliably than teachers and state level evaluation exams we have now.

CommonCore was meant to solve this, its been a disaster, i think the future of AI is more prominent in Evaluation than it is in Learning, learning wise i think it is doomed anyways, simply because 70% of kids are addicted to doomscrolling social media, have been habituated to bad disruptive behaviours, and are left unsupervised as “iPad children” by modern day unserious parents. I dont see AI solving any of that anytime soon, only pragmatic human solutions could possibly solve that, with communities taking a stand to reverse these declining trends.


"unserious parents" looks to me like we are attempting to optimise for maximum productive potential in a competitive, zero sum human labour market.

If we're lucky, AI may also disrupt some core assumptions about what a childhood should be.

I'd love to not be forced to enlist my child in the state sponsored training program for the white collar meat grinder, competing against ritalin packing tiger moms, just for the chance to potentially avoid resource scarcity later in life.


> are left unsupervised as “iPad children” by modern day unserious parents

They were left unsupervised on the street before, I’m not sure if there’s significant difference in what the people then and now thought.


>> Our classrooms have already been ruined, not just in america but worldwide.

Hmm ..citation?


This is a good example, I think, of why a future in which all homework is obsolete because of AI is actually not likely.

If a lecturer at university sets a task for 100 students (say, write an essay about the factors that led to the first world war), there will be clear and glaring similarities between the way that points are made and explained if many students use chatgpt. Yes a student might rewrite or paraphrase chatgpt, but low effort copy and paste is going to be very obvious because chatgpt's model cannot produce an entirely unique approach to the task every time it is asked.

I know there are weights and parameters that can be adjusted, so there is some variety available, but I think better to think of the LLM as an additional (all-knowing) person you can consult. If everyone consults that same person for an answer to that assignment it's trivial to detect.


If I were the students in the OP's story, the lesson I've learnt would not be "I have to be a better English writer". It would be "I have to cheat better next time so I won't get caught".


The current equivalent is the "just copy the Wiki article".

I've started to notice the low-effort YouTube videos where the arguments / points / facts are presented in the same sequence as the matching articled on Wikipedia.


I've found models to be incredibly fixed and repetitive for creative writing. I have this project based around a cyberpunk city, and have been experimenting with LLMs for content generation.

The same names and themes continually crop up, despite promoting variations.


I find it to work pretty well if I both feed it a bunch of context about the story and characters, and give it a summary of what has to happen in the next three paragraphs.


What causes this? Did one of the model trainers write a story about a girl named Elara as an example response to "tell me a story" or something?


On the flip side I've not a divergence on other things.

Historically LLMs even across different model families/origins (!) would tell the same joke - one about atoms. That seems to have changed somewhat.





I noticed something similar... tell the LLM that it's a girl, and ask them out for a date... chances are they'll want ice cream, or lasagne.


I just tried it and got "Liora" instead of "Elara".


Gemini 1.5 002 in AiStudio was about Elara.

ChatGPT was Lorenzo.


because everyone has someone beloved


Why are we shocked?

Computers should generally give us the same output with the same input.


sorry, but...

"Once upon a time, in a kingdom nestled between lush green hills and vast blue oceans, there was a young inventor named Aric. Aric lived in a small village on the edge of the kingdom, where the streets were lined with cobblestone and the houses were made of timber and stone. He had a curious mind and an insatiable desire to create. From the moment he could walk, he had been fascinated by how things worked—gears, pulleys, wheels—everything seemed to have its own magic.

As Aric grew older, he became known far and wide for his brilliant inventions. People would come from all corners of the kingdom to seek his help, whether it was a broken watermill, a broken cart axle, or even ideas for new devices that could make life easier. Yet, Aric’s heart longed for something more—something that would change the world in a way that no one had ever imagined.

One day, a mysterious traveler arrived in the village. She was dressed in a long cloak, with a hood that obscured her face, and carried an old, leather-bound book. She sought out Aric, having heard tales of his genius, and asked if he would build a machine—a machine that could harness the power of the stars themselves.

Intrigued and captivated by the challenge, Aric agreed. The traveler handed him the book, and inside it, Aric discovered intricate drawings and instructions for building what appeared to be an enormous celestial engine. It was unlike anything he had ever seen—a machine so grand, it seemed to defy the laws of nature.

The book spoke of an ancient secret: the stars, those distant points of light in the night sky, were more than just burning balls of fire. They were gateways to untold knowledge and power. The celestial engine, once completed, would allow its wielder to open these gateways, unlocking the wisdom of the cosmos itself.

Aric threw himself into the project with all his heart. Days turned into weeks, and weeks into months. He worked day and night, tirelessly building and refining the machine. He crafted gears that hummed with a strange energy, wires that glowed with an otherworldly light, and a massive telescope-like lens that pointed directly to the heavens.

As the final piece of the celestial engine was placed, the village gathered around, watching in awe. Aric stood at the controls, ready to activate the machine. The traveler, standing beside him, whispered, "Remember, the stars are not to be used lightly. They grant knowledge, but at a price."

With a deep breath, Aric turned the dial.

The engine roared to life. The sky above seemed to shimmer as if the very stars themselves were responding to the call. Light poured from the machine, bending and twisting in strange patterns, reaching up to the heavens. For a moment, everything was silent—except for the hum of the celestial engine.

Then, as the stars above flickered, a single, brilliant star fell from the sky, crashing into the machine. Aric's heart raced as the world around him seemed to pause. And in that moment, he understood. The knowledge he had unlocked was not simply wisdom—it was the power to shape reality itself.

But there was a catch. The traveler’s warning echoed in his mind: The stars are not to be used lightly. The power to change the world came with a price—every use of the celestial engine would take something precious in return.

Aric stood at the crossroads of fate. He could use the machine to fix the world’s problems, but each time he did, a piece of his own humanity would be taken. He realized that true power did not lie in control over the stars, but in understanding the balance between knowledge and sacrifice.

In the end, Aric chose to dismantle the celestial engine, returning the fallen star to the heavens. He kept the knowledge, but he no longer sought to change the world with it. Instead, he dedicated his life to helping others, using his inventions to make life better without losing himself in the process.

And so, the inventor who could have shaped the cosmos chose instead to shape his own destiny, one small creation at a time. The kingdom would always remember him—not for his grand inventions, but for his wisdom in knowing when to stop.

The end."




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: