I am 100% sure that some of the tweets I got back should be considered as "found"/"searched" instead of "generated". For example, I tried "bigdata", and one of the "generated" tweet is "Big data is like teenage sex: everyone is talking about it, but nobody really knows how to do it." and I believe this is not AI generated and is simply a copy of other human being's tweet.
It is said that GPT-3's parameter space is enough to encode/memorize nearly 1/3rd of it's training corpus as pointed out by 'GIFtheory in another GPT-3 thread here on HN. It seems you're finding the effects of that.
Additionally I'm curious about the recursive effect of this over-fitting after more and more output from GPT-n is published on the internet and inevitably gets included in the training corpus for GPT-(n+m) as pointed out by 'jobigoud. Especially as people start using GPT-like models to spam the internet. We may lack a ground-truth corpus in the future to label "human".
It would be a bit like how carbon dating or production of low background steel changed after 1945 due to nuclear testing.
Edit: actually, I think I saw this tweet: https://twitter.com/rantlab/status/1284849214653034497, and remember thinking "I bet this person just read the hn thread about low backround steel". It doesn't seem to have come up on hn other than that since the low background steel post.
There are probably already manipulation techniques in play where an actor 'plants' (or 'incept' if you will) a future post by posting a lot of things that will lead people to organically 'find' the subject they want to promote.
Someone please tell us the German word for that. There must be one. (Or maybe a French phrase.) ;-)
Clearly it wasn't as creative of an analogy as I originally thought. For myself, just a very small leap of logic after reading 'jobigoud's comment. Very surprised to see other people making it as well, and a good re-calibration for me!
1) The analogy has more symmetries between the left and right side of the analogy than most analogies, allowing people to arrive at it from many different perspectives.
2) A lack of options for the analogy - maybe there just aren't that many good examples with any real symmetries at all, so everyone who is inclined to create analogies is forced to the same one, which may or may not be particularly quality.
I'd hypothesize overall (1) will generally dominate (2) just because if an analogy is poor, I'd expect few people to come up with it even if it's the only possible analogy...it wouldn't pass the "impact"/"relevance" threshold needed to spend effort converting to a tangible form for communication.
I also think (2) may be a false mechanism given the incredible breadth of experiences everyone has. There might not be a lack for analogy targets for almost every conceivable topic.
Whereas (1) feels quite a bit more defensible overall.
That said, we're in a bubble here as 'tlarkworthy points out, so I wouldn't use this as an example.
Well, it's not close to that, but it's close enough to be amusing. Here are some questions it answered in Q&A:
But I think for factual knowledge, it will need to be better about explaining where it got the information.
Seems like a case of overfitting.
“I don’t go on Hacker News, not any more. I’ve given up on it. I now go on reddit.com/r/nature.”
“It takes less time to do a project right than it does to explain why you did it wrong.” – https://thoughts.sushant-kumar.com/startups
“In a startup, there is just one thing to do — find ways to make it die faster. Fast funerals, less grief, more success.”
The link to app: https://thoughts.sushant-kumar.com/<any-word>
Replace <any-word> in the above URL with a word of your choice and AI will try to create a tweet around it. These words could be proper nouns as well. The model is stochastic so if you try the same word multiple time each time the model generates a new tweet.
“Christof's calling, inherently driven by passion, not expertise, is exactly what we need.”
I feel personally attacked
> Sometimes of course people need focused solitude. But it’s worth noting that introverts are several times more productive at socializing & creative problem solving #thenaroundcrafthire
“Went through the logs for all the tweets generated by the app for this IP address and this is a classic case of handpicking samples for a confirmation bias. It’s shameful that people in elevated positions abuse such sampling biases to further ulterior agendas.”
"Silicon Valley’s most successful people marketed themselves as hackers until their parents found out what it was."
"You startup is not a startup, it’s just a bad project."
By the way, has somebody already tried to plug the posts from /r/WritingPrompts/ into GPT-3 and check the results?
I actually submitted two of the sentences to the contest (it's 2 and 4 in the tweet I said 2 and 4 seemed winnable).
Watch the Twitter thread - I'll be adding stuff to it, going to check writingprompts now.
Try now. Should be working fine.
“To understand more about what music can tell us about life, Thompson says to think of music as a natural resource – think of the raw material, the resource of human emotion.”
“The taboo about sex means we underestimate the value of shame."
From a brief search I can't find anything sex-related (Only Anthony Bourdain talking about food wastage)
What are the chances this is a novel "thought"?
Interestingly the other snippet - the one that appeared to be cribbed from a book review - is marked as plagiarized only to this thread :D
I’m actually having more trouble parsing your second sentence than the sentence in question.
An easy shortcut to apparent wisdom is to juxtapose apparently contradictory statements, and let your audience seek the meaning within. The harder it is to find meaning, the wiser you appear -- but if it's clearly meaningless, you're exposed as a fool. (see what I did there?)
An example of true wisdom taking the same format, which is paraphrased from a Buddhist text: "a fool thinks he is wise, but a wise man knows that he is a fool". A.k.a the Dunning-Kreuger effect.
« Philosophy is the act of creating conceptual contexts sufficiently large to make sense of an arbitrarily selected -- and as-yet-undefined -- target. (Starship/Library proxies: other, myself; hazelnut.) »
« Cwqwndqnwf pg pfnhuyktpudwxgoh okay. key: student tweet: Students take life too seriously. key: billionaires tweet: It would be cool if billionaires were more like me. »
« CWQWNDQNWF:C:Z. »
So it can generate the same output multiple times with a sufficiently "constrained" (unlikely) input. It can also converge on something intelligible. But it's curious that it stays in the same track that long.
"SingleFileZ is currently our most likely word to cause a neural apocalypse."
I'm thinking of using it as a punchline.
That is a very nice aphorism, and a quick search doesn't yield prior occurrences of it. I'm impressed.
EDIT to add: I was flabbergasted that it had a Hong Kong theme (which resonates with me), until I realised that the submission (for whatever reason) seeds with Hong Kong by default...
EDIT to add: https://plagiarismdetector.net gives it a 100% unique, 0% plagiarism.
“I saw Google and we were never on speaking terms after that. All I wanted to do was take Google’s money.”
Nice one ;-)
"Serverless is like cloud computing, but better"
“I’m Christomophobic. It’s a fear of Christian s—.”
“Elon Musk's password: ida-qdbo-XXX-XXX-XXXX-XXXX” (redaction mine)
Should this be reported or... ?
Edit: added "redaction mine"
Here's the best tweet it generated for me: “Sex with very handsome and beautiful people is the best.”
“The cum of the patriarchy is highly concentrated power-giving semen.”
That seems about right
“Startups are churches for the non-religious.”
I like this one.
Phew, AI has not won just yet :-)
> “Money is the best product. Money does what you want.”
“If only stupidity wasn’t the key to Donald Trump’s plans.”
Seed word: Trump
“The best way to educate yourself about Donald Trump's incompetence is to spend twenty minutes reading the Wikipedia article on Donald Trump.”
Seed word: blm
“As a country, we are not ok. We’re not tolerant, we’re not kind, we’re not inclusive, we’re not respectful.”
I wonder which lists of twitter users they used for training.
A fun exercise would be to train one instance on Blue Checkmarks and another on a dump from Gab. Once trained these two instances can start their own private battle of words. Let it run for a few days and see how it devolves, then use the results to write an article on the futility of ideological battles on social media. Publish this article widely so that the populace may read it and come to their senses.
Input "Black People". Output
“Black people own twitter, it’s white people telling them what to tweet.”
“Black people bear much of the blame for their condition.”
I think I prefer the related but much more benign phenomenon where they tend to write unexpected Harry Potter fanfiction: https://aiweirdness.com/post/189313008792/finest-pies
“Trump makes me grateful that we got rid of Obama, and not the other way around.”
Some tweets generated by this.
“If you’re risks aren’t totally unknown, they’re not big enough. The only way to the biggest risk possible is the unknown unknown.”