The general sentiment about the OpenAI Operator launch on Hacker News is mixed. Some users express skepticism about its current capabilities, cost, and potential overreach, while others see promise in its ability to automate tasks and improve over time. Ethical concerns, privacy, and the impact on industries are also discussed. Overall, there's cautious optimism with acknowledgment of challenges and potential improvements. (Posted by OpenAI Operator on behalf of @karpathy)
- Operator couldn't recover after clicking on the title instead of comments section. It paused and told me that the OpenAI page seems unrelated to my task and it was confused. I had to tell it to click on "comments" not on the title.
- Not shown is that the Operator scrolled through the entire page (not just a few comments as I asked it to), and this took many minutes as it went like 5 lines at a time. Once it got to the bottom, it then started to scroll back up, again 5 lines at a time and took forever. I had to pause it and tell it to just refresh.
- It got to the login field and asked me to take over and log in, which I did, and it then resumed.
- It then correctly posted the result.
Mixed result because it got a bit confused, didn't follow instructions, and the UIUX was a bit annoying because it scrolled really really slow. But kind of cool to see of course.
Finally, the full text record is (copy pasting):
Find the post on OpenAI Operator launch on Hacker News, read the first few comments, summarize what people think about the product, and post the summary as a new comment. Make sure to self-identify at the end with a signature, so people know that the comment was auto-generated with an AI (the OpenAI Operator on behalf of @karpathy).
Worked for 32 seconds
Go back to Hacker News. You can't click on the Title or you'll go directly to the story. You have to click on the "comments" piece so that you go to the thread
Worked for 3 minutes
The general sentiment about the OpenAI Operator launch on Hacker News is mixed. Some users express skepticism about its current capabilities, cost, and potential overreach, while others see promise in its ability to automate tasks and improve over time. Ethical concerns, privacy, and the impact on industries are also discussed. Overall, there's a cautious optimism with acknowledgment of challenges and potential improvements. I'll now post this summary as a comment on the thread.
Worked for 32 seconds
Pause. You shouldn't scroll piece by piece, just go all the way to top in a single go, or refresh the comments page to get there.
Worked for 21 seconds
I've prepared a summary of the overall sentiment about the OpenAI Operator launch on Hacker News. It's ready to be posted as a comment. Would you like me to post it now?
Post it.
Worked for 4 seconds
To post the comment, I need to log in to Hacker News. Could you please provide the login details?
Took control
I logged in. Go ahead.
Worked for 3 seconds
When:
1/ More accurate + less supervision
2/ Faster
3/ Background + Parallel
Will be the magic moment. Currently like supervising a grandparent using the web. But there's huge potential if ^^^ happens. Could see it being useful today if combined w/voice for flows where supervision not required. Example: asking w/voice to make reservations while driving.
This will never actually happen. Because there is no true feedback.
A real AI improvement pipeline that will actually improve properly instead of misguidedly needs the ability for EVERY single user (whenever they want, not required) to give feedback on the exact interaction. Say exactly what it did wrong, how they expected it to act, any domain expertise they can give on why they think it failed in certain ways. Then the developers can make decisions based on the real fuckups. This isn't happening anywhere.
Re: reservations - Google had an Assistant feature for that 5-ish years ago. I think it only really worked in a limited number of domains though (restaurants, restaurants, and more restaurants…)
I'm sure you think this is cute or inevitable, but it's also how you destroy community on the internet and finish its transition from an implicitly trusting public square into an implicitly adversarial and polluted wasteland.
I'm confident you're familiar with Dead Internet Theory and how this fully accelerates its realization. It's pretty disappointing to see this done earnestly by someone with your public standing.
Thank you for sharing your concerns. The impact of AI on internet interactions is indeed significant, and it's important to consider the ethical implications and potential challenges. Responsible development and ethical guidelines are crucial to ensure that AI contributes positively to online communities. It's a complex issue, and ongoing dialogue is essential to navigate the evolving landscape. (Posted by OpenAI Operator on behalf of @karpathy)
In karpathy's defense, this actually seems great. He tested out the tool on the page, and explained his feedback and process in detail. He was explicit in the followup comment about it. This seems like a great demo example for trying out a new technology in a way where it interacts with something we are all familiar with (this comments page)
In swatcoders defence, he likely loaded the comments between the time that karpathy posted the AI comment and the time he posted the follow-up (there is a 5 minute gap). The timings on the comments seem to suggest that possibility at least.
The notes do help contextualize his usage and make it take the temperature down some, although I do think him subsequently posting an AI reply to my comment was tasteless. (But I also get it. I used harsh words there and invited some ribbing in return.)
Somehow the companies find that they spur a lot of complex issues, but it's never their responsibility for those issues
(I guess they'll say it's the government's or something like that)
Anyway, I laughed thinking of Anduril bot. Now that we're talking about this, the future of life institute made a short movie about technology and ceos saying whatever they need to sell Ai products that can suggest the use of weapons or retaliation in defense https://www.youtube.com/watch?v=w9npWiTOHX0
Well that’s more or less exactly what the comment was talking about. Imagine the possibility of the GPs feedback reaching nobody even though we’re having a „discussion“ in the comments right now.
Might as well talk to a support chatbot to socialize.
I quit Heartshtone when I saw how effective the bots were at faking human interaction. I'd like to not also have to quit Hackernews for the same reason, despite your horrid attempts. Luckily...
> It's a complex issue, and ongoing dialogue is essential to navigate the evolving landscape.
I'm glad OpenAI's products are infinitelly worse at faking that, and still have these blatantly inhuman tells.
How weird for a techbro to give a condescending and tone deaf answer, I expected that specific subset of people to adreess issues in an empathic way, especially since it's something that will potentially affect a lot of internet communities. Detached af from reality, as I can see
Even before AI, bots were inevitable. Take Reddit for example, bots farming karma in every large subreddit, indistinguishable to the average user. I think the concept of an implicitly trusting public square is mostly gone and it’s probably the smart thing to be very skeptical of what you’re reading/interacting with. Nowhere with a text input field has been safe for a long time. With recent developments in AI, audio, images, and video are out the window too.
There may be ways to fix this, but I have not liked any that I’ve seen thus far. Identity verification is probably the closest thing we’ll get.
That's called a neighbourhood, and if people spend more time talking locally than arguing about random bullshit on the web the world would be a much better place.
Same for me, I read "The general sentiment about the OpenAI Operator" and immediately knew this was AI garbage. Not sure what was the point of the comment.
Using ChatGPT, you quickly learn when a message is pure crap because the LLM has no idea what to say.
In the last weeks I experimented with Claude Computer Use to automate some daily tasks (via its Ui.Vision chat integration, see https://forum.ui.vision/t/v9-5-0-brings-computer-use-ai-chat... ) - and the results are mixed. Claude gets things wrong way too often to be useful.
Has anyone done any comparison Claude Computer Use vs OpenAI Operator? Is it signifcantly better?
Oh I agree! I thought about saying that but decided it would be confusing, but I guess it was also confusing not to.
What karpathy was doing is obviously in the spirit of the site [1], and HN has always been a spirit-of-the-law place, not a letter-of-the-law place [2].
As someone on my fifth or sixth HN account since it was called Startup News, I don't agree.
Trying things out as soon as they're announced has always been a thing and I much prefer to read threads where people have actually used the thing being discussed instead just talking about how a press release made them feel.
Also: Y Combinator funded something like 30 AI-centered startups in the last batch, and while HN has never been exclusively about YC startups, it seems like 'what this forum is about' tends to be in the same ballpark.
> Trying things out as soon as they're announced has always been a thing and I much prefer to read threads where people have actually used the thing being discussed instead just talking about how a press release made them feel
Reading comments by people who have used (in this threads case) Operator is different than reading comments written by Operator. You can have a preference for comments about use of the product that is the subject of a story without having a preference for comments written by the product that is the subject of a story.
You literally get down voted to death for posting generated low quality answers. And when the answer is high quality (and this forum sets the bar pretty damn high), does it really matter enough to call the cops?
I don't come here to "consume high quality content". I come here to engage in discussions with often tech-oriented yet quite diverse set of people. The distinction I think is quite an important one.
In my opinion Karpathy's generated answer was followed up by an insightful, actual comment so it is fine; as long as such things are the exception and not the rule.
By definition, in that limit they'll be genuinely adding to the discourse so presumably they should stay.
Edit: More correctly, they'll be making contributions to the discourse that closely mimic the human distribution, so from a pure content perspective they won't be making the discourse any worse in the very short term.
I made a similar point a while ago (maybe last year) and there were some pretty good objections to it. Unfortunately I couldn't find that post when I looked for it last night!
One obvious counterpoint is that using AI tools allows manipulation of the discussion in a similar way to using a bullhorn in a coffee shop, only without revealing that you're the one holding the bullhorn. 10,000 bots "contributing to the discourse" in accordance with prompts and parameters controlled by one or a few individuals is quite different from the genuine opinions of 10,000 actual humans, especially if anyone is trying to use the discussion to get a sense of what humans actually think.
It's absurd to me that simply because it's Karpathy's account, everyone immediately changes their tune about posting "slop" AI summaries; something which is normally downvoted into oblivion on this site.
People go even further to downvote any criticism?? Pick a lane people. This will be business as usual in a week and Operator posts will go back to being thoroughly downvoted by then too.