Hacker News new | past | comments | ask | show | jobs | submit login
I replaced all our blog thumbnails using DALL·E 2 (deephaven.io)
677 points by dsmmcken on Aug 8, 2022 | hide | past | favorite | 322 comments



>"While the role of the artist isn’t going away soon, the role of stock image sites might disappear. "

Not, yet. While it's cheap relative to stock images, it's time consuming to generate exactly what you want. Prices for stock images will collapse for the common quick to use images but the price for the specialized high end images will hold their value or even increase in value. Those historical and such images will continue to be valuable.

It will be interesting to see if a specialized job will rise where people will get paid to generate just the right image. It might be called "A.I. image artist " This individual will generate an image with an A.I. but use graphic tools to finalize it for use.


Also there is nothing preventing Stock Image Sites themselves using Dall-E to generate additional images. Heck they can use their own existing images for training (which the other's can't due to copyright issues) to increase the portfolio, but the Stock Image Sites can access free public images.

So, counter-intuitively it may strengthen Stock Image Sites value


Considering how DALL-E is fed with so many stock images that it sometimes spontaneously generates specific stock image websites' watermarks on output,[1] this is the stupidest, blandest possible ouroboros.

[1] https://twitter.com/kevin2kelly/status/1551964984325812224


The latent-diffusion[1] I've got running at home frequently generates stock image watermarks (e.g. "The London Skyline at night in the style of Carboni"[2], images 1, 2, and 6)

[1] https://github.com/CompVis/latent-diffusion [2] https://imgur.com/a/8tOI9QU


Just because "Stock Image Sites" haven't sued Dall-E, doesn't mean it is legal.

The minute Dall-E becomes a threat, these sites will enforce Dall-E to retrain from public / creative commons images. But Stock Image sites will and be one-up on Dall-E

It's almost you don't how to play chess or business strategy.


Is this DALL-E? There’s no DALL-E signature on the image.

The same prompt with DALL-E gives me proper silhouettes with no watermark.


Probably Midjourney which doesn't add the signature and outputs 4 images likes this.


This is such a clever hack: “here, take this answer; it’s good the ‘AI’ generated it”, and if you can’t check like you did you might be able to spoof a self serving recommendation through simply because interprability of the inference is so difficult and reproducing might be costly.


Meh it’s still original artwork. It’s derivative ha.


> which other‘s can‘t due to copyright

Apparently using copyrighted training data is okay in EU: Directive (EU) 2019/790 … on copyright and related rights Article 4 [1]

[1] https://eur-lex.europa.eu/legal-content/EN/TXT/HTML/?uri=CEL...


That exception only applies when the license doesn't explicitly reserve the right. Open source licenses could be updated to prohibit use as training data, or at least to clarify the position.


> (which the other's can't due to copyright issues)

Be sure to pass this along to Microsoft's GitHub Copilot.


Nope copilot is only trained on open source codebases.


open source isn't the same as public domain. the code is still copyrighted and has a specific license that has to be adhered to in order to be allowed to use it


Nobody knows whether or not training a NN is fair use. If it were ruled to be not fair use then it would basically shut down a large chunk of ML research in the US, as it's not just Copilot that is training on copyrighted data. All the large language models require copyrighted data such as web crawls as there is just not enough public domain material around.

So even if you are right, and courts rule that training is not fair use, it seems likely that all the big tech companies would lobby Congress to restore the prior state of affairs, as not having an ML industry is somewhat of a national security issue if the rest of the world is going full steam ahead on making Skynet.


„Dad, Dad! Can you give me your gun? I want to shoot myself in my foot!“ - „Oh no, why would you ever want to do this?!“ - „the neighbours are doing it too!“ - „oh, alrighty then!“

I swear, world politics and economics isn’t much different from that, intellectually….


The problem with this analogy is that the companies are large enough to have leverage over the legal system to likely be able to avoid any consequences. Even if case law eventually rules in favor of the original copyright holders of the training data, it's the customers of the ML companies who would most likely be directly liable for infringement, not the ML companies themselves (though the customers could then try to sue them for damages).

Since there is no legal precedent and the law itself isn't clear about this use case, it's basically a huge gamble on a legal gray area at this point. For VCs the risk doesn't really matter as ML startups only need to exist long enough to provide an exit with high ROI and for enterprise companies it doesn't matter as ML products are just one of many ventures for them.

It's worth noting that unlike Germany, where book and newspaper publishers have won rather unusual copyright claims against companies like Google, in the US the big publishing industries to worry about are movies and music, and most ML projects right now seem to focus on generating images or text rather than music or video. If "AI generated music" caught on like DALL-E 2 did, I think we'd see a lot more contention over how copyright law applies to ML training data.


German quotation marks are now in the house.


We could you know… ask people to donate content to these systems. They could train on Creative Commons and ask Twitter to build image license options in to their UI, then train on all the freely licensed images.

So we could actually follow copyright laws and still have an ML industry. But I’m not sure big tech wants to ask the public for consent. They would rather have free reign to do whatever they want.

By the way I don’t like copyright law or the concept of IP. But I find it a little annoying that I’m supposed to respect IP law and ML stuff can just ignore it. Also if big tech was forced to encourage people to share stuff with an open license, this would be a huge net good for society! Instead nothing changes but big tech gets to take advantage of peoples copyrighted works and artists can’t do anything to stop it. That kinda sucks.


What can AI do but you can't? The copyright violation is in the output published work, not in the method of operation.


As a tangent... a human being who reads up on websites (that are not explicitly in the public domain) is in essence also "training" themselves on that data. It would be really short-sighted (and impossible to start with) to forbid an AI to read it if a human is allowed..


But even human-generated work has all sort of domain-specific “fair use” rules to comply with, including plagiarism (academia), open source licensing (code), attribution for generative works (CC) and so on. People who make too closely generative works are scrutinized and face social and business pressures (art, law).

Today’s ML output throws everything into a mixer and then blanket-calls it ML-generated output, because the original training content has been separated from the social and legal frameworks that govern it.


Absolutely, but it's the output of the human (or AI) that plagiarizes that might be ethically wrong or illegal when used in certain ways, not the human (or AI) reading the input documents in the first place.


> Nobody knows whether or not training a NN is fair use

A NN producing a copyrighted work without respecting the license is not fair use, we know that much.


Eventually, like in the music industry, we will have dedicated legal teams taking anyone to court who is possibly using a combination of those 5 colors out of the 50 million color combinations we "own" simply because filling your work schedule with court cases is the right thing to do.

The funniest ones will be where ML independently reproduces a picture from unrelated materials with humans making a futile effort trying to figure out how it obtained this result.


sure, I agree that the legal status is unclear. My point was mostly just to rebut the dismissal of the comparison by my parent poster.


What about my biological intelligence? Let's say I'm just training myself (learning) by browsing and reading a lot of open source projects, not even bothering to check what the licenses say. Am I possibly violating any license? And I will produce some output based on this, writing code at work (that output will not be copy-pasted, but certainly based on the "training" I received from reading through the open-source code).


Ah yes, I envision a language where every method is someones property with various subscription levels for each. Imagine how well maintained everything would be, how rich in features....


The training data for is partially made from Alamy , I think.


Agreed - having played around with DALL-E 2 a fair bit and having made a lot of usage of stock images over the years (for blog posts with specific subjects), I would say the former takes more work/time than the latter. With stock images I can just do a quick search on Shutterstock and find a lot of high quality options (usually), whereas with DALL-E 2 I need to figure out the exact prompt I want and iterate on it for a while. Stock images are not that expensive -- if you buy many it's as low as $2 per image, or on the high end (if you pay to just download a few per month) it's more like $10 per image. It does cost more, but time is money, so...


Time is only money if you someone will pay you for your time. If that’s not true, then it’s just a good excuse to spend your money when you could (in some cases should) be spending your time.


True - I guess I meant "time is money" in the broader sense really (ie I have other things I would prefer to spend my time doing, whether paid or not).


Here's a vast catalog of Dall-E images and the prompts used to generate them. https://www.krea.ai/

Oh, and if you generate an image with Dall-E and there's a face that is distorted, you can use this tool to restore the facial features. https://arc.tencent.com/en/ai-demos/faceRestoration


> Here's a vast catalog of Dall-E images and the prompts used to generate them. https://www.krea.ai/

What a weird decision to only display prompts in the page title, and nowhere else on the page


Agree, it makes discovery tedious.


> https://arc.tencent.com/en/ai-demos/faceRestoration

Very interesting and promising, thanks! But I always get a 500 error when trying to upload an image... Were you able to make it work past the demo?


-- I agree with you - however it's not that time consuming to get what you want - it's pretty easy once you get used to DALL-E - so far there isn't anything I've not been able to get on a couple tries (granted after spending ~$30 learning the prompt system)- however once you're used to it - it's fairly easy - I agree that the market for very custom work will go through the roof - but "I need a burger" or "I need an American looking hot dog" eeeek!!! =) --

https://i.imgur.com/32Cq2M3.png

https://i.imgur.com/UAlAzfl.png


> While it's cheap relative to stock images, it's time consuming to generate exactly what you want

My dall-e experience is very limited but looking for the right photo out of many is a very time consuming process, at least at designer level.


> It will be interesting to see if a specialized job will rise where people will get paid to generate just the right image.

I don't think so. People just need the result, so the AI will simply become a tool of the trade and you won't have any more AI image engineers than you have dedicated Photoshop artists right now.


I think the point of differentiation is that being adept at Photoshop is a relatively advanced and specialized skill.

Manipulating an AI prompt to get what you want is also a specialized skill, but may require an order of magnitude or two less training, or obviate the need for the job entirely.

An art director for a campaign wants a set of images created, and a set of stock photos used. They may have a junior person on their team create those images, each of which could take hours to create, or they can produce those images with an AI tool, which might take minutes. Or the director may simply use the tool themselves for a few minutes and then hand it off to someone else to clean up "in post".


Would that be similar to a "prompt engineer" role?


I don't know why but I can't help but feel this role is going to exist. "Senior Prompt Engineer", "Principal Prompt Engineer", etc. The future's wild.


As a programmer, it's my precise job description to "engineer google search terms for information I need."


Think people who think this is a job will be deluding themselves, eventually the magic of getting a good image or a bad image will just be a set of known good styles it generates at the start that you can then pick and choose from in the backend invisible to the actual thing you type. Same way how Dall-e 2 solves it's diversity issues.

The goal of OpenAI isn't to build a whole new industry of AI Artists, it's to make the AWS of creativity, which means it has to be so simple that you don't even need anyone who can write a good sentence. Just has to get good results from whatever they type.


Ah, yes, because AWS famously didn't build a whole new industry of engineers


Help me out. What's a "prompt engineer?" It seems like a very good title for the way we will be using AIs in the future. Many of these AIs will need just the right prompt (question?) to get the info we need. I like the title.


People selling their prompts on PromptBase.com might think of themselves as prompt engineers...


Make sure to check PromptOverflow.com if you get stuck.


It's basically a way in which already delusional people are further deluding themselves that writing a part of a sentence while interacting with a black box is engineering.


Engineering is full of highly specialized disciplines. I met radio antenna engineers who only ever worked on one single piece of test equipment their entire working career and who used it to watch a signal bounce while they moved around pieces of copper tape on a PCB. That's literally all they did, and they didn't know what a spectrum analyzer or any other related tool was.

You'll find these specialized positions working with AI as well soon.

We can deride it from not being "traditional" engineering, but engineering as defined as "doing the best you can with tools you have", certainly allows for a prompt engineer to be defined..


>engineering as defined as "doing the best you can with tools you have"

This is not a definition of engineering anyone sane holds. It's so broad it can be used to define a floor cleaner as an engineer (or, if you want higher tech involved, a floor cleaner with a roomba).

This is not logic, this is travesty.


$p$g


>Prices for stock horses will collapse for the common quick to use horses but the price for the specialized high end horses will hold their value or even increase in value.

This was all true when cars became a thing. What’s the market cap of horse production companies before and after?


The future of jobs is Celery Man

https://www.youtube.com/watch?v=maAFcEU6atk


You are right. Humans will move on to building more high quality images. But for regular run of the mill stock images, AI is already there. I had done a small experiment to create stock images[1] [1] https://medium.com/ozonetel-ai/generating-a-landing-page-wit...


OTOH, sometimes stock pictures made by humans aren't even that relevant if the entity using them is mostly just looking for filler: https://www.reddit.com/r/weirdwikihow/


Vocabularies are starting to appear. Something like a pattern language.


This means the result in the best-use case will be developing a suitable in-house pattern language. I'm sure Adobe's all over this since it's related to language arts and will be a more natural fit for the critical eye of a design & branding team.

And probably still, downstream designers are going to be showing how they can convert DALL·E 2 imagery to polished finals. Especially after reviewing the blog post, it's really clear that if you want things to come together well for a refined corporate environment you'll need someone doing that. "I love the whale imagery, but I don't like the DALL·E 2 look, what can we do about the whale teeth" or whatever will definitely be a thing.


Not an artist, but (confirming against DeviantArt) with DALL-E etc it appears much lower cost to take bigger -- actually artistically more impressive -- risks. And dare I say more ego-less?

One thing I have wondered: what will these generators do to the corporate art market -- art bought in bulk for a hospital or office space. Will interior design specialists pick up prompt generation as an added skill?


The #jobs decreasing is an existential threat to any sector


AI DJ?


Honestly, this is an awesome use of DALL-E and I'll probably start doing the same for my blog

It's perfect because:

- The images just need to get across a vibe, they don't need to be perfect

- It's a low-value enough use of images that you'd probably never commission a human artist to do them; instead you'd either use stock photos, or skip having images completely

- The nature of header images for a tech blog tends toward the abstract/surreal, which means it's either hard to find the right stock images, or the ones you do find will be super abstract to the point of being boring

All of these make it a great use of the technology


>The images just need to get across a vibe

I can't speak for everyone, but with my own experience of reading (at least partially) a dozen or two technical articles almost every day for many years, pointless media is a hallmark of low quality. these days, I just immediately bail with Ctrl+W as soon as I encounter a twitter-pop-culture meme/gif in the header or anywhere near the top. sure, it does mean I skip the 5 out of 100 that were worth reading, I save a lot of time by skipping the 95 out of 100 that weren't.


I almost always bail when I see animated gifs in a technical article. They are most often meant to add a touch of humour, so of questionable value to begin with, and I find them painfully distracting from the text I'm actually trying to read.

“Perfection is achieved, not when there is nothing more to add, but when there is nothing left to take away.” — Antoine de Saint-Exupéry


>They are most often meant to add a touch of humour, _so of questionable value to begin with_

This comes across as remarkably sad to me. Humor is of questionable value as a baseline?

To each their own, I suppose. I enjoy a technical article with animations that explain things, or a well-placed joke/humorous anecdote/etc. Especially in cases where the humor also supports a point or illustrates something. I tend to remember things better when they have some sort of emotional impact, and out of the emotional impacts possible (sad, angry, etc.) I definitely prefer joy, humor, etc.


You're right. I definitely worded that poorly. My distaste for animated gifs that are _meant_ to be funny, but I find sensorially painful and detracting rather than enhancing definitely leaves me in poor humour.

All in favour of animations that explain things and definitely enjoy a well-placed joke or anecdote.


I use image thumbnails for my blog posts as well but they don't even show up in the article. Instead when the article is linked to on twitter or a message the thumbnail will show up. That's to say that this is pretty awesome and very valuable for a blog even if not used for cringy pop-gif stuff in the article.


but if that thumbnail is from an AI-generated image, it has no value.

tech blogs don't need to follow the same practices as clickbait farms. the audience is very different


How would you even know it's AI generated at this point? Did you see the images generated by the author of the article? They might as well have been made by a human digital artist


they're great. but if you have no image relevant to an article to put in the header/thumbnail, what value does using a cringy stock photo or generic cutesy cartoon add?

the infotainment/clickbait/gossip websites do that because they know their target audience. you don't need to do that on your tech blog


1. It will attract people by catching their attention

2. It will not affect clickthrough rates of people like you who don't like it


You are probably right about meme/gifs (or xkcd comics) anywhere near the top indicating a low-quality post. Though they are fine if you can get a few pages in before finding memes/comics. Meme style gifs below the fold... probably not.

But we are talking here about headers with abstract "art". They are more there for style and "vibe".

Placing a Meme/gif/comic comic above the fold seems to indicate the author was unconfident about the actual content hooking the reader, and they decided to try and hook them with humour or recognisable memes instead. Which is a bad sign in itself.

All that custom abstract art in the header really tells you is the author cares about style/vibe. Which I'd argue is not a bad sign in itself; though perhaps it's a warning sign to quickly check other things, like does the style/vibe match the content you are expecting? Is this just low-effort content to attract newsletter signups?

It's also annoying when the header takes up more than half the screen. Especially more than half of a desktop screen. Phones are somewhat excusable. But I'm not sure there is a correlation between that and bad articles. Caring about style is not the same thing as being good at style.


Don't want them either, but, please blame the specs? not needfully the writers?

https://ogp.me/#metadata opengraph requires an image, for example


You can use the same image for every page on the website if you want.


I've recently started dabbling in short story narrative writing as a hobby and I found a super interesting usage of dall-e is to generate certain art works or art style to draw inspiration from.

For example I came up with [0] after writing a draft about an old warrior/mercenary in a fantasy-like setting and then put something into dall-e and built upon that just to get the right "vibe". Or if you're into more "cosmic horror" kind of stuff I generated artworks like these [1] which gave me a lot of inspiration for future short stories I'm planning to draft.

I only spent about $15 so far, and a lot of it was just experimenting with artstyle (mostly to get some interesting discord profile pictures and logos) but I feel like I learned a lot. I can't stress how ridiculously cheap it is for the amount of quality artwork you get out of it.

[0] https://twitter.com/xMorgawr/status/1555728353780310017

[1] https://twitter.com/xMorgawr/status/1556667345443049473


> The images just need to get across a vibe, they don't need to be perfect

I dislike super generic stock photo at the beginning of an article. It’s completely pointless, sometimes aesthetically unpleasant, often disconnected with the actual content, and hence a distraction.

If neither you nor the reader cares about the stock photo, why not just forsake the thumbnail or use your website’s logo?


It seems pointless but it works. Our monkey brains like aesthetics. From the article:

> Blog posts with images get 2.3x more engagement


Those metrics imply that what images are being used is completely irrelevant. Either that's true, and we could all just use the same photo everywhere (I suggest the one of St. IGNUcious, Your Mileage May Obviously Not Vary - all hail the irrefutable and infallible 2.3x metrics!) or it could still be true that a "super generic stock photo at the beginning of an article" is, indeed, pointless.


I wonder if that is because of social media. In which case add the silly image to the metadata for embedding as a tweet, but keep it out of the post.


That number refers to actual images in blogs, not specifically stock photo thumbnail.

Perhaps I’m just an outlier; but every time I see a thumbnail that has nothing to do with the title, I scroll past it. It’s a negative for me.


Yes also due to shortening attention spans


Agree, especially since it could bring some image variety into blogging. Sometimes I keep seeing the same, overused image again and again. Now you can create super cool looking, unique images and tailor them to how you'd like them to look like.


Oh jesus. 99,5 percent of sites use terrible, bland, uninteresting images. Like people have no sense of beauty at all. Now people will use these horrible abominations that DALLE shits out, well, a very nice and bright future is ahead of us :D.


I'm hoping websites eventually accept that forcing everyone to scroll past a bunch of useless barely (if at all) relevant images to reach content isn't making their sites better.


Am I one of the few people who finds these generated pictures really bad? They often have weird and unsettling details when you look closely.

I mean, it’s an incredible achievement in AI that we can generate images at this level, but I don’t want them shown to me on a daily basis while I’m reading blogs.


I sunk ~20 hours and $100 playing with DALL-E since last week and I've had a very different experience. Sure--my first dozen attempts with the engine gave bad results, but once I learned to "speak its language" it got easier to generate highly-polished images. The most realistic results come by appending things like "realistic photograph, 4k, in the style of a fashion magazine" to prompts. I suppose any style would work, as long as the body of source material in that style is (mostly) high-quality.

Here's a couple examples I produced with just a little trial and error. FWIW I have an engineering background and zero design experience.

"Frida Kahlo crossed with Julia Child, 4k realistic, expressive photo, hdr" https://labs.openai.com/s/hvFClrAMCXN6zwqJUJwsmYSB

"John Lennon crossed with Paul McCartney, 4k photograph" https://labs.openai.com/s/lb7qw07tdvRPZ9nmkrCmU0RA

Maybe they're not perfect, but I'm impressed as hell. Exploring what's possible by wording prompts differently feels very much like using a search engine for the first time. Give it a year. This technology is going places.


ERROR> https://openailabs-site.azureedge.net/public-assets/d/b3ac74...: Uncaught RangeError: Failed to construct 'ImageData': Out of memory at ImageData creation

I thought 2GB ought be enough for everybody.


Well, for 100 bucks you feel satisfied. That's what counts :DDDDDDDDDDDDDD.


I find the images to be incredible, but it's very unsettling when you focus on certain details like hands, feet, and eyes. The hands and feet that it draws are almost always mangled, and while it does a good job of drawing an individual eye, it doesn't seem to draw two eyes in a well coordinated manner, either one eye is bigger than the other, or there is something weirdly unsymmetrical about the eyes that makes the image look creepy.


I have a great example of this:

"Jesus takes a selfie" - https://imgur.com/a/togE2ko

I'm pretty sure this snap was three days after the resurrection.


Something that truly amazes me, is how well Dall-e handles lighting. The lighting and shadowing is really good, comparable to a path tracer, except it isn't really doing the super expensive light simulation that a path tracer does.


The shadows of the turkeys in OPs post is what impressed me the most so far...


Pretty much the definition of uncanny valley.


> The hands and feet that it draws are almost always mangled

I really, really REALLY don't like this fact and I won't be using or endorsing the technology until it's improved.

I also always hated the "deep dream" pictures of lovecraftian dog horrors


Just recently saw a three fingered spider-man created by DALL-E


I wonder if there’s a potential cottage industry of GANs that then fix up these details — one that knows exactly what a hand should look like and will fix up anything that looks like (or that you identify as) a hand


There are already ones to fix up faces, it'd be interesting to combine these things in a workflow. Dalle -> face detector -> user clicks -> stylegan3.


Online DALL-E generators often are deliberately prevented from producing good faces, because of the potential for abuse.


FYI, "ARC: FACE RESTORATION" is an AI that fixes distorted faces. It won't do miracles but it does a pretty good job when the face is just off a bit.

I suspect that overtime there will be many AIs that will target very specialized functions similar to ARC.


The article isn't loading for me, so I can't really comment on the images it contains, but I've found telling the ai to apply an impressionistic filter does wonders for removing the unsettling aspect. Obviously that limits you to a specific style of image, but I imagine there are other stylistic filters you might apply that achieve the same goal.

I could spend all day looking at the output of "impressionist cats" and similar queries.


I'm over 1000 credits into Dalle so far (I know, I know..) and you're on the money. You can go a lot further than impressionism, though. Specifying the names of famous illustrators, photographers or artists. Specifying the media used. Lens types. Colorways. Film types. Lighting. The right combinations can yield some incredibly realistic looking things, even faces, and then for the rest of it, there's Photoshop, Gigapixel, and other tools to patch things up. (I've had more luck creating 'elements' with Dalle and then montaging them the old fashioned way.)

The images used in the blog linked by OP are okay but stylistically all over the place. OP acknowledges how difficult good prompts are to write. Beyond that, though, you still need to think like an art director and establish a way to set a common style to avoid jarring the readers, and Dalle alone can't do that.


Dall-E is very “first generation” in its design and interactivity with the users.

It’s just a matter of time until setting a “consistent art style” becomes a feature of these things.

Similarly, asking the AI to produce multiple views/poses of the same thing will likely become a common feature.


> Specifying the names of famous illustrators, photographers or artists. Specifying the media used. Lens types. Colorways. Film types. Lighting.

Are training sets prepared with systematic variations in individual axes, as an alternative/addition to tagging each of millions of training images on these axes?


I'm with you. I would hate to see these images all over the place -many are just unpleasant to see.

The cover image generated for the cosmopolitan cover is stunning at first but after seeing it a few times it begins to feel uncomfortable to look at. The uncanny valley is alive and well in many of these images.


The pictures are certainly deep down in the uncanny valley, but I think they would be great for nightmarish games. In fact, game developers (and especially game artists) might be the next profession on the line, to be automated by AI.


I don't understand your assertion. Neither game developers nor artists are in any danger of being automated by AI.

Until Copilot can make the game you want, you cannot replace developers. And until you think AI is ready to replace artists in general, you won't be able to automate game artists.

That's not to say a game with assets largely drawn by AI, and heavily assisted by Copilot, wouldn't be a cool artistic experiment!


assets seem quite possibly ai generatable in the near future. you won't be using AI for the important things yet, but a AA or AAA game has a ton of random assets where for random crap that doesn't matter, but that you need to make the world feel full. that seems like a perfect use of AI assets.


Depends on which kinds of assets. You don't even need AI; plenty of games feature procedurally generated levels, art and even enemies.

This hasn't put game developers out of a job.


> Am I one of the few people who finds these generated pictures really bad?

Well they're bad at not looking like AI generated art. It's impressive, but I've yet to come across an example that doesn't look like AI generated art. A few seconds of surface level inspection and you can see the weird AI psychodelic circling effect (no idea what the technical name is - eye-ball-ification?)


There's a cyberpunk art Facebook group where some people have taken to sharing AI generated cyberpunk cityscapes, and I've been hard pressed to tell it apart from human art on occasion.

To be fair, I think this is because "cyberpunk cityscape" as an artform has become so cliché and generic, it's easy for an AI to copy it!


These are good enough for 99% of blogs out there.

Just like AI generated articles are good enough for 99% of content farms out there.


I agree. They're pretty in the same way as fractals are pretty, but still boring and bland.

I would not have any of the ones that I've seen this far on my wall, or as my blog icons.


The results tend to be residents in the uncanny valley. They are nice if you want something unsettling. They are very impressive, can be very aesthetically pleasing(especially with midjourney) but they look very alien.

Maybe part of the reason we are so impressed with those is because they break our perception of reality. It looks like the renaissance statues that are made from marble but looks like cloth.


Yeah, maybe it's okay for a tech company to be weirdly robotic, but I'd be happier without them random illustrations.


For maximum coherency, you have to make batches of 50 - 100 and pick the best one. Which can be time consuming and expensive.


This. The best things I've made have required that level of generation to get good results.


I truly wonder if there will a 'image picker job' in the future. A person who creates dall-e images for clients and picks the best for them.


Is that fundamentally different from an art director today who hands out assignments to freelancers and then picks out and puts together the best of their output?


It's not about being perfect, it's about having something that doesn't take time to produce. like the article says, searching google and stock image sites looking for a picture that very few people are going to ingest is a huge waste of time.


I would suggest scanning through the r/dalle2 sub-reddit, as the submissions there are rated. There are limitations in the way the current crop of AI generators work, but in the hands of someone who knows these and know what prompts to specify you get completely amazing results that you as a layman can't tell is AI generated (without an expert investigation maybe into pixel-level artifacts).


I had MUCH lower expectations about this article’s images, once I got it to load I was surprised, no, amazed!


They are good enough for most people and over time those details will get better until we have no need for illustrators.

Already I see website agencies and bloggers using DALL-E. What I do see is that it is easy to pick out DALL-E generated images, in that its too fantastic. Way over the top to a fault.


>Way over the top to a fault.

way over the type as a style

it's like the über-modern modern art. the next level of those goofy over-the-top meme images that make the rounds in socials

while you may not like it, you just know that this will be a thing on how to create AI-like images without AI. I used to refer to that as grade school ;P


You might be looking at the limited launched product. The fully powered product has amazing results

https://www.instagram.com/openaidalle/


> Am I one of the few people who finds these generated pictures really bad?

Bad compared with what? They certainly convey a lot more information than a randomly generated gravatar.


I tried it a lot and I think it works ok for simple, mundane, trivial prompts, but when you start to ask for sophisticated stuff it gets weird.


When asked about what jobs the robots would come for first, I would have had to say that digital artist was pretty low on my ranking before now.


It's coming for us, too.

It won't be long before most software engineer positions are eliminated while some are replaced by software "technicians" with enough expertise to command AI to generate working code. Perhaps the technicians will be tasked with building tests and some automation, but even that stuff can be delegated to AI to an extent.

This may seem far off because the present economy is accustomed to paying engineers large sums of money to write apps. Even with the retractions we've been seeing in hiring and venture capital, there's just enough easy money still there and the capabilities of code-writing AIs isn't quite there yet.

All we need is a significant market correction and the next generation of AI to wipe out a large swath of tech jobs.

The next step regardless is applying technologies like DALL-E to web design, and for said technology to be widely used, open and affordable. We won't need web designers or even UXD.

Then we won't need as many engineers when AI can solve a lot of common problems in building software. AI can do it better because it won't spend inordinate amounts of time dillydallying over next-gen frameworks, toolchains, and preprocessors. AI won't even have to worry about writing "clean" and maintainable code because those things will no longer matter.


I don't think we'll see this in our lifetimes.

For that scenario to be possible, general AI needs to be developed first.

A huge (and awful) part of software engineering is figuring out what exactly the stakeholders want you to build or fix. Sometimes, they themselves don't even know.

Dealing with ambiguos jira tickets, poorly reported bugs, non-existent requirements, missing or outdated documentation; these are the "common problems" in building software. Current AI technology isn't even close to being able to sort these types of problems today, and it won't be until a monumental breakthrough in the field is achieved.

Generating art is "easy" in the sense that art can't be wrong or right, it just is.

Generating the backend of a streaming platform? I'd like to live long enough to see it.


> A huge (and awful) part of software engineering is figuring out what exactly the stakeholders want you to build or fix. Sometimes, they themselves don't even know.

Ask any creative out there what the hard part of their job is.


> A huge (and awful) part of software engineering is figuring out what exactly the stakeholders want you to build or fix. Sometimes, they themselves don't even know.

Yeah, but that part can be learned by anyone without a CS degree.

Perhaps not everything in software can be automated, but I could see a team of 10 programmers be replaced by 1 person (programmer or not) skillful enough to control a bunch of AI software tools.


A tool that 10xs programmer productivity will if anything lead to higher demand for programmers, because we're nowhere close to developing 1/10 of the total software the world demands.


Perhaps, but that would be an indirect consequence.


>I don't think we'll see this in our lifetimes.

I already see a clear path that'd take about 20 years to execute properly. That's assuming low pressure conditions and a very large amount of funding though, both of which aren't typically present in reality.

The result is essentially what GP describes, with a path to AGI in the form of extremely competent tool AI. We're going to hit self-assembling programs before we hit true AGI.

I can't say I'm particularly excited to see such things become reality. Fortunately, humans usually find a way to fuck things up. Our species' collective incompetence is the largest barrier to AGI currently, which may be a blessing in disguise depending on how you look at it.


The backend of a streaming platform is trivially summoned by logging into Twitch. GPT3 has already demonstrated its "understanding" between a problem statement, a variable, and it's value (and if you haven't, it's worth finding the tweet/gif). Bridging the gap between the words "streaming platform backend" and an ffmpeg command line may involve a bunch more details, but the gap between the two is only going to shrink.


AI is great for recommendation systems and art because they are fuzzy. "Good enough" results are relatively easy to achieve. There is lots of tolerance for errors, because human preferences are flexible and fuzzy themselves.

Engineering is a different ballgame... If anything, all the code monkeys will simply become QA monkeys/test-engineers, because you need to be really sure that your black box algorithm is actually doing what you think it should be doing.


> software "technicians" with enough expertise to command AI to generate working code

People keep trying to make simplified programming environments for significantly less-trained people and they keep failing. Is mixing in an AI actually going to make it easier to get a result that has no crippling bugs?


No, but when have crippling bugs ever stopped software businesses from shipping it anyway?


>> People keep trying to make simplified programming environments for significantly less-trained people and they keep failing. Is mixing in an AI actually going to make it easier to get a result that has no crippling bugs?

Yeah. I've even worked in one of those environments for a year (not my choice).

I'm of the opinion those kind of environments won't ever work. They'll either be:

1. Extremely cookie-cutter (e.g. make a clone of our "standard app" with a insignificant little tweaks).

2. Require software engineers to get anything useful out of them, and those engineers will feel like they're working with one hand tied behind their backs (or banging their heads against a wall).

IMHO, one of the main skills of a software engineer is translating user requirements into technical requirements that work and understanding when they work. I don't think skill is automatable without a fairy-tale AGI.

> No, but when have crippling bugs ever stopped software businesses from shipping it anyway?

A lot? Depends on your definition of "crippling." A software engineer will gripe and say, "I don't want to use this;" something that's awkward but the people who use it can still get their work done; or the system literally incapable of performing its function?


And how is it going to increase information in the output by having AI involved, if these AIs aren’t actually thinking and pouring out of their own entropy source into outputs?


Judging by the current state of DALL-E, the generated software will look good at first impression, but have lots of weird bugs when examined closely. So yeah, not much different to current software dev.


"Physical" engineering fields will probably come first... think AI-generated architecture, with AI-generated structural engineering, plumbing, electrical wiring, etc... with human-guidance of the generative process, and human-review/accountability of final output. Amplification of humans, not obsolescence.

In software, yeah boiler-plate and function-level code-generation... I could also see generating trivial UIs for CRUD apps, or no-code data-pipelines for small businesses... maybe even generating high-level architectures for new services... but we're far off from AI auto-generating code for enterprise applications or foundational services. The differentiation being making changes within an existing complex domain/code-base, in contrast with generating new assets from nothing.


Most of the math for structural engineering is already done through software, we just don't call it AI. The difficult part and valuable part of being a good structural engineer is translating requirements and dealing with clients. The actual math and engineering work is often not much more difficult than what's done taught in their undergrad, and much of it is offloaded to designers anyway.

Source: My family owns one of the largest civil engineering firms in my home province.


It will never come for us. You think it will, but that’s because you don’t understand software.

Pick any random Jira ticket for a large software project. Could an AI understand and implement that feature into a larger project? Can it correctly deploy it without interruptions to production jobs? Will it correctly implement tests and decent code coverage? If there are regressions will it know how to go in and fix them? If there are bugs reported by users will it be able to investigate and accurately fix them? What about when multiple branches of feature development have to be merged, will it know how to do it correctly? Will it know how to write high performance software or just use shitty random algorithms?

If it can’t do these things AI is basically useless. Because this is basically 90% of software development.


A plane flies and is nothing like a bird.

Most likely, the APP-E or GAME-E, given a prompt generate an application or game, will not generate C++/JavaScript/Swift/Kotlin but directly target the pixel space, running in a 60+ FPS loop a single "function" such as `nextFrame(currentState)`.

It will probably be here in the next few years: write a prompt such as "2D game like Mario but with butterflies" and receive a 2GB blob which opens a window accepting inputs and changing the pixels accordingly. Or, something more serious, a prompt like "invoicing application following the laws of France, Material Design, store the database in AWS using <token>". APP-E or GAME-E doesn't need to totally replace software development, just be good enough to replace in 99% of use cases.

Bugs/tests could probably be solved by some mechanism for injecting localized prompts: given an already generated binary, fine tune it accordingly to a list of other prompts.

As for deployment, it's already pretty much solved with the CI/CD solutions galore all around, not sure why you would need generative statistics for it.

What DALL-E offers is a glimpse of the next 30 years, and probably 99% of the infrastructure required to run it to its full potential is not here yet. Just as in 1992 (3 years after the HTTP proposal, but 2 years before the launch of Netscape) there were only glimpses of what a connected world would look like.


If we had such advanced AI we wouldn’t ask it to build programs, we would just tell it to do those computations directly. So instead of asking for an invoicing application, we’d just ask to generate an invoice to whatever parameters we need, and to remember the reusable data for next time.


Sure, who knows. Although humans are allegedly bad with 7±2 parameters, hence a nice user interface is still required if humans are to be kept in the loop.

My point was that something like APP-E or GAME-E seems very plausible in the near future and it is more likely to render pixels with the underlying logic encoded in an inscrutable sparse matrix, somewhat the consequence of a beefier DALL-E with regard to the data set, the learning modalities, and the attention span, than to write programs to be compiled/interpreted by any current language stack.


I guess the difference is, with the art the human can immediately reject/accept and iterate. And a bad image can be crappy, it doesn't break anything.

With software it might take days of testing to verify the result, and then repeat that for every iteration. Would be cheaper to build the thing!

Where AI might work is in some restricted subset of software, like a web CRUD app where you say "I want an app that stores a TODO list with dates". With the constraints of it being crud, it just needs to AI the database and arrangement of fields and so on.

The AI is not programming so much as it is choosing which "rails-like scaffolds" to initiate.


Simple CRUD apps are mere toys these days. We don’t even need AI to quickly generate them, it can be done with a couple scripts. The only people that would be replaced by an AI that specializes in CRUD would be recent CS graduates, junior developers.

The serious engineers are all working on things that go far deeper, and they could never be replaced.

If you want to build a business big enough to be listed on the NASDAQ, you need real developers, and you need to pay them real money.


Writing software is like writing novels: putting words together to make sentences is easy. Making the story make sense is difficult.

One could think that much of art is just pretty form without sense and that is why DALL-E works.


I don't think that assessment of art is accurate.

Most art is not "pretty form without sense". It actually has sense and meaning more often than not, so we can debate what a particular piece "means".

The difference with engineering is that art's meaning is way more subjective, and that if I "miss the point" or simply disagree with the consensus on its meaning, this doesn't make an airplane go down or a nuclear reactor to melt down.


I don't think it will be like that, for two reasons.

One, coming up with a correct description of a program is what computer programming actually is. Implementation is something we're always looking to do faster, so we can describe more behaviors to the computer.

Two, we're nowhere near the scale of software production which would clear market demand. If everyone who writes code for a living woke up and was ten times as productive, there would be more churn than usual while the talent finds its level, but the result would be ten times as much code per year, not five times and 50% unemployment.

Today I wrote a little bit of code to generate a prefix trie and write it out in a particular format, with some extra attention to getting the formatting 'nice'. This took me about three hours.

It won't be long before something in the nature of Copilot could have gotten this down to, maybe, a half hour for results of the same (minimal, acceptable) quality.

Wonderful! Can't wait, I'll be six times as productive at this kind of task.

This might make it hard, on the margin, for some of the more junior and glue-oriented developers to find work, but I think the main result will be more software gets written, the same way using a structured programming language got people further in a day than writing in assembler did.


I think when AI can do the full job of a programmer we'll have reached the singularity. Programmer will be the last job to go.


Time for me to open up my "AI Code Refurbishing" shop and specialize in fixing all the disasters these "AI technicians" make.


Imma be honest, working as an artist who has to come up with Dall-E prompts and as a programmer who has to maintain a codebase slapped together from GPT-5 output sounds equally horrifying. I think I'll stick to my grug brain engineering.


I am personally bearish on this assumption unless a few hurdles are reached. Being a software engineer involves a lot of translation of intent from a required feature into an efficient and maintainable implementation.

In a good number of cases it is more difficult to communicate what needs to be built rather than actually building the end product.

The recent work with DALL-E 2 echos a similar problem, coming up with a descriptive prompt can be difficult to do and needs fine tuning to be done. Not unlike trying to communicate with a graphic designer your expected intentions and giving similar works to draw from.


I agree. Most people fail to see it, because they see all the effort they need to put into producing good results (regardless of their actual job, BTW). Programmers keep thinking their job is secure, because, after wall, we are the ones, who write the software. Even if it's a ML system. (But ML systems don't necessarily need much coding.)

However, software development is probably the most thoroughly documented job, the job with the most information online how to do it right, the job with the best available training set. There is a lot of quality code (yes, bad code too), a lot of questions and answers with sample code (stackoverflow...) available. Maybe we've even already written most the software needed in the near future, it's just not available to everyone who needs it (because no one knows all the things out there and also these might be in several pieces in several repos).

Now the one critical thing I think is still needed, based on how actually we create software is an agent that has reasonable memory, that can handle back references (to what it has been told earlier), i.e. one that can handle an iterative conversation instead of a static, one time prompt.

This might be a big leap from where we are now or it may seem like one but AI/ML keeps surprising us for the past decade or so with these leaps. Another thing that may be needed is for it to be able to ask clarification questions (again, as a part of an iterative conversation). I'm not sure about this latter one, but this is definitely how we do the work.


The problem with that theory is that writing code is easier than reading code. This is generally not the case in other professions. It is definitely not the case for an artist.

You still need correct code, and the halting problem says you can't prove whether code does what you want it to. At the end of the day, someone needs to be able to go in and fix shit the AI did wrong, and to do that you need to understand the code the AI wrote.


> The problem with that theory is that writing code is easier than reading code. This is generally not the case in other professions. It is definitely not the case for an artist.

> You still need correct code, and the halting problem says you can't prove whether code does what you want it to. At the end of the day, someone needs to be able to go in and fix shit the AI did wrong, and to do that you need to understand the code the AI wrote.

This might have been your point, but chances are the "code the AI wrote" will be an unmaintainable mess, so "fixing it" means throwing it away and re-doing it.


The question is how many developers will fix code after AI and how many developers will grow potatoes.


AI has been over promising and under delivering for 50 years.

There's a reason why the general models aren't being released. The second you look under the hood and start poking the unhappy paths you see that it doesn't understand anything and you're talking to something dumber than a hamster.


There's a weird tension between people who say saying "AI is overblown" and people who say "this is the most magical thing I've seen in my lifetime".

I lean towards the latter but with a healthy dose of "it's deeply weird and hard to get anything useful from". But that doesn't make it any less magical.

And no - it's not "intelligent" in any human sense.

But I can't relate to people who pooh-pooh it as if there's nothing exciting happening. Either they are deliberately cultivating a dismissive air, or they are deeply jaded and weary.

EDIT - There's a 3rd option. People are making a rhetorical point because they perceive a need to correct an imbalance in the general mood. This is actually the most likely explanation and is often under-appreciated as motivator in public statements. I've noticed it in myself frequently.


AI can be overblown and it can also be magical.

This was true for state of the art in 2010: https://xkcd.com/1425/ today you have a free phone app that does both. Of course it also classifies a spoon as a large breasted robin which is why you need a human in the loop. It's even truer in programming.


That xkcd actually claims the opposite: "some things that many people assume to be trivial CS problems actually require advancements in AI"


That's what I said. The converse is that despite the huge advance the model is still incredibly fragile and quite useless outside the niche for which it was trained for.


Programming requires far more breadth and precision than 2D art.

I think that in the very long run programming work will be automated, but by that stage we will either be post-scarcity or reconstituted in computation substrate.


You say that with such confidence.

I'm looking for ways to hedge my reliance on my skills.


I'm confident that society would be so radically different that trying to predict and prepare is a fools game.


>This time will be different.

The 20th time you hear that is when you stop caring.


programming is going to get automated with language models in <5 years

however, I find that my job (SWE) is about 1% programming and 99% strategizing, designing & communicating.


Really? How much money are you willing to put on the claim "LLMs are going to produce >80% of newly written production code in 5 years"? If it's less than your yearly salary, then even you don't believe your own assertion.


For github & google, I'd put my own money on it being >80% in 5 years. Maybe not a whole year's salary though, I'm a big baller


Well, you will lose but I basically don't believe your claim is anything more than hyperbolic bullshit as no one who is working as a SWE is actually spending only 1% of their time programming as per your initial comment.


some software engineers spend most of their time designing, communicating with other teams, and managing operations. It becomes the case as you get more experienced, depending on your staff engineer path.


This is a truism that every senior engineer understands. There is no way to get below 10% coding unless you move into people management/product. At this point, you are no longer a SWE (except by title, maybe).


this thread is funny because I ended up spending my whole day yesterday and today heads down coding


> "This may seem far off..."

After experimenting with GitHub Co-Pilot I can see that day being 50% - perhaps even just 25% - as far as it used to feel.


Wouldn't millions of unemployed developers start creating software that would compete/overthrow existing software companies?


Oh people don't realize that all those movies/series about dystopian societies were blueprints for the elite.


AI still can't translate lol


Maybe, but this is replacing fifty cents of stock photo, not digital artists.


Who do you think created that 50-cent stock art?


There is a semantic issue going on here, "digital art" I thought meant fine art, not illustration for blog posts


Probably not a digital artist. Is there a big fraction of that on stock photo sites?


And digital artists aren't still pretty low on your ranking?

I don't know, after all the predictions about self-driving cars, I'm cautious. Especially considering that back then, it almost seemed obvious that we'd have self-driving cars by now. Cars were certainly capable of driving themselves back in 2016, it just seemed like we needed to iron out a few kinks. How long could that possibly take?

Now, I have no idea when it'll happen.

I'm not necessarily saying that it'll take AI forever to do what humans can do. Rather, I think its very hard to make good predictions with all the hype slightly deceptive marketing.


I think self-driving cars might be one of the hardest domains ever, considering they need to do the right thing (or at least avoid doing the wrong thing) 99.99% of the time to be even considered viable.

An AI artist can get away with a 5% success rate and still be considered viable for replacing humans.

Likewise, and AI programmer can be right only 50% of the time with expert oversight (someone sitting there fixing the code), or 90% with non-expert oversight.


Self-driving cars are reality for a few years already. Issue is with laws and people, much less with tech.


They issue is absolutely with tech. Cars can drive themselves right now in good conditions and with a human taking over ASAP in case it fails and even that isn't working 100%. Put that self driving car on a mountain road or in deep snow and see what happens.

Also, if your bar for accidents is "slightly better than a drunk driver and even less accountable", sure, then we can fully deploy that right now. Unfortunately this really isn't sufficient and car companies are fully aware. There's a reason Mercedes made headlines by taking responsibility for the car for ten seconds after disengaging the auto pilot and why they're the only ones who are doing this so far.


I think this is a comment for which "sources needed" is kind of the bare minimum.

To most people, the "readiness" of self-driving cars doesn't even come about when they would accomplish parity with human drivers across all situations, because they should be better. And we're not even close to parity with human drivers across a large swath of common situations.

Unless you have links showing otherwise...


They works great when conditions are ok. As soon as there is an outstanding parameter they fail spectactularly.

We can also talk about AI support service crap with accounts banned/locked without the user having any way to know why or prove he didn't break any terms of the contracts.


Plus the approach. I'm not sure we need AI at all for self driving.


Same. I would have thought the arts would be the last to move to AI. What do you put at the end of the list now? It's not trucking. Or ordering. Or anything related to porn. Eliza is 50+ years old, not actually a good replacement for a psychotherapist yet but I would imagine it could go a long way.

I'm biased from the terrible experience I had trying to get my kids to learn online in the pandemic, but I think schoolteacher might be one of the mass professions that is least susceptible to being AI Engineered away.

Ethicist is probably a safe career path too, but there aren't that many of those. And Politicians will of course prevent robots from taking over their jobs.


The reason it's focused so much is that art has incredibly low stakes... and people don't want AI making any seemingly important decisions...


God it’s so fucking depressing to see all you techies debating whether or not the skill I have dedicated my life to getting good enough at to earn a quiet, modest living with should be automated away or not. And insisting that surely your jobs are too special and complex for this to ever happen to you.

At least I can take solace in the fact that for now these things aren’t gonna be taking a bite out of the furry art commissions I like to take, since that’s way too associated with crazy cartoon porn for them to not censor relevant keywords.


Of course it should be. You aren't entitled to my money for your services if a computer can do a good enough job for free. And human history is full of examples of innovations displacing humans and temporarily unemploying them until they found other jobs. That's basically what "productivity" is about in economics, and overall it's a good thing we're not still in the stone age in the name of protecting people's jobs.


Please save this comment somewhere and read it to yourself when the successor to GitHub’s Copilot makes you start worrying about your ability to earn a living. Thanks in advance.


Why would this change anything? If I 'm not willing to do my job for the salary offered I have to change jobs. What is upsetting about this?


There absolutely nothing upsetting about that to the CEO of your company. I would expect the destruction of your lifetime of accumulated economic value at an age when you can’t possibly retrain to be able to support your previous lifestyle would be rather upsetting to you. Or if not then I would not call that a typical emotional response.


I would be upset, but that is separate to the question posed of whether we should be trying to innovate and automate. The answer is that innovation and productivity should be sought, even though it's understandably not fun for the people whose skills are being commoditized in the moment. At no point in history was it a good idea to purposefully stop innovation just to preserve some job that was about to be automated. The cumulative impact of jobs being automated and productivity is why we have such easy lives today.


And that is all well and good until the pace of automation outstrips the community's ability to absorb the externality. Then it's not going to be the AI dream future, but heads on spikes. We have hit those points in history before.


When has the pace of automation outstripped the ability for the economy to provide near full employment?

Besides. Even if true, the answer isn't to stop innovating.


Not to mention that the whole goddamn trick here is basically taking the output you and your peers have created, anonymizing it, throwing it in a blender, and then acting like the AI has generated something new and humans aren't needed anymore.


Yep!

“Draw me Elon Musk fellating a Bitcoin in the style of an artist who spent decades refining their craft and producing enough art to train a neural network to make something resembling it, and completely ignore any potential copyright issues because it’s maybe right on the edge of sufficiently transformative, and anyway we have no idea what’s going on in these damn black boxes any more.” That’ll be one compute token, please.


What about artists who are inspired by other artists?


Thats a bit unfair. The discussion is not about what should happen. But a prediction of what will happen, given these technologies exist. This is a tech forum after all, and speculation about tech and society is a big part of that.

Personally my societal concern is yet another industry where we had multiple small jobs will be ruled by a few conglomerates.

Small business are like democracy in a free market. And we keep evolving to it all ending in the hands of a few.


Wait until the world becomes oversaturated with AI generated imagery and then make a killing when people realize that real artists can actually produce something original.


I can only hope that this doesn’t happen after it has become impossible for young artists to make enough money with almost-pro work to be able to get in enough practice to become pros worth paying for. Finding your particular creative voice is a long process; the Internet opened up a lot of opportunities for people to do that.

I also hope this oversaturation doesn’t happen while I’m still alive, or that it happens as quickly as people became obsessed with Mandelbrot sets. You saw them everywhere and then suddenly they weren’t cool any more, and it always feels faintly embarrassing to see one used when you go back to look at something from that era.

I’m fifty years old and I doubt the people making these AI art tools are gonna send a single penny towards supporting and/or retraining the artists they displace, even the ones whose names constantly show up in the lists of things to type into the text box to get a better image.


I sympathize greatly because I think over the short term it's going to be worse all around for artists. I think anything that gets chucked into the ML wood-chipper is going to get old fast.


Have you seen the Dall-E outputs? If anything they do better than the vast majority of artists in terms of originality. Sure there's probably some people out there who can do better, but no where near the amount of people employed in the field today.


Err... more original than the artists that originally conceptualized most of the things it's creating? No, I don't see that at all. Certainly not better than the "vast majority".


Hard disagree


News at 11: jobs are taken away and replaced by advances in technology.

You are not special. I am not special.


> insisting that surely your jobs are too special and complex for this to ever happen to you

This insistence bugs the crap out of me, and shows just how arrogant tech-types (and even more so, business-types) can be. The late David Graeber noted this attitude, and while he's talking about investors/entrepreneurs/financiers, it could equally apply to the cocky software engineers I've met over the years:

"It's possible for futurologists to imagine robots relpacing sports editors, sociologists, or real estate agents, for example, yet I have yet to see one suggest that the basic functions that capitalists are supposed to perform, which mainly consist of figuring out the optimal way to invest resources in order to answer current or potential consumer demand, could possibly be performed by a machine. Why not?"

I optimistically believe that art will never be automated away like software engineering has the potential to be. The human creative element is so core to what art is that replacing it with a machine misses the point entirely, even if a machine could fully replicate an aesthetic.


Trust me, we (the techies) are in strong denial about how much time we have left. I've been programming since the late 1980s and have learned everything from VHDL to Clojure and give us till no later than 2040. Realistically, more like 2030, due to the billions of dollars being thrown at AGI for finance and other monetizations. Of course, money won't be worth anything after that, but I doubt that will stop anyone.

In my own life, I've decided to transition away from shared truth towards manifesting the reality that I want to live in. I try to help people now, I meditate a lot about humans becoming aliens, I try to be in the moment whenever possible and be thankful for consciousness. But I no longer put my energy into the ego-based materialism that captured tech. Since wealth inequality can't be stopped, I feel that the only salvation lies in nonattachment.



As companies in spotlight make it impossible to generate gore or porn, will more commission artists' specialization be in that area?

There are already enough offbeat companies dedicated in generating porn, I am sure.


Don't worry, while the rest of the industry is busy making artsy abstract images of chimpanzees surfing on a wave of coronavirii, you can quietly work on UX design I guess.


God that’s a fucking sure route to suicide right there. From sitting at the top of Maslow’s hierarchy of needs to shitting out buttons for yet another fucking weather app on Dribble.


> furry art

I can‘t wait for a dalle2 level model trained on all of e621, rule34, x-hentai, furaffinity, sofurry, inkbunny and u18chan. It‘s gonna be great.


i feel for ya mate, i can see it coming for me too


I would add to "what I've learned" that DALL-E 2 will start to be used to generate real ideas for real world 3d objects.

Here are cups rendered in the style of a famous architect.

https://www.reddit.com/r/dalle2/comments/w7buyx/a_coffee_cup...

Dalle 2's influence will be felt outside the graphic artist realm too.


This is very interesting, but why are all the images just showing up as blurred out pre-loads for me? Makes it a lot less communicative, since it's literally about the images!

I mean, I'm guessing these aren't the intended images, since you don't need DALL-E to generate blurry splotches!


IMO, Using AI for image generation is the trend, whether we like it or not. But for me, as a software engineer, I still insist on creating pictures for my own blogs or presentations, etc.

I'm not a painter, of course I paint badly. But I find that it is not so difficult to create a picture that expresses my ideas. Mostly I try to use https://excalidraw.com/ for sketching hand-drawn, and find the free stock pictures from https://www.freepik.com/.

I think this is another joy of creation, like programming :-)


Archive.org: https://web.archive.org/web/20220808212448/https://deephaven...

Unfortunately the images are cached at a very low resolution


Visit their main blog page to see a subset of the resulting images all at once

https://deephaven.io/blog/

https://deephaven.io/blog/page/2/

etc


I find the images fairly distracting and they take all the focus on the above page. The artwork also isn't very consistent which makes it feel like a jumbled mess. When you click a post the image takes the entire screen and pushes all the content below the fold. It's more like browsing a community art portfolio instead of a tech blog.

I'd prefer smaller thumbnails or icons that give more context to the actual post. This way they could add some benefit, such as helping to visually categorize the content. As of now, they're just a bunch of random illustrations taking up valuable screen real estate.

That being said, thanks for sharing, it's interesting to see an example of someone integrating DALL·E 2 into their workflow.


Looks like a hipster LA artist drew them all.


> Maybe it was because this was my first attempt, but with 100 more posts to go, I hoped I could get better with practice. It would be super cool to just feed DALL-E a whole blog post and have something great pop out, but even with some GPT-3 magic we probably aren’t there yet.

I see a business opportunity here. Feed text into GPT-3 and have it generate DALL-E prompts to make appropriate images.

Then you have it do the same thing for a children's book.



I've done something similar a while ago, turned out great (according to my toddler): https://www.reddit.com/r/dalle2/comments/ueizwz/i_printed_a_...


Ugh, this immediately drives home the realization that representative images are soon going to be devalued and useless, to the point that we'll all be ignoring them soon. Possibly even stripping them with ad blockers or similar tools.

I actually think it's really awesome to be able to do this with a series of blog posts, and even if you look past the stylistic inconsistencies and oddities, this particular usage is good and adds value.

Which is kind of the problem. Relatively low cost, currently high benefit? It's going to be driven into the ground.

We've seen this over and over again. Some reliable form of signal, or of value, becomes inexpensive enough to produce that it gets commoditized, monetized, and weaponized against us all.

Email is a major productivity advance that gives a low-friction way of communicating for mutual gain? Well, now we're drowning in spam and phishing attempts and people won't read random unsolicited messages—if they even make it pass the automated filters. Same for text messages. Bold images and lettering used to be good for highlighting and accentuating important information. Now we don't see them, even if they make it past our ad blockers, because the neural networks living in our skulls know to filter them out as negative-value advertisements.

The same thing will happen here. Nearly all blogs will soon be sprouting cutesy images to go along with the posts. Initially, many of them will be useful and add value, suggesting a metaphor or analogy or simply providing a visual anchor to make the content more memorable.

But they'll quickly become expected and necessary and we'll have the usual race to the bottom. Everyone will have some image because it boosts engagement by 8%... wait now 6%... oops it's too common, we're awash in crappy irrelevant images just added for the boost, which is down to 2%... oh crap, now the absence of an image is a good signal for content quality, we're at -1%!

(If you put work into the prompt and curate carefully, it will still be a net positive to your content. But it won't matter for long in terms of traffic/engagement, because everyone will be mentally ignoring it.)


> Email is a major productivity advance that gives a low-friction way of communicating for mutual gain? Well, now we're drowning in spam and phishing attempts and people won't read random unsolicited messages—if they even make it pass the automated filters.

Actually, in my experience, and the experience of most of the people in my life, email is a valuable tool that we consistently use, and rarely have any problems with. Spam is a minor inconvenience , at most. (And certainly less of one than the overwhelming number of ads on Facebook.) Phishing is almost non-existent, and very easy to ignore when it pops up a couple of times a year.

In fact, without email, I think the internet would be pretty horrible.

People love to criticize email, but it's actually quite great. It's distributed, low cost, easy to use, and nearly universal. It has mature tools that make the obvious problems (such as spam) not really a problem. And most email senders actually make it easy to unsubscribe. I regularly subscribe/unsubscribe to email lists -- and it's nice having control over what comes into my email inbox.

Email, frankly, is what more of the internet should be like...!


> But they'll quickly become expected and necessary and we'll have the usual race to the bottom. Everyone will have some image because it boosts engagement by 8%... wait now 6%... oops it's too common, we're awash in crappy irrelevant images just added for the boost, which is down to 2%... oh crap, now the absence of an image is a good signal for content quality, we're at -1%!

Too late. It already started happening a while ago. Tons of blogs with annoying animated gifs and now browsers have the ability to block them.


Yep. My analogy for this kind of effect is McDonald’s (and other fast or prepackaged food). Fast food never eclipsed good food because, well, it’s not that good. But it spreads. It’s just the result of unfettered capitalism (tech is no different in this regard to any other industry).

But it’s also a culture thing…like people trying to convince us that mediocre stuff is good or cool. I mean in this case it’s new because a machine is doing it, but what are OpenAI’s real economic motivations behind all their press releases I wonder…?



That is missing the images.


I assume those who currently have access to DALL.E 2 are legally prevented from selling second-hand use of their access? I'm currently on the waitlist and I've been curious if there's any way to bill access like this as "design consulting" or something?

Posts like this are absolutely fascinating!


is there a way to know for sure if something was generated by it?


Provably better than 80% of fiverr "artists" ;)


Images look very nice but also, browsing the blog archive, pointless, since they mostly have nothing to do with a post's context. The later is something that low effort blogs tend to have in common and I immediately go away when see it.


I take your point about images and increased engagement but I'm so over useless imagery in blog/article content. I think I first really noticed it with Medium. So many pointless header images that very rarely have anything to do with the article. Just visual noise.

My prediction is that a lot of blogs will do what you've done too, and then they'll all look/feel the same. New models will come out I guess, but then they'll proliferate too and everything will look the same again. And then maybe to differentiate, those few that value it/can afford it will make the effort to find actually relevant images or commission artwork.


How do you think copyrights around these images will evolve? No copyrights? Full ownership to the one creating the prompt? Copyrights based on specific context?


I find it quite ironic that a AI company like OpenAI seems to use word lists to prevent inappropriate content creation.


Private word lists right.


I guess these images cannot be copyrighted for a blog post since dalle could generate either the same or very similar ones.

Some less capable algorithms are quite repetitive in such aspect. Perhaps after some time it will be like blog theming where everyone has the same style.


I think this just shows the pointlessness of the junk pictures on posts, like the ones Medium apparently requires.

Now that pictures are cheap ven NYT articles are stuffed with photos that don’t add to the article.


> But how do you even have a signature style when it is the AI generating your images?

Style transfer? It was one of the first CNN applications, and it would be nice to see that included in image generation pipelines.


I've been using Crayon AI (similar project, free) to create images for a pet hobby project, free and for my own private enjoyment.

Think a D&D RPG: I'm writing a bestiary and book of spells, and because I don't care about an exact fit, I can prototype tons of bizarre creatures, wizardly types, weird spells, and colorful scenery. It's pretty cool, and way easier than drawing the prototypes myself.

The one thing these AIs really struggle with is faces. It's hard to get a human without a monstrous face. The best you can hope for is an abstract or blurry face.


First time to hear "Prompt engineering" but I feel familiar with this term. It's my daily job to adjust google search queries to get the information I need, even though sometimes I don't know what I need.First time to hear "Prompt engineering" but I feel really familiar with this term. It's my daily job to engineer google search queries to get the information I nee.


Prompt "engineering" will last a couple of years, until the UI improves to the point any dummy can do it, just like with instagram filters.


Right, the best prompt engineers might develop the best prompt UIs.


The "Bonus killer feature request for Google" kinda exists in PowerPoint in Office365, except that it's based on stock pictures, rather than synthetic images. It selects a picture based on the text currently present in the slide. So, unlike the "feature request", you don't even need to type a description. The feature is called "Design Ideas".


I used this feature so much when I was in school.

And I was thinking exactly this.


I thought it was going to autogenerate a prompt pased on the article's text. This seems no better than using stock photos. Actually worse.


> Our approach so far has been to spend 10 minutes scrolling through tangentially related but ultimately ill-fitting images from stock photo sites, download something not terrible,

This just seems… Not interesting, well done you’ve recreated images that clearly even before AI everyone ignored.


Serious question, although everyone seems to avoid it, when will this or a similarly advanced system allow porn? In fact, the porn companies have been awfully quiet for a while. What are they doing? Usually they are the ones on the brink of new technology.


Already out


I'll wait to read it when it actually comes up but until then I have to say when I finally got access to the DallE2 beta, I was underwhelmed. The mini-Dalle, while not as good in the image fidelity department, is much more fun.


The beta is publicly available now, everyone has to pay tho


The problem with using DALLE2 is that seemingly the only publicly accessible version available on the "hugging face" site constantly hits limits.


I'd like to see O'Reilly switch to using DALL•E to generate their book covers, using their existing corpus as training material.


What about royalties to the designers / artists that created the art being used to train these systems?


Every time an artist looks at a painting, gets inspiration from that experience and creates an image themselves, they need to pay royalties?


Gety will become a trainingset vendor.


Exactly, "We don't have exctly what you ask for but maybe you would like these GettyAI(tm) generated images:"


I signed up months ago for the beta and I'm still not in.


Me too, that's awkward


They choose who to accept based on their shared accounts (linkedin, instagram, etc).


Real name LinkedIn probably. Basically the user needs to be traceable, that’s the Microsoft/US “safe anything policy” right?


Did anybody try to generate desktop icon sets for Linux?


I haven’t seen anybody generating “useful creations” except cover albums and thumbnails, yet.


What service did they use for $45? Is this legal to include in published works too like a book?


From the Dall-E terms of use:

> Use of Images. Subject to your compliance with these terms and our Content Policy, you may use Generations for any legal purpose, including for commercial use. This means you may sell your rights to the Generations you create, incorporate them into works such as books, websites, and presentations, and otherwise commercialize them.


They must be well-funded if they're leaving money on the table like this. I'd pay a few a dollars for commercial use.


Legally its still unclear if the image can be copyrighted. If you use it commercially, others may be able to use the same images if they’re generated by DALL-E.


It takes several tries to generate something actually meaningful, so they're definitely making money by having people buy credits to run it


You should've invested in CloudFront first because your site isn't loading.


I know. I wasn't actually expecting to hit front-page. Trying to reinforce it best I can.


Fwiw, in my experience you don't actually need a CDN just to survive HN. It may be enough to just make sure you're not hitting a DB on every request; ideally you'd be caching the HTML output wholesale (via static site generator or otherwise)

For reference: with cached HTML, my single Node.js process running on Heroku's cheapest tier has weathered the front page multiple times without breaking a sweat


I switched backends a bunch of times because everything I tried (Go stdlib HTTP, Tornado, etc.) kept getting taken out whenever I would hit the front page, either due to CPU overload or some sort of resource leak. I ended up using Warp+Wai+Servant (https://github.com/yesodweb/wai) and it has been smooth sailing since then off my $3/mo VPS. It can take thousands of req/sec without flinching (which is higher than what you see from top of HN - that maxes out at a few hundred req/s).


Did you use some software/service for load-testing the alternatives?


Yes, I first tested locally with httperf and some other tools. I took it as a good sign when the load testing tools crashed before my server did. Then, I found a few services by searching for something like "website load test" and using their free tier (which would typically generate something like a few hundred req/sec - sufficient to simulate HN).


What VPS are you using and can you recommend it?


Vultr. They have been totally solid for me. Nothing fancy - just a reliable cloud VM. Have had no reason to look for alternatives for web hosting.


Right, like something like Varnish.


Hadn't heard of Varnish, but yeah, it looks like a good solution the OP could probably layer over their current setup without too much trouble


At that point, why not just have a static blog hosted on an AWS bucket?


For me: it's less to manage, it's less to learn (AWS is a nightmare from my perspective), and I enjoy other benefits like the fact that one codebase can generate and then serve up the site, and the fact that it's vendor-agnostic (just clone/npm install/run). Also allows easy customization of headers and redirects, allows for the odd dynamic route, and makes local dev/previewing super simple

OP may or may not feel the same! Just wanted to communicate that a simple server can definitely do the job


Right. So a CDN?


A CDN is its own thing- it's distributed across a provider; it can't just be served off a simple box. It requires having or gaining familiarity with a specific provider, as well as other constraints like you have to statically export to the file system (can't just cache responses in memory), and you can't have any dynamic content without standing up a separate server, etc

Makes sense for a lot of things, but it comes with downsides, especially for hobbyists! I've found I prefer sticking with a simple server for my website, and OP might find it's easier to do that too


That's not at all true. A CDN is a content delivery network. There is nothing that says it isn't a network of a single host on the same machine as the original content.

It's just a cache that returns content faster than the original content.


Nope, that's just a proxy/cache.

> A content delivery network, or content distribution network, is a geographically distributed network of proxy servers and their data centers. The goal is to provide high availability and performance by distributing the service spatially relative to end users.

Emphasis mine. CDN implies some sort of edge-hosting topology (or at least more-proximal than the main servers).


Not sure why you think CDN implies geographically distributed edge hosting. Are you getting that from Cloudflare's marketing? A company that sells a geographically distributed network?


CDN is a term used for years before Cloudflare was funded[0][1] to mean exactly that. Specifically[0]:

>CDNs act as trusted overlay networks that offer high-performance delivery of common Web objects, static data, and rich multimedia content by distributing content load among servers that are close to the clients.

and:

>CDNs first emerged in 1998 to address the fact that the Web was not designed to handle large content transmissions over long distances.

[0]: https://ieeexplore.ieee.org/document/1250586

[1]: https://link.springer.com/book/10.1007/978-3-540-77887-5


No, I'm getting it from wikipedia. It has nothing (directly) to do with Cloudflare or any other commercial interests - bittorrent is a content delivery network. You could run a private CDN. The whole purpose is moving the data to be consumed away from some centralized primary host, to nodes which are more proximal to the data consumer (either spatially, or solely in terms of bandwidth, torrent bandwidth is decoupled from the primary server). Bittorrent sort of automatically works out "proximity" by pulling from the highest bandwidth seeders. Also it's geographically distributed, providing redundancy and availability, which is arguably the more important part than proximity.

I think the criteria are that it a) delivers/distributes content b) is a network, implying multiple nodes c) lowers the latency and/or bandwidth cost of data consumption, by d) leveraging geographically distributed redundancy and/or proximity. I think the key feature is geographically distributed redundancy which differentiates it from a regular cache.

https://en.wikipedia.org/wiki/Content_delivery_network


Aws clountfront is configurable to be one region or global. So would cloudfront not count as a cdn in your mind if it’s configured to be a single region?


A region has multiple access points.


For reference, it's a fully static site on a low-end shared host. The post had quite a few images, which were pngs from DALL-E, but I've just now recompressed as smaller jpg.


Just curious, is there any general number range for how much traffic a front page post might get? Less than 10k, 10-100,000, 100,000+, etc


I'll see on the order of 10k-25k hits (hard to say exactly, most of HN uses adblockers/tracker blockers and I use CloudFlare for caching) from an article on the HN frontpage. It's not that bad, and I could almost certainly serve it off my colo'd server without any trouble - bandwidth just isn't that high.

But as my blog is entirely static (except for the comment threads, hosted on my Discourse forum), I just let CloudFlare serve it. I had to do some tweaks to the configs to say, "No, really, cache everything!" (it doesn't do that by default for a range of very valid reasons, none of which apply to me), but once that change went in, I'll see 98.5% or higher "served out of cache" ratios when I'm seeing a lot of traffic from HN or somewhere.

I'd originally designed it to be hosted out of a Google Cloud bucket with CloudFlare (egress traffic is cheaper that way than out to the internet), but I eventually decided to host on my server, as I could then do Tor and some other stuff more easily. I've got the server anyway...

One of these days, I may play with dropping analytics entirely and just passing requests through to my server, let images remain cached as that's the bulk of my bandwidth. Then I can go even more oldskool and parse my server logs for stats and referrers and such!


> Then I can go even more oldskool and parse my server logs for stats and referrers and such!

Expect to see a bunch of bots. I tried setting up server-side analytics for a WordPress-based website, but I had to get rid of it as the bot traffic made it essentially useless.


I can tell you it's more than a $5 dreamhost box can handle right now.


My $3/mo vultr box can handle HN loads easily when using a fast and well-designed (namely resource-leak-free) backend (I've settled on https://github.com/yesodweb/wai based apps - the only thing that has worked well for me so far).


I front paged a couple times back in the day. In the neighborhood of thousands of pageloads and hundreds of concurrent users. Totally trivial for static HTML, but most people get into trouble with hand-rolled or poorly tuned blog frameworks that make multiple database calls on every visitor.


Second place for a few hours and ~1k points resulted in around 50k unique visits.

If your website is a collection of static files and you're hosting them on S3+CloudFront or something similar (GitHub pages works too), then it'll work without any issues and cost pennies for the whole thing.


I've gotten on the front page more than a few times. In my experience, it usually peaks around 1.5k concurrents for a blog post. Peak was 50k total visits over a couple days, but has been much less too. Depends on the content and how interesting it is to the wider HN audience.


I once frontpaged with a funny article that readers also shared with others. Back when people still did that instead of taking a screenshot.

1,000,000 uniques over 3 days in ~2011.

I have been trying to re-create that high ever since, lol. Going viral is one hell of a drug.


Wow thats insane! I always think this is what going viral on social media is like, although I am probably better off not having experienced that.


I just hit #1 last week and frontpaged a bunch in the past. Peaked at around 250-300 concurrent visitors, totaling around 10k in a 24-hr period, which is on par with past experience.


10-100k iirc. Peak requests maybe 200-500req/sec (but not sustained). I had a few posts get >250k but those were on Reddit as well.


Hit the front page twice with two articles in the past.

Total traffic both times was around 60k over the course of 2-3 days.


Would nginx (caching everything) work on a $5/month VPS?


Should work fine. I personally avoided using a reverse proxy like nginx or apache because they tend to have a ton of vulnerabilities (check out the CVE database results for "nginx"), making them a security management headache.


Any serious vulnerability in NGINX will be big news since it is so widespread. CVE database shows some entries by searching for "nginx" but I looked at all 2022 entries and the only ones affecting NGINX itself are in NJX plugin so actually not affecting NGINX core functionality.

https://nginx.org/en/security_advisories.html shows one "medium severity" vulnerability in the last 4 years.


Huh, guess I haven't really checked on this since mid-early 2010s. Was a lot worse back then.


Suggestion for the next blog entry: How my blog post going viral on Hacker News pushed me to finally use a CDN


I wrote that one, but haven't gotten enough traffic again to see if it works... https://l-o-o-s-e-d.net/hug-of-death


How I generated static html for my static site and it got served and cached.


this. I couldn't even load the page.


Is DALL-E 2 generating the images on every page request? ;)


An important lesson: >99% of your blog readers will visit during <1% of the time.

If your goal is to allow people to read the stuff you write, you must be capable of serving your content to >1000 people per second.

Use a good backend (I recommend Servant+Warp or Snap - the only two I've tried that could handle it, out of probably 15 popular options tested) or a good caching reverse proxy (insofar as such a thing can be said to exist) and/or a CDN.


Static html, from a cloud bucket.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: