Hacker News new | past | comments | ask | show | jobs | submit login
Show HN: A beetle generator made by machine-learning zoological illustrations (cunicode.com)
308 points by belforn 3 months ago | hide | past | web | favorite | 47 comments



I had the thought of how obfuscating this technology could be to hide information. Imagine this going off and publishing an unlimited number of images with generated descriptions that are indistinguishable from what a human would do. There would be no way to verify what the correct information is if you are someone just casually searching the Internet. (This could also almost apply for any type of information...)


-> https://thisrentaldoesnotexist.com

All of the dynamic content on each listing is generated via a series of different machine-learned AI models.


"The laundry are converted into a mini central heating for every guest."


I've found that creating Markov chains are more human feeling (but maybe a bit overfit) if I use a n[0]-level deep look ahead for my generators.

[0] where n is normally 2 or 3


Yeah, the listings on this page don't quite make it up to a human level of coherence:

"Thus serve a 24 hour security pournising. Also hype is requested as much as you please empty and have a prestige less um be restriction, day or night."

It's close, though! But maybe not close enough for people to worry -- the last 10% of the security pournising always takes 90% of the development time.


“there will be a BREAKFAST Walk into the block in the room”


If that was going to be a problem, people would already be generating fake numerical data like material properties, physical constants, statistics, or whatever that's also subtly wrong. You don't need machine learning to do that.

This fear of fake information from ML misleading everyone is ridiculous and kind of arrogant. It assumes that the world is full of "other people" who are too stupid to make decisions for themselves and usually concludes that "us smart people" have to somehow control what they see to protect them from themselves. We've had fake information since forever and we've developed systems to deal with it. Citing sources, trustworthy organizations, multiple sources agreeing with each other, people pointing out mistakes, Google favoring popular sites, confirming it yourself, etc. Some fake information still gets through and it is a problem but it already happens and the world keeps turning. For casual internet searchers who don't care how reliable their information is, let them believe whatever nonsense satisfies them. They aren't trying to be right, they're just entertaining themselves.


I agree that "we must control others" is a bad approach. But I think you're underestimating the problem. Fake news has a body count, and it will only get worse: https://www.wired.com/story/how-whatsapp-fuels-fake-news-and...

At the very least, I think we need to train people in a lot more media literacy. But traditional approaches to that rely on media being scarce and expensive to create, which gave people enough time to carefully vet what they were consuming. As it becomes cheaper to create media than to vet it, we'll have the same problem as spam: it'll be impossible for humans to effectively filter it manually.

I think the real solution is automated vetting tools, so no information is presented without provenance. Basically, any time somebody sees an image or a video, there should be a link that lets you find out about the source, the editing, and who, specifically is vouching for it. And warnings for things that lack that. That still gives the viewer agency, but brings the problem back to human scale.


What a clever idea!

By the way, there are over 400,000 naturally occurring species of beetles. Beetles make up 25% of all animal forms according to Wikipedia.


There's been some new work that throws that percentage into doubt: https://www.insidescience.org/news/move-over-beetles-new-cha...


Reminds me of the (possibly apocryphal) Haldane quote, on what nature revealed about the mind of God: "An inordinate fondness for beetles."

https://en.wikipedia.org/wiki/J._B._S._Haldane


Humans: by mass, a very insignificant species.


Not among land mammals: https://xkcd.com/1338/


I wonder if this could be used as an identification tool for beetles like a phantom sketch. You could move some sliders to get it closer to the bug you are thinking of or which you have in front of you.

Of course, this method would have to compete with a related model just trying to classify a photo.


They already have similar apps to this for a range of things. I guess the most popular ones are for plants [1]

[1] https://www.picturethisai.com/


it could be used like that. Technically, the model already learned the features of the dataset. (although it was unlabelled). There are some implementations to find features in StyleGAN and enhance/modify those: StyleGan Encoder https://github.com/spiorf/stylegan-encoder


Reminds me of Dawkins' biomorphs: http://www.emergentmind.com/biomorphs


For more info on things like this:

https://en.wikipedia.org/wiki/L-system


I thought it was now basically the law that this must be named thisbeetledoesnotexist.com


Someone already set it up: http://thisbeetledoesnotexist.com/


haha, about time!


This is a beautiful write-up. And I don't wish to detract from the author's work. Nature's result side-by-side with that of the Machine. Makes one feel as though we've taken a step backwards from Alan Turing's "The Chemical Basis of Morphogenesis".

Consider the "morphogenetic puzzle" of a bi-valved seashell that shuts with perfect water-tight seal. There is a constraint to this design: survival!

https://twitter.com/AlainGoriely/status/1207210428344029184


Reminds me of video game art: every game has two sets of art, the “real” art which is part of the mechanics of the game world, and the “pretty bits” which dangle off the game objects and try to trick you into believing there’s more to the game world than there is.

A lot of gameplay involves testing for this boundary... Trying to figure out whether you can actually do things that are implied by the art.

Are there any modern games where 100% of the art exists inside the game world?


I'd argue that that's definitely the case for Dwarf Fortress.


Check out Noita -- it's a pixel-art-style game where every pixel is subject to the physics engine. Leads to some cool emergent gameplay.


> Are there any modern games where 100% of the art exists inside the game world?

Baba Is You


In Lone Echo, a VR game, you can grab any and all surfaces with your hands.


Deep Rock galactic. If it's not all game objects its damn close.


Minecraft?


There are e.g. particle effects which don't (afaiu) influence anything else in the game, and are just visual.

Also, the clouds.


Excellent. I did something similar a few months ago using a dataset of zoological silhouettes, resulting in a menagerie of mammals, bugs, spiders and other mutant wonders.

https://twitter.com/cortesi/status/1153075801891278848

https://twitter.com/cortesi/status/1153088629972934656


nice results! could be interesting to see another network naming the mutant creatures :)


Was this model tested for overfitting? I do not have any sense of whether the beetles that I'm seeing match some source pictures exactly.

I noticed that the transformations seem to be fast through a transition and then seemingly paused. Is this intentional or does this have something to do with the model?


Is there a good way to test a gan for overfitting?


To the creator: since most (all?) of the beetles are symmetrical, couldn’t you generate left halves and then reflect it to create the right half? This could help you prevent asymmetric generations


Most of the images have lighting which isn’t symmetrical and I think that makes it pretty obvious when an image is made by mirroring half a beetle.

But maybe there’s some way to deal with that.


Naively, you could train another network to correct the lighting on a mirrored beetle.

Although I’m sure there’s a smarter way.


The network should just learn that they all have symmetry and only encode the unique information. Once you start hand-coding priors like this, where do you stop? Maybe also constrain the range of colors? Size? Other geometric features? Eventually you're just doing old fashioned programming, not ML. And since generating beetle images isn't really the important goal anyway, why would you use dirty tricks to achieve it?


Would be interesting to see the result trained on images from this book[1] about mutation on insects in the chernobyl area.

[1]https://www.amazon.com/Heteroptera-Beautiful-Other-Images-Mu... [2]https://atomicphotographers.com/cornelia-hesse-honegger/


What I never understand about these things is ... what actually does the drawing? The AI decides what the beetle looks like, at what level of abstraction? When/how does it go from beetle idea to pixels? Does this network "know" what the beetle's "leg" is, or does it just "know" "this pixel here should be this color"?


From what I understand there are two networks in a GAN like this one.

One (the discriminator) is trained with a bunch of images showing what beetles can look like. It detects a real or fake image of a beetle.

The other (the generator) is just generating images with a convolutional neural network. The generator optimizes itself based on how close it is to passing the discriminators test - that is its "loss function".

So over time, the generator gets better and better at making things that look like beetles. The process takes a very long time and is aided by many GPUs (as mentioned in the article)


Much closer to the latter. I haven’t read this yet, but it sounds like an encoder model.


Correct.

The machine here doesn’t even know that those are beetles (because nobody told it), it is “just” arranging pixels in a similar manner as the pixels from the source images. It does understand that each generated image must have “legs”, “eyes”, “shells”... and other features that it detected are common in the original images.


Interesting to see how it handled the legs. Looks like it had problems with the fact insects have 3 pairs of symmetrical legs.


Is the source/model/network available to run?


yes, it is published. Available via @RunwayML https://open-app.runwayml.com/?model=cunicode/confusing_cole...


Very clever. Do you know if a machine does the drawing? Wasn't totally clear on that bit.




Guidelines | FAQ | Support | API | Security | Lists | Bookmarklet | Legal | Apply to YC | Contact

Search: