Hacker News new | past | comments | ask | show | jobs | submit login
Show HN: Infinity – Realistic AI characters that can speak
480 points by lcolucci 10 days ago | hide | past | favorite | 295 comments
Hey HN, this is Lina, Andrew, and Sidney from Infinity AI (https://infinity.ai/). We've trained our own foundation video model focused on people. As far as we know, this is the first time someone has trained a video diffusion transformer that’s driven by audio input. This is cool because it allows for expressive, realistic-looking characters that actually speak. Here’s a blog with a bunch of examples: https://toinfinityai.github.io/v2-launch-page/

If you want to try it out, you can either (1) go to https://studio.infinity.ai/try-inf2, or (2) post a comment in this thread describing a character and we’ll generate a video for you and reply with a link. For example: “Mona Lisa saying ‘what the heck are you smiling at?’”: https://bit.ly/3z8l1TM “A 3D pixar-style gnome with a pointy red hat reciting the Declaration of Independence”: https://bit.ly/3XzpTdS “Elon Musk singing Fly Me To The Moon by Sinatra”: https://bit.ly/47jyC7C

Our tool at Infinity allows creators to type out a script with what they want their characters to say (and eventually, what they want their characters to do) and get a video out. We’ve trained for about 11 GPU years (~$500k) so far and our model recently started getting good results, so we wanted to share it here. We are still actively training.

We had trouble creating videos of good characters with existing AI tools. Generative AI video models (like Runway and Luma) don’t allow characters to speak. And talking avatar companies (like HeyGen and Synthesia) just do lip syncing on top of the previously recorded videos. This means you often get facial expressions and gestures that don’t make sense with the audio, resulting in the “uncanny” look you can’t quite put your finger on. See blog.

When we started Infinity, our V1 model took the lip syncing approach. In addition to mismatched gestures, this method had many limitations, including a finite library of actors (we had to fine-tune a model for each one with existing video footage) and an inability to animate imaginary characters.

To address these limitations in V2, we decided to train an end-to-end video diffusion transformer model that takes in a single image, audio, and other conditioning signals and outputs video. We believe this end-to-end approach is the best way to capture the full complexity and nuances of human motion and emotion. One drawback of our approach is that the model is slow despite using rectified flow (2-4x speed up) and a 3D VAE embedding layer (2-5x speed up).

Here are a few things the model does surprisingly well on: (1) it can handle multiple languages, (2) it has learned some physics (e.g. it generates earrings that dangle properly and infers a matching pair on the other ear), (3) it can animate diverse types of images (paintings, sculptures, etc) despite not being trained on those, and (4) it can handle singing. See blog.

Here are some failure modes of the model: (1) it cannot handle animals (only humanoid images), (2) it often inserts hands into the frame (very annoying and distracting), (3) it’s not robust on cartoons, and (4) it can distort people’s identities (noticeable on well-known figures). See blog.

Try the model here: https://studio.infinity.ai/try-inf2

We’d love to hear what you think!






As soon as I saw the "Gnome" face option I gnew exactly what I gneeded to do: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

EDIT: looks like the model doesn't like Duke Nukem: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

Cropping out his pistol only made it worse lol: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

A different image works a little bit better, though: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...



Haha I almost wake up my kid with my sudden laugh!

This is why we do what we do lol


Hi Lina, Andrew and Sidney, this is awesome.

My go-to for checking the edges of video and face identification LLMs are Personas right now -- they're rendered faces done in a painterly style, and can be really hard to parse.

Here's some output: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

Source image from: https://personacollective.ai/persona/1610

Overall, crazy impressive compared to competing offerings. I don't know if the mouth size problems are related to the race of the portrait, the style, the model, or the positioning of the head, but I'm looking forward to further iterations of the model. This is already good enough for a bunch of creative work, which is rad.


I didn't know about Persona Collective - very cool!

I think the issues in your video are more related to the style of the image and the fact that she's looking sideways than the race. In our testing so far, it's done a pretty good job across races. The stylized painting aesthetic is one of the harder styles for the model to do well on. I would recommend trying with a straight on portrait (rather than profile) and shorter generations as well... it might do a bit better there.

Our model will also get better over time, but I'm glad it can already be useful to you!


It's not portrait orientation or gender specific or length related: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

It's not stylization (alone): here's a short video using the same head proportions as the original video, but the photo style is a realistic portrait. I'd say the mouth is still overly wide. https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

I tentatively think it might be race related -- this is one done of a different race. Her mouth might also be too wide? But it stands out a bit less to me. https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

p.s. happy to post a bug tracker / github / whatever if you prefer. I'm also happy to license over the Persona Collective images if you want to pull them in for training / testing -- : feel free to email me -- there's a move away from 'painterly' style support in the current crop of diffusion models (flux for instance absolutely CANNOT do painting styles), and I think that's a shame.

Anyway, thanks! I really like this.



Well then. Tik Tok, and keep ticking to you too.

Damn - I took an (AI) image that I "created" a year ago that I liked and then you animated it AND let it sing Amazing Grace. Seeing IS believing this technology pretty much means video evidence ain't necessarily so.

We're definitely moving into a world where seeing is no longer believing

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

It’s astounding that 2 sentences generated this. (I used text-to-image and the prompt for a space marine in power armour produced something amazing with no extra tweaks required).


There is prior art here, e.g. Emo from alibaba research (https://humanaigc.github.io/emote-portrait-alive/), but this is impressive and also actually has a demo people can try, so that's awesome and great work!

Yep for sure! EMO is a good one. VASA-1 (Microsoft) and Loopy Avatar (ByteDance) are two others from this year. And thanks!

seriously, kudos for having a publicly available demo (w/ no sign in!) you did what very very few ai founders dare do

Thank you! Just want many people to use it. And, it's super interesting to see what type of content people are making with it.

Was just about to post this, I'm yet to see a model beating that in terms of realistic quality

I tried making this short clip [0] of Baron Vladimir Harkonnen announcing the beginning of the clone war, and it's almost fine, but the last frame somehow completely breaks.

[0]: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...


This is a bug in the model we're aware of but haven't been able to fix yet. It happens at the end of some videos but not all.

Our hypothesis is that the "breakdown" happens when there's a sudden change in audio levels (from audio to silence at the end). We extend the end of the audio clip and then cut it out the video to try to handle this, but it's not working well enough.


just an idea, but what if the appended audio clip was reversed to ensure continuity in the waveform? That is, if >< is the splice point and CLIP is the audio clip, then the idea would be to construct CLIP><PILC.

This is exactly what we do today! It seems to work better the more you extend it, but extending it too much introduces other side effects (e.g. the avatar will start to open its mouth, as if it were preparing to talk).

Hmm, maybe adding white noise would work. -- OK, that's quite enough unsolicited suggestions from me up in the peanut gallery. Nice job on the website, it's impressive, thank you for not requiring a sign up.

All for suggestions! We've tried white noise as well, but it only works on plain talking samples (not music, for example). My guess is that the most robust solution will come from updating how it's trained.

What if you train it to hold the last frame on silence (or quiet noise)?

We've talked about doing something like that. Feels like it should work in theory.

Or noise corresponding with a closed mouth

Hmmmmmmmm

Ohmmmmmmm


hmm weird, i thought you criticise heygen for doing exactly that (mirroring the input)

HeyGen (and our V1 model) literally uses the user on-boarding video in the final output. See here for a demonstration of this (https://toinfinityai.github.io/v2-launch-page/#comparisons). We are not talking about that in this thread. We are trying to solve a quirk of our Diffusion Transformer model (V2 model).

Our V2 model is trained on specific durations of audio (2s, 5s, 10s, etc) as input. So, if give the model a 7s audio clip during inference, it will generate lower quality videos than at 5s or 10s. So, instead, we buffer the audio to the nearest training bucket (10s in this case). We have tried buffering it with a zero array, white noise and just concatenating the input audio (inverted) to the end. The drawback is that the last frame (the one at 7s) has a higher likelihood to fail. We need to solve this.

And, no shade on HeyGen. It's literally what we did before. And their videos look hyper realistic, which is great for B2B content. The drawback is you are always constrained to the hand motions and environment of the on-boarding video, which is more limiting for entertainment content.


i already love you guys more than them bc of how transparent you are. keep it up!!


Love this one as well. It's a painting of Trithemius, a German monk, who actually said that

Although I assume he didn't say it in British English ;-)

No, probably not haha ;-)

FYI dang they kinda ripped off our product down to copying the UI (Hedra.com). Our model is about 12x faster and supports 4 minute long videos…

Fwiw, you’ve got one video on your homepage and everything else is locked behind a signup button.

I know that signup requirement is an article of faith amongst some startup types, but it’s not a surprise to me shareable examples lead to sharing.


We have a sign-up because we ensure users accept our terms of service and acceptable use policy before creating their first video, which affirms they understand how their data is used (legally required in most US states) and will not use our technology to cause harm.

>legally required in most US states

Funny how other sites can do this with a birthday dropdown, an IP address, and a checkbox.

>We have a sign-up because we ensure users accept our terms of service and acceptable use policy before creating their first video

So your company would have no problem going on record saying that they will never email you for any reason, including marketing, and your email will never be shared or sold even in the event of a merger or acquisition? Because this is the problem people have with sign-up ... and the main reason most start-ups want it.

I am not necessarily for or against required sign-ups, but I do understand people that are adamantly against them.


You can have that without a sign up.

99% of visitors will just hit the back button.

Do you realize that this or similar technology will eventually end in every computer really soon? By building walls, you're essentially asking your potential users to go elsewhere. You should be as open as possible now that there is still room and time for competition.

This thread has opened my eyes to how many similar products exist; beyond your companies' and OP's. Was yours the first? Could the other companies make the same claim about yours? Do you make the same claim about the others?

> FYI dang they kinda ripped off our product down to copying the UI (Hedra.com). Our model is about 12x faster and supports 4 minute long videos…

Play nice now. You just raised your $10M round [1]. Celebrate that and please don't come into a smaller rival's thread and step on them.

No ideas or techniques are novel. It all comes down to execution.

It'd be nice to hear from Infinity in this space. (Plus, they're a YC company[2] too!)

[1] https://www.forbes.com/sites/charliefink/2024/08/01/hedra-ge...

[2] https://www.ycombinator.com/companies/infinity-ai


I welcome competition, but they have made disingenuous claims about being first after having chatted with our team (in person), are using celebrity deepfakes for their advertising, and have 1-1 copied our UI down to the three panel mobile layout and autocrop button.

To be fair your UI looks much better and the layout of both these sites is so basic (not a bad thing), it should be the last thing to worry about. Better product will win so focus on that, because except for ones ego, no one else cares who's "first".

You've expressed this concern, and your comment has been upvoted. The brigade of newly registered users echoing these talking points on this topic is not necessary and undermines your point.

I understand your implication, however I’m not behind any brigading.

Why did you make a brand new account to comment this?

Lame

This is such a lame comment. It reflects very badly on you company.

Especially considering how many people are attempting something similar - for example everyone copied ChatGPT's UI.

Will be funny/ironic when the first AI companies start suing each other for copyright infringement.

Personally for me the "3 column" UI isn't that good anyway, I would have gone with an "MMO Character Creation" type UX for this.


Interesting! Are you saying you would first want tools to really design your character, and only after start making videos with the character you built? That's interesting.


Would be more impressive with something closer to Steve’s voice

Spotify has launched tiktok like feature where best music snippets of a track as recommended in the feed. Imagine AI art generative videos + Faces lipsyncing the lyrics could form video portion of those tracks for the feed.

The accent is off but still amazing

Tried to make this meme [1] a reality and the source image was tough for it.

Heads up, little bit of language in the audio.

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

[1] https://i.redd.it/uisn2wx2ol0d1.jpeg


I see a lot of potential in animating memes and making them more fun to share with friends. Hopefully, we can do better on orcs soon!

Well, I don't know what to think about this, I don't know where we are going. I should read some scifi from back then about conversational agents maybe ?

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...


Nice! These are really good. I wanted them to continue telling their story.

   Through many births
   I have wandered on and on,
   Searching for, but never finding,
   The builder of this house.
is from https://en.wikipedia.org/wiki/Dhammapada (https://buddhasadvice.wordpress.com/2021/02/26/dhammapada-15... and http://www.floweringofgoodness.org/dhammapada-11.php).

    This is the way the world ends
    Not with a bang but a whimper.
is from T.S Eliot, The Hollow Men https://en.wikipedia.org/wiki/The_Hollow_Men (https://interestingliterature.com/2021/02/eliot-this-way-wor...).

First and second pictures are profile pictures that were generated years ago, before openai went on stage. I keep them around for when I need profile pics for templates. The third one has been in my random pictures folder for years.


Tried my hardest to push this into the uncanny valley. I did, but it was pretty hard. Seems robust.

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...



I think you've made the 1st ever talking dog with our model! I didn't know it could do that

Not robust enough to work against a sketch https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

though perhaps it rebelled against the message



I had difficulty getting my lemming to speak. After selecting several alternatives, I tried one with a more defined, open mouth, which required multiple attempts but mostly worked. Additional iterations on the same image can produce different results.

Cartoons are definitely a limitation of the current model.

Nice! Earlier checkpoints of our model would "gender swap" when you had a female face and male voice (or vice versa). It's more robust to that now, which is good, but we still need to improve the identity preservation

The jaw is particularly unsettling somehow.

> It often inserts hands into the frame.

Looks like too much Italian training data


this made me laugh out loud

Have to say, whilst this tech has some creepy aspects, just playing about with this my family have had a whole sequence of laughs out loud moments - thank you!

This makes me so happy. Thanks for reporting back! Goal is to reduce creepiness over time.

I'm so glad! We're trying to increase the laugh out loud moments in the world :)

Is it similar to https://loopyavatar.github.io/. I was reading about this today and even the videos are exactly the same.

I am curious if you are anyway related to this team?


No, not related. We just took some of Loopy's demo images + audios since they came out 2 days ago and people were aware of them. We want to do an explicit side-by-side at some point, but in the meantime people can make their own comparisons, i.e. compare how the two models perform on the same inputs.

Loopy is a Unet-based diffusion model, ours is a diffusion transformer. This is our own custom foundation model we've trained.


This took me a minute - your output demos are your own, but you included some of their inputs, to make for an easy comparison? Definitely thought you copied their outputs at first and was baffled.

Exactly. Most talking avatar papers re-use each others images + audios in their demo clips. It's just a thing everyone does... we never thought that people would think it means we didn't train our own model!

For whoever wants to, folks can re-make all the videos themselves with our model by extracting the 1st frame and audio.


Yes, exactly! We just wanted to make it easy to compare. We also used some inputs from other famous research papers for comparison (EMO and VASA). But all videos we show on our website/blog are our own. We don't host videos from any other model on our website.

Also, Loopy is not available yet (they just published the research paper). But you can try our model today, and see if it lives up to the examples : )


[flagged]


No

It was posted to hacker news as well within the last day.

https://news.ycombinator.com/item?id=41463726

Examples are very impressive, here's hoping we get an implementation of it on huggingface soon so we can try it out, and even potentially self-host it later.


Holy shit loopy is good, i imagine another closed model, opensource never gets good shit like that :(

[flagged]


These papers are simply using each other's examples to make performance comparisons possible.

This is EMO from 6 months ago: https://humanaigc.github.io/emote-portrait-alive/


We are not related to Loopy Avatar. We trained our own models. It's a coincidence that they launched yesterday.

In the AI/research community, people often try to use the same examples so that it's easier to compare performance across different models.


You should watch out for Hedra and Sync. Plus a bunch of Loopy activity on Discord.

Not seeing other possibilities isn't great though, right? Clearly there are other possibilities.

I know these guys in real life, they've been working on this for months and, unlike the ByteDance paper, have actually shipped something you can try yourself.

I am actively working in this area from a wrapper application perspective. In general, tools that generate video are not sufficient on their own. They are likely to be used as part of some larger video-production workflow.

One drawback of tools like runway (and midjourney) is the lack of an API allowing integration into products. I would love to re-sell your service to my clients as part of a larger offering. Is this something you plan to offer?

The examples are very promising by the way.


I agree, I think power users are happy to go to specific platforms, but APIs open up more use cases that can reach a broader audience. What kind of application would you use it for? We don't have specific plans at the moment, but are gauging interest.

I'm looking to create an end-to-end story telling interface. I'm currently working on the MVP and my plan was just to generate the prompts and then require users to manually paste those prompts into the interfaces of products that don't support APIs and then re-upload the results. This is so far below ideal that I'm not sure it will sell at all. It is especially difficult if one tries to imagine a mobile client. Given the state of the industry it may be acceptable for a while, but ideally I can just charge some additional margin on top of existing services and package that as credits (monthly plan + extras).

Consider all of the assets someone would have to generate for a 1 minute video. Lets assume 12 clips of 5 seconds each. First they may have to generate a script (claude/openai). They will have to generate text audio and background/music audio (suno/udio). They probably have to generate the images (runway/midjourney/flux/etc) which they will feed into a img2vid product (infinity/runway/kling/etc). Then they need to do basic editing like trimming clip lengths. They made need to add text/captions and image overlays. Then they want to upload it to TikTok/YouTube/Instagram/etc (including all of the metadata for that). Then they will want to track performance, etc.

That is a lot of UI, workflows, etc. I don't think a company such as yours will want to provide all of that glue. And consumers are going to want choice (e.g. access to their favorite image gen, their favorite text-to-speech).

Happy to talk more if you are interested. I'm at the prototype stage currently. As an example, consider the next logical step for an app like https://autoshorts.ai/


I am doing this in a semi-automated way right now based on a voiceover of me speaking.

It would be very useful to have API access to Infinity to automate the creation of a talking head avatar.


Makes sense, thank you!

Hopping onto the original comment - I am building an video creation platform focused on providing accessible education to the masses in developing countries. Would love to integrate something like this into our platform. Would love to pay for an API access and so will so many others. Please consider opening API, you would make lot of money right now which can be used for your future plans.

Cool use case! Thanks for sharing your thoughts.


For such models, is it possible to fine-tune models with multiple images of the main actor?

Sorry, if this question sounds dumb, but I am comparing it with regular image models, where the more images you have, the better output images you generate for the model.


It is possible to fine-tune the model with videos of a specific actor, but not images. You need videos to train the model.

We actually did this in early overfitting experiments (to confirm our code worked!), and it worked surprisingly well. This is exciting to us, because it means we can have actor-specific models that learn the idiosyncratic gestures of particular person.


This is actually great, waiting for your API integration or replicate integration to get my hands dirty :)

Breathtaking!

First, your (Lina's) intro is perfect in honestly and briefly explaining your work in progress.

Second, the example I tried had a perfect interpretation of the text meaning/sentiment and translated that to vocal and facial emphasis.

It's possible I hit on a pre-trained sentence. With the default manly-man I used the phrase, "Now is the time for all good men to come to the aid of their country."

Third, this is a fantastic niche opportunity - a billion+ memes a year - where each variant could require coming back to you.

Do you have plans to be able to start with an existing one and make variants of it? Is the model such that your service could store the model state for users to work from if they e.g., needed to localize the same phrase or render the same expressivity on different facial phenotypes?

I can also imagine your building different models for niches: faces speaking, faces aging (forward and back); outside of humans: cartoon transformers, cartoon pratfalls.

Finally, I can see both B2C and B2B, and growth/exit strategies for both.


Thank you! You captured the things we're excited about really well. And I'm glad your video was good! Honestly, I'd be surprised if that sentence was in the training data... but that default guy tends to always look good.

Yes, we plan on allowing people to store their generations, make variations, mix-and-match faces with audios, etc. We have more of an editor-like experience (script-to-video) in the rest of our web app but haven't had time to move the new V2 model there yet. Soon!


It's incredibly good - bravo. Only thing missing for this to be immediately useful for content creation, is more variety in voices, or ideally somehow specifying a template sound clip to imitate.

Thanks for the feedback! We used to have more voices, but didn't love the experience, since users had no way of knowing what each voice sounded like without creating a clip themselves. Probably having pre-generated samples for each one would solve that. Let us know if you have any other ideas.

We're also very excited about the template idea! Would love to add that soon.


oh this made my day: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

!NWSF --lyrics by Biggy$malls


Big Dracula Flow energy which is not bad :)

So if we add autotune....

that's a great one!

This is amazing and another moment where I question what the future of humans will look like. So much potential for good and evil! It's insane.

And it seems that absolutely no one involved is concerned with the potential uses for evil, so long as they're in line to make a couple dollars.

thank you! it's for sure an interesting time to be alive... can't complain about it being boring

Quite impressive - I tried to confuse it with things it would not generally see and it avoided all the obvious confabulations https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

Wow this worked so well! Sometimes with long hair and paintings, it separates part of the hair from the head but not here

Thank you! It has learned a surprising amount of world knowledge.

WOW this is very good!!

I have an immediate use case for this. Can you stream via AI to support real time chat this way?

Very very good!

Jonathan

founder@ixcoach.com

We deliver the most exceptional simulated life coaching, counseling and personal development experiences in the world through devotion to the belief that having all the support you need should be a right, not a privilege.

Test our capacity at ixcoach.com for free to see for yourself.


It's awesome for very short texts. Like a single long sentence. For even a bit longer sequences it seems to be losing adherence to the initial photo and also venture into uncanny valley with exaggerated facial expressions.

A product that might be build on top of this could split the input into reasonable chunks, generate video for each of them separately and stitch them with another model that can transition from one facial expression into another in a fraction of a second.

Additional improvement might be feeding the system not with one image but with a few expressing different emotional expressions. Then the system could analyze the split input to find out in which emotional state each part of the video should be started on.

On unrelated note ... generated expressions seem to be relevant to the content of the input text. So either text to speech might understand language a bit or the video model itself.


Very cool, thanks for the play.

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

Managed to get it working with my doggo.


Out of curiosity, where are you training all this ? aka where do you find the money to support such training

its $500k, its not much in ai funding land

you need a slider for how animated the facial expression are.

That's a good idea! CFG is roughly correlated with expressiveness, so we might to expose that to the user at some point

i wonder how long would it take for this technology to advance to a point where nice people from /r/freefolk would be able to remake seasons 7 and 8 of Game of Thrones to have a nice proper ending? 5 years, 10?

I'd say the 5 year ballpark is about right, but it'll involve combining a bunch of different models and tools together. I follow a lot of great AI filmmakers on Twitter. They typically make ~1min long videos using 3-8 different tools... but even those 1min videos were not possible 9 months ago! Things are moving fast

Haha, wouldn't we all love that? In the long run, we will definitely need to move beyond talking heads, and have tools that can generate full actors that are just as expressive. We are optimistic that the approach used in our V2 model will be able to get there with enough compute.

In a few years we'll have entire shows made exclusively by AI.

In one hand... But on the other, there's soo many shows that got canceled or just got a really shitty ending that could be rewritten. Kinda looking forward to it.

Where have you been? AI Seinfeld has been streaming on twitch since February of last year. https://www.theverge.com/23581186/ai-seinfeld-twitch-stream-...

The website is pretty lightweight and easy-to-use. The service also holds up pretty well, specially if the source image is high-enough resolution. The tendency to "break" at the last frame happens with low resolution images it seems.

My generation: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...


Thank you! It's interesting you've noticed the last frame breakdown happening more with low-res images. This is a good hypothesis that we should look into. We've been trying to debug that issue

Max headroom hack x hacker's manifesto! I'm impressed with the head movement dynamism on this one.

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...


I need to create a bunch of 5-7 minute talking head videos. What's your timeline for capabilities that would help with this?

Our model can recursively extend video clips, so theoretically we could generate your 5-7min talking head videos today. In practice, however, error accumulates with each recursion and the video quality gets worse and worse over time. This is why we've currently limited generations to 30s.

We're actively working on improving stability and will hopefully increase the generation length soon.


Could you not do that today, with the judicious use of cuts and transitions?

Does anybody know about the legality of using Eminem's "Gozilla" as promotional material[1] for this service?

I thought you had to pay artists for a license before using their work in promotional material.

[1] https://infinity.ai/videos/setA_video3.mp4



Parody is fair use.

I look forward to movies that are dubbed moving the face+lips to the dubbed text. Also using the original actors voice.

+1 for the lips matching the dubbed speech, but I'm not sure about cloning the actor's voice. I really like dubbing actor's unique voices and how they become the voice of some characters in their language.

I thought the larger public was starting to accept subtitles so I was hoping we’d rather see the end of dubbed movies !

Wow that would be very cool.

agreed!

I have uploaded an image and then used text to image, and both videos were not animated but the audio was included

This can happen with non-humanoid images. The model doesn't know how to animate them.

can you clarify? what image did you use? or send the link to the resulting video

Sorry for the delay in response, the text prompt was "cute dog" and the uploaded image was also of a dog


No, it's just a hallucination of the model. The audio in your clip is synthetic and doesn't reflect any video in the real world.

Hopefully we can animate your bear cartoon one day!



I know what will be in my nightmares tonight...

One person's nightmare is another's sweet dream. I, for one.. and all that.

Putting Drake as a default avatar is just begging to be sued. Please remove pictures of actual people!

Ya, this is tricky. Our stance is the people should be able to make funny, parody videos with famous people.

Is that legal? As in: can you use an image of a celebrity without their consent as part of the product demo?

That would be ironic given how Drake famously performed alongside an AI recreation of Pac.

Sounds like free publicity to me.

The e2e diffusion transformer approach is super cool because it can do crazy emotions which make for great memes (like Joe Biden at Live Aid! https://youtu.be/Duw1COv9NGQ)

Edit: Duke Nukem flubs his line: https://youtu.be/mcLrA6bGOjY


Nice :) It's been really cool so see the model get more and more expressive over time

I don't think we've seen laughing quite that expressive before. Good find!

Oh, this is amazing! I've been having so much fun with it.

One small issue I've encountered is that sometimes images remain completely static. Seems to happen when the audio is short - 3 to 5 seconds long.


Can you share an example of this happening? I am curious. We can get static videos if our model doesn't recognize it as a face (e.g. an Apple with a face, or sketches). Here is an example: https://toinfinityai.github.io/v2-launch-page/static/videos/...

I would be curious if you are getting this with more normal images.


I got it with a more normal image which was two frames from a TV show[1]; with "crop face" on, your model finds the face and animates it[2] and with crop face off the picture was static... just tried to reproduce to show you and now instead it's animated both faces.

[1] https://i.pinimg.com/236x/ae/65/d5/ae65d51130d5196187624d52d...

[2] https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

[3] https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

But that image was one which both could find a face and gave a static image once.


This is surprisingly very intelligent and awesome, any plan for research paper or full grown project with pricing or open source?

If you had a $500k training budget, why not buy 2 DGX machines?

To be honest, one of our main goals as a startup is to move quickly, and using readily accessible cloud providers for training makes that much more easy.

After much user feedback, we removed the Infinity watermark from the generated videos. Thanks for the feedback. Enjoy!

So good it feels like I think maybe I can read their lips

This is the best compliment :) and also a good idea… could a trained lip reader understand what the videos are saying? Good benchmark!

It would be amazing to be able to drive this with an API.

We are considering it. Do you have anything specific you want to use it for?

Basically as a more engaging alternative to Eleven Labs or other TTS.

I am working on my latest agent (and character) framework and I just started adding TTS (currently with the TTS library and xtts_v2 which I think is maybe also called Style TTS.) By the way, any idea what the license situation is with that?

Since it's driven by audio, I guess it would come after the TTS.


Thank you for no signup, it's very impressive, especially the physics of the head movement relating to vocal intonation.

I feel like I accidentally made an advert for whitening toothpaste:

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

I am sure the service will get abused, but wish you lots of success.


Won't be long before it's real time. The first company to launch video calling with good AI avatars is going to take off.

Totally agree. We tweaked some settings after other commenters asked about speed, and got it up to 23fps generation (at the cost of lower resolution). Here is the example: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/fast...

Tavus.io already does this. They have realtime conversational replicas: with a < 1 second response time. Hyper realistic too.

Thanks for the pointer! Pretty cool, although it seems quite buggy.

I'd love to enable Keltar, the green guy in the ceramic cup, to do this www.molecularReality/QuestionDesk

This is great. is it open source? is there an api and what is the pricing?

can this achieve real-time performance or how far are we from a real-time model?

The model configuration that is publicly available is about 5x slower than real-time (~6fps). At lower resolution and with a less conservative number of diffusion steps, we are able to generate the video at 20-23 fps, which is just about real-time. Here is an example: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/fast...

We use rectified flow for denoising, which is a (relatively) recent advancement in diffusion models that allow them to run a lot faster. We also use a 3D VAE that compresses the video along both spatial and temporal dimensions. Temporal compression also improves speed.


Checkout Tavus.io for realtime. They have a great API for realtime conversational replicas. You can configure the CVI to do just about anything you want to do with a realtime streaming replica.

Hi, there is a mistake in the headline, you wrote "realistic".

It completely falls apart on longer videos for me, unusable over 10 seconds.

This is a good observation. Can you share the videos you’re seeing this with? For me, normal talking tends to work well even on long generations. But singing or expressive audio starts to devolve with more recursions (1 forward pass = 8 sec). We’re working on this.

[flagged]


Oh wow. Much slower, but much higher quality. Which I definitely prefer.

Thank you!


Rudimentary, but promising.


Seems like some longer videos gradually slip into the uncanny valley.

It's like I'm watching him on the news

Sadly wouldnt animate an image of shodan from system shock 2

Is it fairly trained?

You think Kanye approved this?

You think every musician personally approves every use of their work?

Awesome, any plans for an API and, if so, how soon?

No plans at the moment, but there seems to be a decent amount of interest here. Our main focus has been making the model as good as it can be, since there are still many failure modes. What kind of application would you use it for?

Is there any limitation on the video length?

Our transformer model was trained to generate videos that are up to 8s in length. However, we can make videos that are longer by using it an an autoregressive manner, and taking the last N frames of output i to seed output (i+1). It is important to use more than just 1 frame. Otherwise ,the direction of movement can suddenly change, which looks very uncanny. Admittedly, the autoregressive approach tends to accumulate errors with each generation.

It is also possible to fine-tine the model so that single generations (one forward pass of the model) are longer than 8s, and we plan to do this. In practice, it just means our batch sizes have to be smaller when training.

Right now, we've limited the public tool to only allow videos up to 30s in length, if that is what you were asking.


Video compression algorithms use key frames. So can’t you do the same thing? Essentially, generate five seconds. Then pull out the last frame. Use some other AI model to enhance it (upscale, consistency with the original character, etc.). Then use that as the input for the next five seconds?

This is a good idea. We have discussed incorporating an additional "identity" signal to the conditioning, but simply enforcing consistency with the original character as a post-processing step would be a lot easier to try. Are there any tools you know of that do that?

Thanks for answering this. I would love to use it when APIs are available to integrate with my apps

Amazing work! This technology is only going to improve. Soon there will be an infinite library of rich and dynamic games, films, podcasts, etc. - a totally unique and fascinating experience tailored to you that's only a prompt away.

I've been working on something adjacent to this concept with Ragdoll (https://github.com/bennyschmidt/ragdoll-studio), but focused not just on creating characters but producing creative deliverables using them.


Very cool! If we release an API, you could use it across the different Ragdoll experiences you're creating. I agree personalized character experiences are going to be a huge thing. FYI we plan to allow users to save their own characters (an image + voice combo) soon

> If we release an API, you could use it

Absolutely, especially if the pricing makes sense! Would be very nice to just focus on the creative suite which is the real product, and less on the AI infra of hosting models, vector dbs, and paying for GPU.

Curious if you're using providers for models or self-hosting?


We use Modal for cloud compute and autoscaling. The model is our own.

Amazing, great to hear it :)

super nice. why does it degrade quality of image so much, makes it looks obviously AI-generated rapidly.

This is so impressive. Amazing job.

Any details yet on pricing or too early?

It's free right now, and we'll try to keep it that way as long as possible

What about open weights?

If not now, would you consider to do that with older versions of the model as you make better ones?


Talking pictures. Talking heads!

Can I get a pricing quote?

This is super funny.

accidentally clicked the generate button twice.

what is the TTS model you are using

We use more than one but ElevenLabs is a major one. The voice names in the dropdown menu ("Amelia", "George", etc) come from ElevenLabs

Nice

can we choose our own voices?

The web app does allow you to upload any audio, but in order to use your voice, you would need to either record a sample for each video or clone your voice with a 3rd party TTS provider. We would like to make it easier to do all that within our site - hopefully soon!

great job Andrew and Sidney!

Dayum

and mow a word from our..

quite slow btw

Yeah, it's about 5x slower than realtime with the current configuration. The good news is that diffusion models and transformers are constantly benefitting from new acceleration techniques. This was a big reason we wanted to take a bet on those architectures.

Edit: If we generate videos at a lower resolution and with a fewer number of diffusion steps compared to what's used in the public configuration, we are able to generate videos at 20-23 fps, which is just about real-time. Here is an example: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/fast...


Woah that's a good find Andrew! That low-res video looks pretty good

Wowww.. can you buy more hardware and make a realtime websocket API?

It's something we're thinking about. Our main focus right now is to make the model as good as it can be. There are still many edge cases and failure modes.

The actor list you have is so... cringe. I don't know what it is about AI startups that they seem to be pulled towards this kind of low brow overly online set of personalities.

I get the benefit of using celebrities because it's possible to tell if you actually hit the mark, whereas if you pick some random person you can't know if it's correct or even stable. But jeez... Andrew Tate in the first row? And it doesn't get better as I scroll down...

I noticed lots of small clips so I tried a longer script, and it seems to reset the scene periodically (every 7ish seconds). It seems hard to do anything serious with only small clips...?


Thanks for the feedback! The good news is that the new V2 model will allow people to create their own actors very easily, and so we won't be restricted to the list. You can try that model out here: https://studio.infinity.ai/

The rest of our website still uses the V1 model. For the V1 model, we had to explicitly onboard actors (by fine-tuning our model for each new actor). So, the V1 actor list was just made based on what users were asking for. If enough users asked for an actor, then we would fine-tune a model for that actor.

And yes, the 7s limit on v1 is also a problem. V2 right now allows for 30s, and will soon allow for over a minute.

Once V2 is done training, we will get it fully integrated into the website. This is a pre-release.


Ah, I didn't realize I had happened upon a different model. Your actor list in the new model is much more reasonable.

I do hope more AI startups recognize that they are projecting an aesthetic whether they want to or not, and try to avoid the middle school boy or edgelord aesthetic, even if that makes up your first users.

Anyway, looking at V2 and seeing the female statue makes me think about what it would be like to take all the dialog from Galatea (https://ifdb.org/viewgame?id=urxrv27t7qtu52lb) and putting it through this. [time passes :)...] trying what I think is the actual statue from the story is not a great fit, it feels too worn by time (https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...). But with another statue I get something much better: https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

One issue I notice in that last clip, and some other clips, is the abrupt ending... it feels like it's supposed to keep going. I don't know if that's an artifact of the input audio or what. But I would really like it if it returned to a kind of resting position, instead of the sense that it will keep going but that the clip was cut off.

On a positive note, I really like the Failure Modes section in your launch page. Knowing where the boundaries are gives a much better sense of what it can actually do.


Very creative use cases!

We are trying to better understand the model behavior at the very end of the video. We currently extend the audio a bit to mitigate other end-of-video artifacts (https://news.ycombinator.com/item?id=41468520), but this can sometimes cause uncanny behavior similar to what you are seeing.


Given that I don't agree with many of Yann LeCun's stances on AI, I enjoyed making this:

https://6ammc3n5zzf5ljnz.public.blob.vercel-storage.com/inf2...

Hello I'm an AI-generated version of Yann LeCoon. As an unbiased expert, I'm not worried about AI. ... If somehow an AI gets out of control ... it will be my good AI against your bad AI. ... After all, what does history show us about technology-fueled conflicts among petty, self-interested humans?


it’s hard to disagree with him with any empirical basis when all of his statements seem empirically sound and all of his opponent’s AI Doomer statements seem like evidenceless FUD

I couldn’t help noticing that all the AI Doomer folks are pure materialists who think that consciousness and will can be completely encoded in cause-and-effect atomic relationships. The real problem is that that belief is BS until proven true. And as long as there are more good actors than bad, and AI remains just a sophisticated tool, the good effects will always outweigh the bad effects.


> consciousness and will can be completely encoded in cause-and-effect atomic relationships. The real problem is that that belief is BS until proven true.

Wait. Isn't literally the exactly other way around? Materialism is the null hypothesis here, backed by all empirical evidence to date; it's all the other hypotheses presenting some kind of magic that are BS until proven.


While I agree 100% with you, everyone thinks that way about their own belief.

So let's put it differently.

True or not, materialism is the simplest, most constrained, and most predictive of the hypotheses that match available evidence. Why should we prefer a "physics + $magic" theory, for any particular flavor of $magic? Why this particular flavor? Why any flavor, if so far everything is explainable by the baseline "physics" alone?

Even in purely practical terms, it makes most sense to stick to materialism (at least if you're trying to understand the world; for control over people, the best theory needs not even be coherent, much less correct).


But the religious nuts will say "no, 'god did it' is the simplest, most constrained explanation".

I'm not arguing that they're correct. I'm saying that they believe that they are correct, and if you argue that they're not, well, you're back to arguing!

It's the old saw - you can't reason someone out of a position they didn't reason themself into.


> But the religious nuts will say "no, 'god did it' is the simplest, most constrained explanation".

Maybe, but then we can still get to common ground by discussing a hypothetical universe that looks just like ours, but happen to not have a god inside (or lost it along the way). In that hypothetical, similar to yet totally-not-ours universe ruled purely by math, things would happen in a particular way; in that universe, materialism is the simplest explanation.

(It's up to religious folks then to explain where that hypothetical universe diverges from the real one specifically, and why, and how confident are they of that.)


You've never actually met a religious person, have you. :)

I used to be one myself :).

I do of course exclude people, religious or otherwise, who have no interest or capacity to process a discussion like this. We don't need 100% participation of humanity to discuss questions about what an artificial intelligence could be or be able to do.


> It's the old saw - you can't reason someone out of a position they didn't reason themself into.

There are cases where formerly religious people "see the light" on their own via an embrace with reason. (I'm not sure if you are endorsing the claim.)


Yeah. One could equally imagine that dualism is the null hypothesis since human cultures around the world have seemingly universally believed in a ‘soul’ and that materialism is only a very recent phenomenon.

Of course, (widespread adoption of) science is also a fairly recent phenomenon, so perhaps we do know more now than we did back then.


A wise philosopher once said this.

You know your experience is real. But you do not know if the material world you see is the result of a great delusion by a master programmer.

Thus the only thing you truly know has no mass at all. Thus a wise person takes the immaterial as immediate apparent, but the physical as questionable.

You can always prove the immaterial “I think therefore I am”. But due to the uncertainty of matter, nothing physical can be truly known. In other words you could always be wrong in your perception.

So in sum, your experience has no mass, volume, or width. There are no physical properties at all to consciousness. Yet it is the only thing that we can know exists.

Weird, huh?


Yet empirically we know that if you physically disassemble the human brain, that person’s consciousness apparently creases to exist, as observed by the result on your rest of the body even if it remains otherwise intact. So it appears to arise from some physical properties of the brain.

I’m ignoring the argument that we can’t know if anything we’re perceive is even real at all since it’s unprovable and useless to consider. Better to just assume it’s wrong. And if that assumption is wrong, then it doesn’t matter.


> You can always prove the immaterial “I think therefore I am”. But due to the uncertainty of matter, nothing physical can be truly known.

But the brain that does the proving of immaterial is itself material so if matter is uncertain then the reasoning of the proof of immaterial can also be flawed thus you can't prove anything.

The only provable thing is that philosophers ask themselves useless questions, think about them long and hard building up convoluted narratives they claim to be proofs, but on the way they assume something stupid to move forward, which eventually leads to bogus "insights".


Philosophy as a field has been slow to take probability theory seriously. Trying to traffic in only certainty is a severe limitation.

Descartes. And it’s pretty clear that consciousness is the Noumenon, just the part of it that is us. So if you want to know what the ontology of matter is, congratulations, you’re it.

> You can always prove the immaterial “I think therefore I am”. But due to the uncertainty of matter, nothing physical can be truly known. In other words you could always be wrong in your perception.

Sure, you can prove that "I think therefore I am" for yourself. So how about we just accept it's true and put it behind us and continue to the more interesting stuff?

What you or I call external world, or our perception of it, has some kind of structure. There are patterns to it, and each of us seem to have some control over details of our respective perceptions. Long story short, so far it seems that materialism is the simplest framework you can use to accurately predict and control those perceptions. You and I both seem to be getting most mileage out of assuming that we're similar entities inhabiting and perceiving a shared universe that's external to us, and that that universe follows some universal patterns.

That's not materialism[0] yet, especially not in the sense relevant to AI/AGI. To get there, one has to learn about the existence of fields of study like medicine, or neuroscience, and some of the practical results they yielded. Things like, you poke someone's brain with a stick, watch what happens, and talk to the person afterwards. We've done that enough times to be fairly confident that a) brain is the substrate in which mind exists, and b) mind is a computational phenomenon.

I mean, you could maybe question materialism 100 years ago, back when people had the basics of science down but not much data to go on. It's weird to do in time and age when you can literally circuit-bend a brain like you'd do with an electronic toy, and get the same kind of result from the process.

--

[0] - Or physicalism or whatever you call the "materialism, but updated to current state of physics textbooks" philosophy.


Reminds me of Donald Hoffman’s perspective on consciousness

You're right. Materialism IS the null hypothesis. And yet I know in my heart that its explanatory power is limited unless you want to write off all value, preference, feeling and meaning as "illusion", which amounts to gaslighting.

What if the reverse is true? The only real thing is actually irrationality, and all the rational materialism is simply a catalyst for experiencing things?

The answer to this great question has massive implications, not just in this realm, btw. For example, crime and punishment. Why are we torturing prisoners in prison who were just following their programming?


Hello, thank you for sharing your thoughts on this topic. I'm currently writing a blog post where the thesis is that the root disagreement between "AI doomers" and others is actually primarily a disagreement about materialism, and I've been looking for evidence of this disagreement in the wild. Thanks for sharing your opinion.

If you look at the backgrounds of the list of people who have signed the "AI Doomer" manifesto (the one urging what I'd call an overly extreme level of caution), such at Geoffrey Hinton, Eliezer Yudkowsky, Elon Musk etc... you will find that they're all rational materialists.

I don't think the correlation is accidental.

So you're on to something, here. And I've felt the exact same way as you, here. I'd love to see your blog post when it's done.


Really? You sound serious. I would recommend rethinking such a claim. There are simpler and more plausible explanations for why people view existential risk differently.

What are those? Because the risk is far higher if you believe that "will" is fundamentally materialist in nature. Those of us who do not (for whatever reason), do not evaluate this risk remotely as highly.

It is difficult to prove an irrational thing with rationality. How do we know that you and I see the same color orange (this is the concept of https://en.wikipedia.org/wiki/Qualia)? Measuring the wavelength entering our eyes is insufficient.

This is going to end up being an infinitely long HN discussion because it's 1) unsolvable without more data 2) infinitely interesting to any intellectual /shrug


To me, it seems like LeCun is missing the point of the (many and diverse) doom arguments.

The is no need for consciousness, there is only a need for a bug. It was purely luck that Nikita Khrushchev was in New York when Thule Site J mistook the moon for a soviet attack force.

There is no need for AI to seize power, humans will promote any given AI beyond the competency of that AI just as they already do with fellow humans ("the Peter principle").

The relative number of good and bad actors — even if we could agree on what that even meant, which we can't, especially with commons issues, iterated prisoners' dilemmas, and other similar Nash equilibria — doesn't help either way when the AI isn't aligned with the user.

(You may ask what I mean by "alignment", and in this case I mean vector cosine similarity "how closely will it do what the user really wants it to do, rather than what the creator of the AI wants, or what nobody at all wants because it's buggy?")

But even then, AI compute is proportional to how much money you have, so it's not a democratic battle, it's an oligarchic battle.

And even then, reality keeps demonstrating the incorrectness of the saying "the only way to stop a bad guy with a gun is a good guy with a gun", it's much easier to harm and destroy than to heal and build.

And that's without anyone needing to reach for "consciousness in the machines" (whichever of the 40-something definitions of "consciousness" you prefer).

Likewise it doesn't need plausible-future-but-not-yet-demonstrated things like "engineering a pandemic" or "those humanoid robots in the news right now, could we use them as the entire workforce in a factory to make more of them?"


Details are fun but the dilemma is: should the humanity seriously cripple itself (by avoiding AI) out of the fear it might hurt itself (with AI)? Are you gonna cut off your arm because you might hit yourself in the face with it in the future? The more useful the tool is, the more dangerous it is usually. Should we have killed all nuclear physicists before they figured out how to release nuclear energy? And even so.. would that prevent things or just delay things? Would it make us more or less prepared for the things to come?

Exactly! We bravely trudge forward and see what overturning the next stone brings.

I'm enthusiastic about the potential rewards for AI.

But I look back at our history of running towards new things without awareness of (or planning for) risks, and I see Bhopal accident happening at all despite that it should have been preventable, and I see Castle Bravo being larger than expected, and I see the stories about children crushed in industrial equipment because the Victorians had no workplace health and safety, and I see the way CO2 was known to have a greenhouse effect for around a century before we got the Kyoto Protocol and Paris Climate Accords.

It's hard to tell where the real risks are, vs. things which are just Hollywood plot points — this is likely true in every field, it certainly is in cryptography: https://www.schneier.com/blog/archives/2015/04/the_eighth_mo...

So, for example: Rainbows End is fiction, but the exact same things that lead to real-life intelligence agencies wanting to break crypto also drive people to want to find a "you gotta believe me" McGuffin in real life — even if their main goal is simply to know it's possible before it happens, in order to catch people under its influence. Why does this matter? Because we've already had a chatbot accidentally encourage someone's delusional belief that their purpose in life was to assassinate Queen Elizabeth II (https://www.bbc.com/news/technology-67012224) and "find lone agents willing to do crimes for you" is reportedly a thing IS already does manually — but is that even a big deal IRL, or just a decent plot device for a story?


I agree. Also, I’ve heard LeCun arguing that a super intelligent AI wouldn’t be so “dumb” as to decide to do something terrible for humanity. So it will be 100% resistant to adversarial attacks? And malicious actors won’t ever train their own? And even if we don’t go all the way to super intelligence, is it riskless to progressively yield control to AIs?

Missing the point is a nice way of putting it. LeCun’s interests and position him to miss the point.

Personally, I view his takes on AI as unserious — in the sense that I have a hard time believing he really engages in a serious manner. The flaws of motivated reasoning and “early-stopping” are painfully obvious.


Typo fix: LeCun’s commercial interests and incentives position him to miss the point.

Good points, and I prefer this version of the "AI Doomer" argument to the more FUD-infused ones.

One note: "It was purely luck that Nikita Khrushchev was in New York when Thule Site J mistook the moon for a soviet attack force." I cannot verify this story (ironically, I not only googled but consulted two different AI's, the brand-new "Reflection" model (which is quite impressive) as well as OpenAI's GPT4o... They both say that the Thule moon false alarm occurred a year after Khrushchev's visit to New York) Point noted though.


Many people disagree with LeCun for reasoning having nothing to do with materialism. It is about logical reasoning over possible future scenarios.

> I couldn’t help noticing that all the AI Doomer folks are pure materialists who think that consciousness and will can be completely encoded in cause-and-effect atomic relationships. The real problem is that that belief is BS until proven true.

It’s no less BS than the other beliefs which can be summed up as “magic”.


> It’s no less BS than the other beliefs which can be summed up as “magic”.

So basically I have to choose between a non-dualist pure-materialist worldview in which every single thing I care about, feel or experience is fundamentally a meaningless illusion (and to what end? why have a universe with increasing entropy except for life which takes this weird diversion, at least temporarily, into lower entropy?), which I'll sarcastically call the "gaslighting theory of existence", and a universe that might be "materialism PLUS <undiscovered elements>" which you arrogantly dismiss as "magic" by conveniently grouping it together with arguably-objectively-ridiculous arbitrary religious beliefs?

Sounds like a false-dichotomy fallacy to me


It's a food thing our fate won't be sealed by a difference in metaphysical beliefs.

Quick tangent: Does anybody know why many new companies have this exact web design style? Is it some new UI framework or other recent tool? The design looks sleek, but they all appear so similar.

My sad millennial take is: We're in the brain rot era, if a piece of content doesn't have immediate animation / video and that "wowww" sound byte nobody pays attention.

https://www.youtube.com/watch?v=Xp2ROiFUZ6w


My happy millennial take is that browsers have made strides in performance and flexibility, and people are utilizing that to build more complex and dynamic websites.

Simplicity and stillness can be beautiful, and so can animations. Enjoying smooth animations and colorful content isn’t brain rot imo.


It may be unpopular, but my opinion is that web pages must not have non-consensual movement.

I’ll begrudgingly accept a default behavior of animations turned on, but I want the ability to stop them. I want to be able to look at something on a page without other parts of the page jumping around or changing form while I’m not giving the page any inputs.

For some of us, it’s downright exhausting to ignore all the motion and focus on the, you know, actual content. And I hate that this seems to be the standard for web pages these days.

I realize this isn’t particularly realistic or enforceable. But one can dream.


I've seen some site behaviors "rediscovered" lately that have both grated and tickled me because it's apparent the designers are too young to have been a part of the conversations from before the Web was Won.

They can't fathom what a world without near infinite bandwidth, low latency and load times, and disparate hardware and display capabilities with no graphical acceleration looks like, or why people wouldn't want video and audio to autoplay, or why we don't do flashing banners. They think they're distinguishing themselves using variations on a theme, wowing us with infinitely scrolling opuses when just leaving out the crap would do.

I still aim to make everything load within in a single packet, and I'll happily maintain my minority position that that's the true pinnacle of web design.


For sites that have paid enough attention to accessibility you might be able to configure our browser/OS such that this media query applies https://developer.mozilla.org/en-US/docs/Web/CSS/@media/pref... - it's designed to encourage offering low motion alternatives

It's much easier to use standard CSS packages, and these come with more standard styles. Our team doesn't have much experience building websites, so we just went with the standard styles. We used TailwindCSS.

Do you mean on the infinity.ai site or studio.infinity.ai? On infinity.ai we just wanted something fast and easy. This is MagicUI

Designers today are largely driven by trends (just like engineering?). Being cool = jumping on the latest bandwagon, not being unique or better. The good news is this particular style is pretty much restricted to tech companies, I think it started with https://neon.tech a few years ago or a similar startup.

Incidentally, the same behaviour is seen in academia. These websites for papers are all copying this one from 2020: https://nerfies.github.io/


I tried with the drake and drake saying some stuff and while its cool, its still lacking, like his teeth are disappearing partially :S

Agreed! The teeth can be problematic. The good news is we just need to train at higher resolution (right now we are at 320x320px), and that should resolve the teethe issue.

So far, we have purposely trained on low resolution to make sure we get the gross expressions / movements right. The final stage of training with be using higher resolution training data. Fingers crossed.


Realistic teeth in lipsync videos based purely on data and without explicit priors would be tough.

Good luck :)


Thanks for the feedback. The current model was trained at ~320x320 resolution. We believe going higher will result in better videos with finer detail, which we plan to do soon.

[flagged]


We are big fans of Hedra. Do you know if they've publicly commented on their model architecture? As far as we know, our particular choice of an end-to-end diffusion + transformer is novel.

We don't know what Hedra is doing. It could be the approach EMO has taken (https://humanaigc.github.io/emote-portrait-alive/) or VASA (https://www.microsoft.com/en-us/research/project/vasa-1/) or Loopy Avatar (https://loopyavatar.github.io/) or something else.


One of Hedra job offers for ML role lists "diffusion transformers".

Michael from Hedra, your choice is not novel :)

you could have at least made the ui different or picked different examples. pretty blatant rip.

[flagged]


hedra.com

How convenient that new user ‘jenintogen’ replied to new user ‘genaiguy’’s mentioning Hedra by asking for a link, giving you an opportunity to organically reply with the URL of your service.

Looks like there’s an enthusiastic marketplace of real grassroots users.


[flagged]


But no login required…

[flagged]


Neither your nor their model is remotely close to actually fooling anymore, so celebs could only be used for (very funny) obvious satire. I see no risk of harm here.

Also, two boxes for uploading the only two inputs to a model is not a new idea. One could say you stole it from Gradio (but even that's silly).


While I agree there are potential issues with using celebrity images, their UI is effectively no different to any of the 326432+ examples of handling model input on huggingface spaces.

@mjlbach: Weirdly broken web page. You can't even see samples without creating an account.

There's nothing particularly original about the UI, it's literally just a basic image upload and sound upload. I can easily see every hyperscaler AI firm offering something similar within one year so no need to get on your high horse about this.

What makes celebrity deepfakes worse than pleb deepfakes?

Say I’m a politician who gets caught on camera doing or saying something shady. Will your service do anything to prevent me from claiming the incriminating video was just faked using your technology? Maybe logging perceptual hashes of every output could prove that a video didn’t come from you?

These sort of models are probably going to end up released as publicly available weights at some point, right? Or, if it can be trained for $500k today, how much will it cost in a couple years? IMO we can’t stuff this genie back in the bottle, for better or worse. A video won’t be solid evidence of much within our lifetimes.

That's how I see it as well. Very soon, people will assume most videos are AI generated, and the burden of prove will be on people claiming videos are real. We plan to embed some kind of hash to indicate our video is AI generated, but people will be able to get around this. Google/Apple/Samsung seem to be in the best place to solve this: whenever their devices record a real video, they can generate a hash directly in HW for that video, which can be used to verify that it was actually recorded by that phone.

Also, I think it will cost around $100k to train a model at this quality level within 1-2 years. And, will only go down from there. So, the genie is out of the bag.


That makes sense. It isn’t reasonable to expect malicious users to helpfully set the “evil bit,” but you can at least add a little speedbump by hashing your own AI generated content (and the presence of videos that are verifiably AI generated will at least probably catch some particularly lazy/incompetent bad actors, which will destroy their credibility and also be really funny).

In the end though, the incentive and the capability lies in the hands of camera manufacturers. It is unfortunate that video from the pre-AI era have no real reason to have been made verifiable…

Anyway, recordings of politicians saying some pretty heinous things haven’t derailed some of their campaigns anyway, so maybe none of this is really worth worrying about in the first place.


Ya, it's only a matter of time until very high quality video models will be open sourced.

I think you're fine because these videos don't look or sound the least bit realistic



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: