this post was submitted on 28 Nov 2023
87 points (93.9% liked)

Technology

59298 readers
4871 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
top 8 comments
sorted by: hot top controversial new old
[–] randon31415@lemmy.world 13 points 11 months ago (2 children)

... How is this better than Animatediff?

[–] Zarxrax@lemmy.world 11 points 11 months ago

At the moment, it doesn't really seem to be that much better. Maybe the frames generated are a little more coherent, but you have less control over it. And from what I have seen so far, it just does very simple motions, like billowing smoke, a turning head, or clouds moving through the sky.

[–] Aeryn@lemmy.blahaj.zone 1 points 11 months ago (1 children)

This is the first I've heard of Animatediff. Where do you hear about stuff like that? Lemmy? Reddit? I try to stay up to date on AI stuff, but it feels like I have to do so much digging to get new info.

[–] randon31415@lemmy.world 3 points 11 months ago

I just keep an eye on the extensions list for Automatic1111: https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Extensions

[–] autotldr@lemmings.world 4 points 11 months ago

This is the best summary I could come up with:


It's an open-weights preview of two AI models that use a technique called image-to-video, and it can run locally on a machine with an Nvidia GPU.

Last year, Stability AI made waves with the release of Stable Diffusion, an "open weights" image synthesis model that kick started a wave of open image synthesis and inspired a large community of hobbyists that have built off the technology with their own custom fine-tunings.

They can operate at varying speeds from 3 to 30 frames per second, and they output short (typically 2-4 second-long) MP4 video clips at 576×1024 resolution.

In our local testing, a 14-frame generation took about 30 minutes to create on an Nvidia RTX 3060 graphics card, but users can experiment with running the models much faster on the cloud through services like Hugging Face and Replicate (some of which you may need to pay for).

We've previously covered other AI video synthesis methods, including those from Meta, Google, and Adobe.

Stability AI says it is also working on a text-to-video model, which will allow the creation of short video clips using written prompts instead of images.


The original article contains 553 words, the summary contains 187 words. Saved 66%. I'm a bot and I'm open source!

[–] LifeInOregon@lemmy.world 2 points 11 months ago

Mmmm. I want a cold, refreshing can of Slurge!

[–] Zoomboingding@lemmy.world 2 points 11 months ago

I should give this a try, but 30 minutes to generate 14 frames sounds much worse than what I can do currently. Usually a SD gen will take about 15 sec for 1 frame.

[–] r3df0x@7.62x54r.ru 1 points 11 months ago

As soon as they add text to speech AI and video syncing, I'm going to use it for memes and shitposts.