Skip Navigation

Search

Technology @lemmy.world
realitista @lemmy.world

Google demos new Lumiere text to video engine. Results are a huge leap forward from previous engines.

Google’s new video generation AI model Lumiere uses a new diffusion model called Space-Time-U-Net, or STUNet, that figures out where things are in a video (space) and how they simultaneously move and change (time). Ars Technica reports this method lets Lumiere create the video in one process instead of putting smaller still frames together.

Lumiere starts with creating a base frame from the prompt. Then, it uses the STUNet framework to begin approximating where objects within that frame will move to create more frames that flow into each other, creating the appearance of seamless motion. Lumiere also generates 80 frames compared to 25 frames from Stable Video Diffusion.

Beyond text-to-video generation, Lumiere will also allow for image-to-video generation, stylized generation, which lets users make videos in a specific style, cinemagraphs that animate only a portion of a video, and inpainting to mask out an area of the video to change the color or pattern.

Google’s Lumiere paper, th

Technology @lemmy.world
Memento Mori @lemmy.world
  • The AI wars might have an armistice deal sooner than expected
Technology @lemmy.world
Memento Mori @lemmy.world
  • Readers pick the best tech books of all time
Technology @lemmy.world
Richard @lemmy.world

Apple is a $3 trillion company — again

"While developers start work on building Vision Pro apps, the potential for people upgrading to the iPhone 15 this year is a big reason for investor optimism."

Technology @lemmy.world
BigFig @lemmy.world

Twitter has started blocking unregistered users