Close Menu
    What's Hot

    Small language models

    April 16, 2024

    Llama 2: Open Foundation and Fine-Tuned Chat Models

    April 1, 2024

    The developer’s guide to open source LLMs and generative AI

    March 19, 2024
    Facebook X (Twitter) Instagram
    Facebook Instagram LinkedIn
    AI VentunoAI Ventuno
    • Home
    • AI Giants
      1. Meta (Facebook)
      2. Google
      3. Amazon
      4. View All

      Llama 2: Open Foundation and Fine-Tuned Chat Models

      April 1, 2024

      Introducing Gemini: Google’s AI Gets a Fresh Identity!

      February 10, 2024

      Google’s Bard chatbot gets the Gemini Pro update globally

      February 2, 2024

      Google’s Lumiere brings AI video closer to real than unreal.

      January 28, 2024

      Google Introduces Gemini, a Cutting-Edge Language Model Set

      January 10, 2024
      8.9

      DJI Avata Review: Immersive FPV Flying For Drone Enthusiasts

      January 15, 2021
      8.9

      Bose QuietComfort Earbuds II: Noise-Cancellation Kings Reviewed

      January 15, 2021

      Thousands Of PC Games Discounted In New Black Friday Sale

      January 15, 2021

      Take Your Photography to The Next Level with This Drone

      January 14, 2021

      Will Using a VPN on Phone Helps Protect You from Ransomware?

      January 14, 2021

      Popular New Xbox Game Pass Game Being Review Bombed With “0s”

      January 14, 2021

      Google Says Surveillance Vendor Targeted Samsung Phones

      January 14, 2021

      Why Are iPhones More Expensive Than Android Phones?

      January 14, 2021
    • Papers
    • Tools
      • Prompts
    • About us
    AI VentunoAI Ventuno
    Home » Google’s Lumiere brings AI video closer to real than unreal.
    AI Video

    Google’s Lumiere brings AI video closer to real than unreal.

    ai_adminBy ai_adminJanuary 28, 2024No Comments3 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Google Imagen clip
    Google Imagen clip
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Google’s new AI video generation model Lumiere uses a new diffusion model called Space-Time-U-Net, or STUNet, that figures out where things are in a video (space) and how they simultaneously move and change (time). Ars Technica reports this method lets Lumiere create the video in one process instead of putting smaller still frames together. 

    Lumiere starts with creating a base frame from the prompt. Then, it uses the STUNet framework to begin approximating where objects within that frame will move to create more frames that flow into each other, creating the appearance of seamless motion. Lumiere also generates 80 frames compared to 25 frames from Stable Video Diffusion.

    AI video generation and editing tools have gone from uncanny valley to near realistic in just a few years. It also establishes Google’s tech in the space already occupied by competitors like Runway, Stable Video Diffusion, or Meta’s Emu. Runway, one of the first mass-market text-to-video platforms, released Runway Gen-2 in March last year and has started to offer more realistic-looking videos. Runway videos also have a hard time portraying movement.

    Google published clips and prompts on the Lumiere site, which let me put the same prompts through Runway for comparison.

    Here are the results:

    Google Lumiere-generated video
    Runway-generated video

    Some of the clips presented have a touch of artificiality, especially if you look closely at skin texture or if the scene is more atmospheric. But look at that turtle! It moves like a turtle actually would in water! It looks like a real turtle!

    Other models stitch videos together from generated key frames where the movement already happened (think of drawings in a flip book), while STUNet lets Lumiere focus on the movement itself based on where the generated content should be at a given time in the video.

    Google has not been a big player in the text-to-video category, but it has slowly released more advanced AI models and leaned into a more multimodal focus. Its Gemini large language model will eventually bring image generation to Bard. Lumiere is not yet available for testing, but it shows Google’s capability to develop an AI video platform that is comparable to — and arguably a bit better than — generally available AI video generators like Runway and Pika. And just a reminder, this was where Google was with AI video two years ago.

    AI video Generators

    Google Imagen clip from 2022 Image: Google

    Beyond text-to-video generation, Lumiere will also allow for image-to-video generation, stylized generation, which lets users make videos in a specific style, cinemagraphs that animate only a portion of a video, and inpainting to mask out an area of the video to change the color or pattern. 

    Google’s Lumiere paper, though, noted that “there is a risk of misuse for creating fake or harmful content with our technology, and we believe that it is crucial to develop and apply tools for detecting biases and malicious use cases to ensure a safe and fair use.” The paper’s authors didn’t explain how this can be achieved.

    AI Video Google Lumiere
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email

    Related Posts

    Facebook

    Llama 2: Open Foundation and Fine-Tuned Chat Models

    April 1, 2024
    Gemini

    Introducing Gemini: Google’s AI Gets a Fresh Identity!

    February 10, 2024
    Chatbot

    Google’s Bard chatbot gets the Gemini Pro update globally

    February 2, 2024
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs

    January 23, 202468 Views

    Single-View 3D Human Digitalization with Large Reconstruction Models

    January 23, 202446 Views

    Llama 2: Open Foundation and Fine-Tuned Chat Models

    April 1, 202436 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews
    85
    Featured

    Pico 4 Review: Should You Actually Buy One Instead Of Quest 2?

    ai_adminJanuary 15, 2021
    8.1
    Uncategorized

    A Review of the Venus Optics Argus 18mm f/0.95 MFT APO Lens

    ai_adminJanuary 15, 2021
    8.9
    Editor's Picks

    DJI Avata Review: Immersive FPV Flying For Drone Enthusiasts

    ai_adminJanuary 15, 2021
    Our Picks

    Small language models

    April 16, 2024

    Llama 2: Open Foundation and Fine-Tuned Chat Models

    April 1, 2024

    The developer’s guide to open source LLMs and generative AI

    March 19, 2024
    Most Popular

    Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs

    January 23, 202468 Views

    Single-View 3D Human Digitalization with Large Reconstruction Models

    January 23, 202446 Views

    Llama 2: Open Foundation and Fine-Tuned Chat Models

    April 1, 202436 Views
    Latest Papers

    Llama 2: Open Foundation and Fine-Tuned Chat Models

    April 1, 2024

    Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs

    January 23, 2024

    Single-View 3D Human Digitalization with Large Reconstruction Models

    January 23, 2024
    AI Ventuno
    Facebook X (Twitter) Instagram Pinterest
    • Home
    • Technology
    • Language Models
    • Tools
    • About us
    © 2025 AI Ventuno. Designed by Ventuno Studio.

    Type above and press Enter to search. Press Esc to cancel.