Thursday, 16 Apr 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI > From reality to fantasy: Live2Diff AI brings instant video stylization to life
AI

From reality to fantasy: Live2Diff AI brings instant video stylization to life

Last updated: July 18, 2024 5:29 am
Published July 18, 2024
Share
From reality to fantasy: Live2Diff AI brings instant video stylization to life
SHARE

Be part of our each day and weekly newsletters for the most recent updates and unique content material on industry-leading AI protection. Be taught Extra


A staff of worldwide researchers has developed an AI system able to reimagining stay video streams into stylized content material in close to real-time. The brand new know-how, known as Live2Diff, processes stay video at 16 frames per second on high-end shopper {hardware}, probably reshaping functions from leisure to augmented actuality experiences.

Live2Diff, created by scientists from Shanghai AI Lab, Max Planck Institute for Informatics, and Nanyang Technological University, marks the primary profitable implementation of uni-directional consideration modeling in video diffusion fashions for live-stream processing.

Live2Diff is the primary try that permits uni-directional consideration modeling to video diffusion fashions for stay video steam processing.

It achieves 16FPS on RTX 4090 GPU ?

Hyperlinks ⬇️ pic.twitter.com/L2HP4QOK8j

— Dreaming Tulpa ?? (@dreamingtulpa) July 17, 2024

“We current Live2Diff, the primary try at designing a video diffusion mannequin with uni-directional temporal consideration, particularly focusing on live-streaming video translation,” the researchers clarify of their paper published on arXiv.

This novel strategy overcomes a big hurdle in video AI. Present state-of-the-art fashions depend on bi-directional temporal consideration, which requires entry to future frames and makes real-time processing not possible. Live2Diff’s uni-directional technique maintains temporal consistency by correlating every body with its predecessors and some preliminary warmup frames, eliminating the necessity for future body knowledge.

Live2Diff in motion: A sequence exhibiting the AI system’s real-time transformation capabilities, from an authentic portrait (left) to stylized variations together with anime-inspired, angular inventive, and pixelated renderings. The know-how demonstrates potential functions in leisure, social media, and inventive industries. (Video Credit score: Live2Diff)

Actual-time video model switch: The subsequent frontier in digital content material creation

Dr. Kai Chen, the undertaking’s corresponding writer from Shanghai AI Lab, explains within the paper, “Our strategy ensures temporal consistency and smoothness with none future frames. This opens up new prospects for stay video translation and processing.”

See also  Augmented/virtual reality can help extend critical infrastructure lifespan

The staff demonstrated Live2Diff’s capabilities by reworking stay webcam enter of human faces into anime-style characters in real-time. Intensive experiments confirmed that the system outperformed present strategies in temporal smoothness and effectivity, as confirmed by each quantitative metrics and person research.

A schematic diagram of Live2Diff’s revolutionary strategy: (a) The coaching stage incorporates depth estimation and a novel consideration masks, whereas (b) the streaming inference stage employs a multi-timestep cache for real-time video processing. This know-how marks a big leap in AI-powered stay video translation. (Credit score: live2diff.github.io)

The implications of Live2Diff are far-reaching and multifaceted. Within the leisure {industry}, this know-how might redefine stay streaming and digital occasions. Think about watching a live performance the place the performers are immediately reworked into animated characters, or a sports activities broadcast the place gamers morph into superhero variations of themselves in real-time. For content material creators and influencers, it affords a brand new instrument for artistic expression, permitting them to current distinctive, stylized variations of themselves throughout stay streams or video calls.

Within the realm of augmented actuality (AR) and digital actuality (VR), Live2Diff might improve immersive experiences. By enabling real-time model switch in stay video feeds, it might bridge the hole between the actual world and digital environments extra seamlessly than ever earlier than. This might have functions in gaming, digital tourism, and even in skilled fields like structure or design, the place real-time visualization of stylized environments might help in decision-making processes.

A Comparative Evaluation of AI Video Processing: The unique picture (high left) is reworked utilizing varied AI methods, together with Live2Diff (high proper), in response to the immediate ‘Breakdancing within the alley.’ Every technique showcases distinct interpretations, from stylized animation to nuanced actuality alterations, illustrating the evolving panorama of AI-driven video manipulation. (Video Credit score: Live2Diff)

Nevertheless, as with every highly effective AI instrument, Live2Diff additionally raises necessary moral and societal questions. The flexibility to change stay video streams in real-time might probably be misused for creating deceptive content material or deepfakes. It could additionally blur the strains between actuality and fiction in digital media, necessitating new types of media literacy. As this know-how matures, it is going to be essential for builders, policymakers, and ethicists to work collectively to ascertain tips for its accountable use and implementation.

See also  Why your enterprise AI strategy needs both open and closed models: The TCO reality check

The way forward for video AI: Open-source innovation and {industry} functions

Whereas the full code for Live2Diff is pending launch (anticipated to launch subsequent week), the analysis staff has made their paper publicly available and plans to open-source their implementation quickly. This transfer is anticipated to spur additional improvements in real-time video AI.

As synthetic intelligence continues to advance in media processing, Live2Diff represents an thrilling leap ahead. Its potential to deal with stay video streams at interactive speeds might quickly discover functions in stay occasion broadcasts, next-generation video conferencing techniques, and past, pushing the boundaries of real-time AI-driven video manipulation.


Source link
TAGGED: brings, Fantasy, Instant, life, Live2Diff, reality, stylization, Video
Share This Article
Twitter Email Copy Link Print
Previous Article Don't overlook unstructured data - Data Centre Review Don’t overlook unstructured data – Data Centre Review
Next Article JaxBio Technologies JaxBio Technologies Secures €2.5M Grant from European Innovation Council Accelerator
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Triton Anchor Completes $2.2M Seed; Secures $3.5M Grant

Triton Anchor, a Chelmsford, MA-based offshore anchoring techniques firm, raised $2.2M in Seed funding and…

December 17, 2024

Apple plans big Siri update with help from Google AI

Apple is planning to make use of a customized model of Google’s Gemini mannequin to…

November 8, 2025

Belfast launches UK hub to drive digital twin technology innovation

A cutting-edge innovation hub has launched in Belfast, positioning the UK on the forefront of…

May 2, 2025

OurRitual Raises $5.2M in Seed Funding

Dr. Orna Guralnik, OurRitual’s Chief Scientific Officer OurRitual, a NYC-based platform targeted on serving to…

July 24, 2024

Geostar pioneers GEO as traditional SEO faces 25% decline from AI chatbots, Gartner says

The second Mack McConnell knew the whole lot about search had modified got here final…

November 1, 2025

You Might Also Like

AI Safety Benchmarks Are Falling Behind
AI

AI Safety Benchmarks Are Falling Behind

By saad
Citizen developers now have their own Wingman
AI

Citizen developers now have their own Wingman

By saad
Commvault launches a ‘Ctrl-Z’ for cloud AI workloads
AI

Commvault launches a ‘Ctrl-Z’ for cloud AI workloads

By saad
Agricultural drones get smarter for large farm holdings
AI

Agricultural drones get smarter for large farm holdings

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.