• Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
Thursday, July 2026, 03, 05:09 AM
  • Login
mashdigi-Technology, new products, interesting news, trends
  • Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
No Result
View All Result
  • Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
No Result
View All Result
mashdigi-Technology, new products, interesting news, trends
No Result
View All Result
Home App

Adobe has teamed up with researchers from the Chinese University of Hong Kong and Johns Hopkins University to create a tool that can intuitively modify complex details in videos using AI.

Author: Mash Yang
2025-10-27
in App, Market dynamics, audio and video, Life, software
A A
0
Share to FacebookShare on TwitterShare to LINE

A new AI model developed by researchers from the Chinese University of Hong Kong, Adobe Research, and Johns Hopkins UniversityEditVerseThe biggest breakthrough is that it attempts to break the huge gap between traditional image editing and video editing, and proposes a unified framework that allows users to edit and generate complex details for videos with intuitive operations similar to editing pictures (Photoshop).

Adobe has teamed up with researchers from the Chinese University of Hong Kong and Johns Hopkins University to create a tool that can intuitively modify complex details in videos using AI.

The research team pointed out that the development of AI video editing has been limited in the past mainly due to architectural barriers (models are mostly image- or video-specific) and data scarcity (high-quality, annotated video data is far less than image data). This AI model, called "EditVerse," aims to simultaneously address these two major challenges.

Core technologies: Universal visual language and contextual learning

EditVerse's core methodology includes:

• Creating a "universal visual language":The model innovatively converts text, images, and videos into a unified, one-dimensional "token sequence" (data stream), enabling AI to understand and process visual information of different modalities in the same way.

• Powerful "context learning ability":Based on the Transformer model architecture and full self-attention, EditVerse can stitch together entire token sequences, including commands and original images. Through full self-attention, it accurately understands the relationships between various components (such as command text, specific objects in the video, and the style of the reference image). This design also enables it to flexibly handle inputs of varying resolutions and durations.

• Building a "knowledge transfer bridge":By adopting a unified framework, EditVerse can seamlessly transfer and apply the knowledge learned from massive image editing data (such as style and special effects) to film editing tasks, greatly alleviating the problem of scarce film data.

Adobe has teamed up with researchers from the Chinese University of Hong Kong and Johns Hopkins University to create a tool that can intuitively modify complex details in videos using AI.

Overcoming Data Scarcity and Establishing the EditVerseBench Benchmark

To address the problem of insufficient training data, the research team established a data production line, using a variety of dedicated AI models to automatically generate a large number of video editing samples. These samples were then screened through a visual language model (VLM), ultimately producing 23.2 high-quality video editing samples.

This batch of data was mixed with 600 million image editing samples and 390 million video generation samples for training, which enhanced the model's knowledge transfer capabilities.

To scientifically evaluate the model's effectiveness, the team also launched the industry's first comprehensive benchmark for prescriptive video editing: "EditVerseBench." This benchmark includes 100 videos of varying resolutions, covering 20 editing tasks.

Adobe has teamed up with researchers from the Chinese University of Hong Kong and Johns Hopkins University to create a tool that can intuitively modify complex details in videos using AI.

The effect exceeds Runway, demonstrating "emergence capability"

In the EditVerseBench performance test, EditVerse is ahead of existing open source models (such as TokenFlow, InsV2V, etc.) in multiple automated evaluation indicators (including video quality, text alignment, time consistency, VLM score, etc.).

More notably, EditVerse even outperformed the closed-source commercial model Runway Aleph in the VLM score (assessed by GPT-4o), which is closest to human preferences. In the real-person evaluation phase, EditVerse also received 51.7% user preference, surpassing Runway Aleph.

Researchers also discovered that EditVerse exhibits surprising "emergent abilities." Even when its video training data didn't include specific examples of "material transformation" or "special effects addition" (e.g., turning a turtle into crystal or adding a time-lapse effect to the sky), the model still understood the instructions and successfully completed the task.

Through ablation experiments (where the model's capabilities dropped significantly after removing image editing data), the team demonstrated that this "self-taught" ability primarily stems from the deep visual principles learned by the model from massive amounts of image data, and successfully transferred these principles to the field of video editing.

A new era of creation

The emergence of EditVerse not only provides a powerful new tool, but also may herald the arrival of a new content creation paradigm that moves from separation to unification, from cumbersome to simple, and is expected to popularize professional-level video editing capabilities to more creators.

Current related papers, project homepages and test codesAll have been made public.

Tags: AdobeAdobe ResearchAIEditVerseRunwayArtificial wisdomJohns Hopkins University researchersChinese University of Hong Kong
ShareTweetShare
Mash Yang

Mash Yang

Founder and editor of mashdigi.com, and student of technology journalism.

Leave a Reply Cancel Reply

The email address that must be filled in to post a message will not be made public. Required fields are marked as *

This site uses Akismet service to reduce spam.Learn more about how Akismet processes website visitor comments.

Translation (Tanslate)

Recent updates:

Quick Hands-on Review: The Google Pixel 10a in Strawberry Pink – A Great Value Choice for Enhanced Gemini AI Experience and Cross-Platform Transfer.

Quick Hands-on Review: The Google Pixel 10a in Strawberry Pink – A Great Value Choice for Enhanced Gemini AI Experience and Cross-Platform Transfer.

2026-03-05
Tired of frequent crashes? OpenAI reportedly secretly building a code hosting platform, aiming to challenge Microsoft's GitHub.

Tired of frequent crashes? OpenAI reportedly secretly building a code hosting platform, aiming to challenge Microsoft's GitHub.

2026-03-05
Apple drops a bombshell at $599! The all-new 13-inch MacBook Neo, powered by the A18 Pro processor, is unveiled, aiming to capture the affordable AI PC market.

Apple drops a bombshell at $599! The all-new 13-inch MacBook Neo, powered by the A18 Pro processor, is unveiled, aiming to capture the affordable AI PC market.

2026-03-04
mashdigi-Technology, new products, interesting news, trends

Copyright © 2017 mashdigi.com

  • About mashdigi.com
  • Place ads
  • Contact mashdigi.com

Follow us

Welcome back!

Login to your account below

Forgotten Password?

Retrieve your password

Hãy nhập tên người dùng hoặc địa chỉ email để mở mật khẩu

Log In
No Result
View All Result
  • About mashdigi.com
  • Place ads
  • Contact mashdigi.com

Copyright © 2017 mashdigi.com