Adobe Unveils New AI Tools for Video and Image Editing at Max 2025
At its annual Max conference, Adobe introduced a new generation of AI-powered creative tools, including Project Frame Forward for video editing, Project Light Touch for lighting control, and several other experimental projects that push the boundaries of digital creativity.
Project Frame Forward: editing without masks
The standout reveal was Project Frame Forward — a tool that enables creators to edit video content without manually applying masks. Instead, the system uses generative AI to detect and isolate objects in each frame automatically.
The technology allows users to modify movement, background, or texture directly, significantly reducing post-production time. Adobe claims the system learns from context, understanding where the subject ends and the environment begins — effectively “thinking” like a human editor.
Project Light Touch and Clean Take
Another highlight was Project Light Touch, which applies generative AI to control lighting in photographs. With it, photographers can adjust light direction, tone, and intensity as if they were operating a digital studio.
The company also unveiled Project Clean Take, an experimental speech-editing tool capable of subtly modifying pronunciation and pacing in recorded dialogue — ideal for correcting errors without re-recording.
Together, these tools represent Adobe’s push toward more intuitive, context-aware editing systems that merge professional control with AI automation.
AI meets 3D and document automation
Beyond creative editing, Adobe announced several projects integrating AI into 3D workflows and document management. New experimental features can transform object textures, adapt materials dynamically, and simulate physical lighting within 3D spaces.
One of the biggest surprises was the introduction of Acrobat Studio — a unified platform combining the classic Acrobat, Adobe Express, and generative assistants. This ecosystem will allow users to automate document analysis, generate summaries, and create visual layouts directly from textual content.
Availability and future release
All newly presented tools are currently in the research or early preview stage and are not yet available to the general public. Adobe did not specify release dates but hinted that integration into Creative Cloud could begin in 2026.
Developers and beta testers at Max 2025 were given early access to prototype builds, emphasizing real-world feedback to refine usability and creative control.
Conclusion
Adobe’s Max 2025 lineup shows how far generative AI has come in creative industries. From frame-precise video edits to intelligent lighting and speech correction, the company is positioning itself at the forefront of AI-assisted content creation.
While the tools remain experimental, their potential signals a future where AI becomes not just a helper, but a genuine creative partner for editors, photographers, and designers alike.
Editorial Team — CoinBotLab