Adobe claims its AI tools are 'commercially safe', but does that stack up?

Image generated with Adobe Firefly showing a smiling cartoon boy with glasses, wearing an orange shirt and red shorts, walking down a pink and orange path in a colorful, candy-themed landscape with whimsical buildings
(Image credit: Adobe)

Adobe has built its reputation in artificial intelligence on a simple but powerful promise: its Firefly AI models are "commercially safe" because they're trained exclusively on licensed content. It's a compelling pitch for businesses worried about copyright lawsuits, and Adobe has marketed this heavily as their key differentiator in the crowded AI market.

But this week's announcement includes a new move that sounded like it was undermining that promise.

What does commercially safe mean?

This isn't just a theoretical concern. if you're a marketing agency creating content for a major client, you want assurance that your AI-generated videos won't trigger a copyright lawsuit. Adobe's promise gave businesses that confidence... or at least, it appeared to.

Now, though, Adobe are offering what they call "partner models" within the Firefly app. This means customers can choose between Adobe's own Firefly Video Model and third-party options like Runway's Gen-4 Video or Google's Veo3 with Audio. More partnerships are coming soon, including models from Topaz Labs, Moonvalley, Luma AI and Pika.

Image generated with Adobe Firefly showing an octopus with textured, patterned skin in shades of brown and blue, surrounded by coral and marine growth.

(Image credit: Adobe)

What exactly does Adobe mean by "commercially safe"? Well, unlike most AI companies, which scrape vast amounts of internet content to train their models, Adobe says it only uses images, videos and audio it has explicit permission to use. This includes stock photos from Adobe Stock, public domain content, and material that's been properly licensed.

But here's the issue: these partner companies haven't made the same training commitments as Adobe. Runway, for instance, has been deliberately vague about their training data, and Google's approach to AI training has involved scraping enormous amounts of web content. These models were built using different standards and potentially different sources of copyrighted material.

When Adobe processes content through their own Firefly models, it can stand behind its licensing claims because it controls the entire pipeline. But when a user selects a third-party model through the Firefly app, Adobe becomes merely a middleman. They can't guarantee the copyright safety of content they didn't create using data they didn't curate.

Adobe's response

When we contacted Adobe, a spokesperson was keen to stress that "Adobe’s Firefly family of creative models, spanning image, video, vector, audio, and 3D, has been designed from the ground up to be commercially safe, with models trained only on content we have the rights to use."

But crucially, in Adobe's mind at least, these 'Firefly models' are distinct from the Firefly app, which is where the third-party services are being integrated. (I was told these integrations are designed to "give customers the freedom to explore a wider range of distinct aesthetic styles and model personalities.")

The spokesperson went on to emphasise several safeguards Adobe has implemented for partner model integration. Most notably, they guarantee that "no matter which model a creator chooses to use within Adobe products, the content they generate or upload is never used to train generative AI models."

Image generated with Adobe Firefly showing a cartoon cow in an astronaut suit standing on the moon with Earth in the background, with controls for voice recording and a timeline with video and audio tracks visible below

(Image credit: Adobe)

Additionally, the spokesperson explained how Adobe automatically attaches Content Credentials to all AI-generated outputs in the Firefly app, which "clearly indicate whether content was created using our commercially safe Firefly model or a partner model".

All of which is very interesting. But none of which, to my mind, addresses the concern that if you're generating content via the Firefly app, it's not necessarily going to be commercially safe. And while Adobe thinks the distinction between 'the Firefly app' and 'the Firefly models' is clear, I'm not convinced it'll be obvious to most people.

It does make business sense

Don't get me wrong; I think Adobe's decision makes business sense, even if it compromises its core message. The AI video generation market is moving incredibly quickly, and Adobe's own models sometimes lag behind competitors in quality and features. By integrating third-party tools, it can offer users the latest capabilities without waiting for its own technology to catch up.

As for "commercial safety" of the Firefly models vs the Firefly app as a whole, Adobe's spokesperson believes that users can see the nuance. "This is how we’re seeing customers who want commercially safe content use the Firefly app," they explained. "Partner models are leveraged during the ideation phase to explore creative directions, but when it’s time to move into production, they turn to Adobe’s commercially safe Firefly models."

I truly hope this is indeed happening, and that no customer is accidentally putting themselves at legal risk while using the Firefly app. But I have to be honest, I've found this all quite confusing, and I doubt I'm alone. So I still think Adobe has its work cut out to ensure everyone using the Firefly app remains "commercially safe".

Thank you for reading 5 articles this month* Join now for unlimited access

Enjoy your first month for just £1 / $1 / €1

*Read 5 free articles per month without a subscription

Join now for unlimited access

Try first month for just £1 / $1 / €1

TOPICS
Tom May
Freelance journalist and editor

Tom May is an award-winning journalist and author specialising in design, photography and technology. His latest book, The 50th Greatest Designers, was released in June 2025. He's also author of the Amazon #1 bestseller Great TED Talks: Creativity, published by Pavilion Books, Tom was previously editor of Professional Photography magazine, associate editor at Creative Bloq, and deputy editor at net magazine. 

You must confirm your public display name before commenting

Please logout and then login again, you will then be prompted to enter your display name.