AI Image Generation - The Fun Stuff and a Few Ethical Questions.

Adobe Firefly AI generated image: Red Teaming bear unhappy that their initial access payload isn't working! :-(
Hmmmm ... why won't this initial access payload work on my target? It was ok in my lab.

I’ve started taking a closer look at AI generated imagery more recently, firstly out of a sense of fun, but also due to a curiosity of the ethical issues. Spoiler alert ... I didn't create any of the images 'from scratch' for this post. I faked it all with Adobe Firefly 2.

💡
While this article features the use of ‘Adobe Firefly’. I paid for it out of my own pocket. Adobe has not sponsored me to write this article, nor should it be considered as a review of the software.

For those that know me, while I’ve got a deep appreciation for all things creative, I’m certainly more of technical person. I am familiar with the mechanics of using tools such as Photoshop due to my earlier years as a digital media developer, where I would ‘cut up’ images created in Photoshop and bring them to life in all manner of things such as interactive kiosks, demos, CD-ROMs and the like.

While I could compose a specific user interface view by configuring the visibility of appropriately labelled layers, I never really ‘created’ anything original of my own. More lately I’ve been using Lightroom to enhance images due to my growing interest in photography. But 'creating' images myself is just something else.

One evening I was curious as to what ‘Firefly’ was all about after seeing it pop-up on my Adobe Cloud homepage. The tool comes as part of a paid-for subscription. I was amazed that I could just type in something and the software would turn out an image based on my prompt text. The results were pretty amazing indeed …. I could specify the ‘what’ through a text prompt and Firefly would take care of the ‘how’.

Adobe Firefly before the login: https://www.adobe.com/uk/products/firefly.html
Adobe Firefly (before login): https://www.adobe.com/uk/products/firefly.html

For somebody like me, on the surface it looked like a game changer … the ability to create new images for this site quickly and easily, with very limited artistic skills in the conventional sense. All manner of tweaks and options in Firefly can help to create everything from extremely detailed photos to futuristic cyberpunk style art. All you need is imagination and the ability to express this via a prompt.

While it was indeed fun, there were a few drawbacks - some more obvious than others. First of all, due to the sheer number of configuration options in Firefly and its ability to interpret very detailed prompts, the results were unpredictable. I suspect that the lack of predictability is actually part of the attraction though.

If you look really closely at the image at the top of this article, the computer equipment looks like it was stuck in an oven beforehand (it's a wonder the bear didn't complain about burns). Things look a bit more complicated with the image below:

Adobe Firefly AI Generated Image: Red Team asks for approval to buy more servers to attack AI Machine Learning Models
How much did the Red Team spend on servers to attack AI Machine Learning Models?

It looks like there is a human hand trying to escape from the bear. Mildly creepy. Obviously the AI generated image had used some other real image as a basis to create something 'unique' for me. That image was obviously scraped from somewhere else - such as a stock photo library or from someone's personal site where they could have created the actual original image themselves.

💡
If the quality of AI generated content depends on the quality of the prompt, will prompt engineering become the new search engine optimisation?

Did Adobe get the permission from the publisher of the original content? I can't answer that, but it reflects a bigger argument around generative AI platforms where large amounts of original material were scraped from sites without the knowledge or permission of the site owner.

While some may argue that if you publish something on the Internet, then it is fair game for other people to copy it. But what if it is copied and used to train AI machine learning models on an industrial scale by a service that ultimately seeks to make money? Will the original creators get paid for their efforts? I doubt it. There are quite a few law suits happening because of the very same issue.

I doubt any of my concerns will stop generative AI tools from becoming more popular - either for fun meme generation or commercial purposes. Permission as well as payment and attribution to original creators shouldn't be forgotten about though.

Although specific to 'conventional' text-based generative AI tools, I do believe that that the approach taken by Brave to integrate AI with search capabilities is a great step in the right direction. From the image below, you can see that I searched for how to setup port forwarding in linux.

Original content attribution by Brave
Original content attribution by Brave

There was an impressive amount of step-by-step instructions returned - 457 words. On first glance, it looked pretty comprehensive (I didn't reproduce the steps on how to setup port forwarding), but it was a nice touch on the part of Brave to quote their sources at the very least.

There is also something more important than attribution at stake here - transparency. It helps the user to determine if the content is trustworthy and accurate. Those two words I fear could get lost in the run-up to elections, which is certainly going to see generative AI being used (abused?) as a tool (weapon?) to aggressively pursue votes.