oer25-demystifying-ai

1

Hello, I’m Tim Fransen:

Technical Tutor in Generative AI, Web Publishing and Graphic Design at London South Bank University (LSBU), teaching visual communication in the School of Art and Social Science & the School of Engineering and Design.

Member of the LSBU AI Working Group.

Researcher at LSBU’s Digital x Data Research Centre.

Co-founder of the Creative and Responsible AI Lab (CRAIL).

Co-Investigator & Technical Lead on a Responsible AI (RAi) Skills Project: Ways of (Machine) Seeing (WoMS). Join us for the Teachers' Resource Launch Event at The Photographers’ Gallery on 28 June.

Defining AI

2

Generative AI is a type of artificial intelligence (AI) designed to generate content, such as text, images, sound or video. There are lots of applications that use generative AI, including the production of art or music, or generating text for chatbots. For example, generative AI art applications can generate an image based on a prompt, such as “make me a picture of a dragon reading a book”. Generative AI art is created using machine learning models trained on millions of images of existing art. The resulting images may replicate the style of an artist, without the original artist knowing or approving. Generative AI applications are becoming more and more commonplace and often you cannot tell that generative AI is being used.

Machine learning (ML) is an approach used to design and build artificial intelligence (AI) systems. ML is said to ‘learn’ by using examples in the form of data, instead of executing step-by-step instructions. In other words, ML applications are data-driven. For example, an ML application is used to recognise speech. It is based on many examples of people speaking in different accents and tones of voice. Other ML applications include identifying objects in images or playing complex games. Each ML application is designed to solve a specific problem.

A machine learning (ML) model is used by an ML application to complete a task or solve a problem. The ML model is a representation of the problem being solved. ML developers use vast amounts of data representative of a specific problem to train a model to detect patterns. The result of the training is a model, which is used to make predictions about new data in the same context. For example, self-driving cars are built using ML models to predict when to stop. The models are trained using millions of examples of situations in which cars need to stop. There are many different types of models, using different kinds of data, and different ways of training the models. All ML models are trained to detect patterns in the training data to make predictions about new data.

A machine learning (ML) model card is a way of documenting essential information about ML models in a structured way. ML model cards are written by ML developers for both experts and non-experts. For example, an ML application is developed to translate different languages, such as from Arabic to French and vice versa. A model card includes information on the model’s translation accuracy, as well as the model’s performance around jargon, slang, and dialects. Other model card information might include the type of ML model, different performance indicators, and even known bias. Model cards are created during the explanation stage of the AI project lifecycle to expose information on the model’s capabilities and limitations, in a way that is easy to understand.

Example Model Card: Stable Diffusion v1

For instance, we can discern from the calculations provided by the Machine Learning Impact Calculator that an estimated 11,250 kg of CO₂-eq (kilograms of carbon dioxide equivalent) was required to train Stable Diffusion v1. This is roughly equivalent to the annual emissions of 4.2 average UK households.

You can find a further 28 explanations in a glossary that is part of Raspberry Pi Foundation’s Experience AI Lessons.

3

Founded in 2019, Stability AI is a UK-based company behind Stable Diffusion, one of the most widely used text-to-image AI models. While often considered "open-source," it is technically open-weight, meaning the model weights are publicly available, but the full training process is not.

Stable Diffusion – Pros and Cons

Pros

Open-Weight Accessibility – Stability AI provides pre-trained models with open weights via platforms like Hugging Face. This enables users to run, fine-tune and modify models freely, supporting innovation and experimentation within the AI community.

Energy-Efficient AI – Stable Diffusion is optimised for consumer-grade GPUs, significantly reducing computational energy use compared to larger AI models. This makes generative AI more environmentally sustainable.

Lower Financial Barriers – Unlike proprietary models that depend on costly cloud infrastructure, Stable Diffusion is freely downloadable and can be run locally. This lowers costs for individuals, educators, researchers and small businesses.

Advancing AI Research – Stability AI collaborates with universities and research institutions to promote open and ethical AI development. These partnerships help advance the field of generative models while maintaining a degree of transparency.

Cons

Potential for Misuse – The open-weight nature of Stable Diffusion, while empowering, also makes it vulnerable to misuse. It has been used to create misleading, harmful or explicit content, raising concerns around content moderation and responsible deployment.

Copyright and Ethical Concerns – Stable Diffusion’s use of the LAION-5B dataset has led to litigation in the UK, particularly Getty Images v Stability AI at the High Court, concerning the unconsented use of millions of copyrighted photographs. Furthermore, the LAION-5B dataset contained problematic material, ranging from links to child sexual abuse content to hateful and private imagery, raising serious ethical and safety concerns about data provenance and the need for stricter curation.

4

Inside the Stable Diffusion Black Box:

C

E

D

F

A

B

A Load Checkpoint Node

Loads the main components required for image generation: the denoising model (e.g. Stable Diffusion 1.5), the text encoder (CLIP) and the VAE.

B Empty Latent Image Node

Generates a random Gaussian noise tensor (a multi-dimensional array of numbers used to represent image data) and sets the output image dimensions.

C CLIP Text Encoder Nodes

Converts text prompts (positive/negative) into numerical data that guide image generation and defines which features to include or exclude from the generated image.

D KSampler Node

Performs step-by-step denoising on the latent image using guidance from the prompt and model settings such as seed (for repeatability) and guidance scale (cfg) for prompt influence strength.

DENOISING

SEED

E VAE Decode Node

Converts the denoised latent numerical data into a viewable pixel-based image.

F Save Image Node

Saves the final image to the local output folder for access and use.

5

Hugging Face offers cloud hosting for AI models through its Inference API and Spaces, a platform where users can deploy, share and interact with machine learning models, including Stable Diffusion, via web-based applications.

Visit:

https://tinyurl.com/jxm42mdv

Try Out Stable Diffusion on HuggingFace:

1. Visit one of the following Spaces:

i Stable Diffusion 3.5 Large Turbo (8B)

ii Stable Diffusion 3.5 Large Turbo (8B)

2. Add your ‘Professor’ image to the MURAL

pinboard: https://tinyurl.com/mvm7yrd4

Stable Diffusion 3.5 Prompt Guide

6

Bias in AI Image Generation:

AI image generation models can reflect and amplify biases present in their training data, leading to stereotypical, exclusionary, or inaccurate representations of people, cultures, and concepts.

Bias can arise from imbalanced datasets, algorithmic assumptions, or societal influences, affecting diversity in AI-generated images. Addressing these issues requires inclusive datasets, ethical model training, and transparency in AI development to promote fairness and representation in generated content.

8

Anthropomorphism Hinders AI Education:

Generative AI models do not create, imagine, or understand images as humans do; they analyse data and generate outputs based on patterns and probabilities.

Using human-like terms such as "AI sees" or "AI learns" can create misconceptions, making it seem as though AI has intent or perception. In reality, Generative AI processes pixel data and statistical relationships to produce images without artistic expression or understanding.

Educators should describe AI as a system that generates images by recognising and replicating patterns, helping students critically engage with its true capabilities and limitations.

9

Environmental Impact:

Fawzi Ammache, known online as @futurewithfawzi, is a designer and engineer committed to making AI more accessible and understandable. Through educational content across multiple platforms, he simplifies complex AI concepts, helping a broader audience engage with and demystify AI.

Sustainable AI Image Generation – Some Best Practices

Use Efficient AI Models – Choose lightweight models (e.g., Stable Diffusion Turbo) and lower sampling steps.

Run on Energy-Efficient GPUs – Use modern GPUs (Apple M-series) for lower power use.

Use Local Processing to Reduce Water Waste – Cloud data centres require heavy cooling; local computers use less water.

Pick Cold-Climate Cloud Servers – If using the cloud, choose data centres in Iceland, Canada, or Finland for lower water use.

Avoid Unnecessary Computation – Don’t generate excessive images; use AI upscaling instead of full-resolution renders.

Power AI with Renewable Energy – Use solar/wind energy for local GPUs or select green cloud providers.

This helps keep AI creative work sustainable and responsible!

Going Local:

10

ComfyUI is a free, open-source, local-hosted, node-based graphical interface focused on building customisable AI image-generation workflows using Stable Diffusion models, with support for video and audio workflows through extensions and custom nodes.

THANKS FOR LISTENING!

Projects:

Here are several projects, including a board game, a concept album, sketches, and a CustomGPT example, which serves as a formative assessment tool designed to assist students produce more effective event posters. These projects explore how Generative AI outputs can be integrated with traditional digital design tools and techniques, as well as assessment for learning.