Human-AI Co-Creation: Evaluating the Impact of Large-scale Text-to-Image Generative Models on the Creative Process

Tommaso Turchi, Silvio Carta, Luciano Ambrosini, Alessio Malizia

Research output: Contribution to conferencePaperpeer-review

Abstract

Large-scale Text-to-image Generative Models (LTGMs) are a cutting-edge class of Artificial Intelligence (AI) algorithms specifically designed to generate images from natural language descriptions (prompts). These models have demonstrated impressive capabilities in creating high-quality images from a wide range of inputs, making them powerful tools for non-technical users to tap into their creativity. The field is advancing rapidly and we are witnessing the emergence of an increasing number of tools, such as DALL-E, MidJourney and StableDiffusion, that are leveraging LTGMs to support creative work across various domains. However, there is a lack of research on how the interaction with these tools might affect the users' creativity and their ability to control the generated outputs. In this paper, we investigate how the interaction with LTGMs-based tools might impact creativity by analyzing the feedback provided by groups of design students developing an architectural project with the help of LTGMs
Original languageEnglish
Publication statusAccepted/In press - 30 Apr 2014

Fingerprint

Dive into the research topics of 'Human-AI Co-Creation: Evaluating the Impact of Large-scale Text-to-Image Generative Models on the Creative Process'. Together they form a unique fingerprint.

Cite this