Maps: Difference between revisions
No edit summary |
No edit summary |
||
(9 intermediate revisions by the same user not shown) | |||
Line 58: | Line 58: | ||
What is particular about the map of this catalogue of objects of interest and necessity, is that it purely attempts to map autonomous and decentralised generative AI, serving as a map for a guided tour and experience of autonomous AI. However, both Hugging Face' dependency in venture capital and Stable Diffusion's dependency on hardware and infrastructure point to the fact that there are several planes that are not captured in the above map of this catalogue, but which are equally important. For instance, The [https://artificialintelligenceact.eu EU AI Act] or laws on copyright infringement, which Stable Diffusion (like any other AI ecology) will also depend on, point to a plane of governance and regulation. AI, including Stable Diffusion, also connects to the depends on the organisation of human labour, or the extraction of resources. | What is particular about the map of this catalogue of objects of interest and necessity, is that it purely attempts to map autonomous and decentralised generative AI, serving as a map for a guided tour and experience of autonomous AI. However, both Hugging Face' dependency in venture capital and Stable Diffusion's dependency on hardware and infrastructure point to the fact that there are several planes that are not captured in the above map of this catalogue, but which are equally important. For instance, The [https://artificialintelligenceact.eu EU AI Act] or laws on copyright infringement, which Stable Diffusion (like any other AI ecology) will also depend on, point to a plane of governance and regulation. AI, including Stable Diffusion, also connects to the depends on the organisation of human labour, or the extraction of resources. | ||
In describing the plane of objects of interest and necessity, we attempt to describe how Stable Diffusion and autonomous AI image generation build on dependencies to these different planes, but an overview of the many planes of AI and how it 'stacks' can of course also be the centre of a map in itself. One example of this is Kate Crawford's ''[https://katecrawford.net/atlas Atlas of AI]'', a book that displays different maps (and also images) that link AI to 'Earth' and the exploition of energy and minerals, or 'Labour' and the workers who do micro tasks ('clicking' tasks) or the workers in Amazon's warehouses. In continuation, Crawford's book also contains chapters on 'Data', 'Classification', 'Affect', 'State' and 'Power'. | In describing the plane of objects of interest and necessity, we attempt to describe how Stable Diffusion and autonomous AI image generation build on dependencies to these different planes, but an overview of the many planes of AI and how it 'stacks' can of course also be the centre of a map in itself. One example of this is Kate Crawford's ''[https://katecrawford.net/atlas Atlas of AI]'', a book that displays different maps (and also images) that link AI to 'Earth' and the exploition of energy and minerals, or 'Labour' and the workers who do micro tasks ('clicking' tasks) or the workers in Amazon's warehouses. In continuation, Crawford's book also contains chapters on 'Data', 'Classification', 'Affect', 'State' and 'Power'. | ||
[[File:Planes of ai.jpg|alt=A photo of a map and illustration of material, technical and other planes that AI objects exist between| none|thumb|640x640px||A map of how objects not only relate pixel space to latent space, but how they are always suspended between different planes - not only a technical one, but also an organisational one, a material one, and potentially many others (capital, labour, knowledge, governance, etc.) (by Christian Ulrik Andersen, Nicolas Maleve, and Pablo Velasco)]] | |||
[[File:Map_objects_and_planes.jpg|none|thumb|640x640px|A sketch map showing the same as above from collaborative workshop (by Christian Ulrik Andersen, Nicolas Maleve, and Pablo Velasco) ]] | |||
[[File:Gertraud Koch layers activity.jpg|none|thumb|640x640px|Five sets of technological activities. The layers wrap around technology as a material entity with its own agency, but the layers are permeable and interdependent | |||
Another abstraction of the layered nature of generative AI is found in Gertraud Koch's a map of all layers that she and her coauthors connects to "technological activity", and which would also pertain to AI.[5] On top of a layer of technology (the 'data models and algorithms') one will find other layers that are interdependent, and which contribute to the political and technological qualities of AI. As such, the map is also meant for navigation – to identify starting points for rethinking its concepts or reimagining alternative futures (in their work, particularly in relation to a potential delinking from a colonial past, and reimagining a pluriversality of technology) | |||
[[File:Gertraud Koch layers activity.jpg|none|thumb|640x640px|Five sets of technological activities. The layers wrap around technology as a material entity with its own agency, but the layers are permeable and interdependent (by Gertraud Koch).]] | |||
Within the many planes and stacks of AI one can find many different maps that build other types of overviews and conceptual models of AI – perhaps pointing to how maps themselves take part in making AI a reality. | Within the many planes and stacks of AI one can find many different maps that build other types of overviews and conceptual models of AI – perhaps pointing to how maps themselves take part in making AI a reality. | ||
Line 67: | Line 72: | ||
==== The corporate landscape ==== | ==== The corporate landscape ==== | ||
The entrepreneur, investor and pod cast host Matt Turck has made the “ultimate annual market map of the data/AI industry”. Since 2012 he has documented the corporate landscape of AI not just to identify key corporate actors, but also developments of trends in business. As he also notes in his blog, the first map from 2012 has merely 139 logos, whereas the 2024 version has 2,011 logos.[6] This reflects the massive investment in AI entrepreneurship, following first 'big data' and now 'generative AI' (and machine learning) - how AI has become a business reality. Comparing the 2012 version with the most recent map from 2024, one can see the corporate landscape of AI | The entrepreneur, investor and pod cast host Matt Turck has made the “ultimate annual market map of the data/AI industry”. Since 2012 he has documented the corporate landscape of AI not just to identify key corporate actors, but also developments of trends in business. As he also notes in his blog, the first map from 2012 has merely 139 logos, whereas the 2024 version has 2,011 logos.[6] This reflects the massive investment in AI entrepreneurship, following first 'big data' and now 'generative AI' (and machine learning) - how AI has become a business reality. Comparing the 2012 version with the most recent map from 2024, one can see the corporate landscape of AI changes over time. How, for instance, the division of companies dealing with infrastructure, data analytics, applications, data sources, and open source AI becomes fine grained over the years, forking out into, applications in health, finance and agriculture; or how privacy and security become of increased concern. Clearly, AI reconfigures and intersects with many different realities. [[File:Big Data Landscape 2012.png|none|thumb|640x640px|The corporate landscape of Big Data in 2012 (by Matt Turck and Shivon Zilis).[7] ]] | ||
==== Critical cartography in the mapping of AI ==== | ==== Critical cartography in the mapping of AI ==== | ||
In mapping AI there are also | In mapping AI there are also 'counter maps' or 'critical cartography'.[8] Conventional world maps are built on set principles of, for instance, North facing up, and Europe at the centre. The map is therefore not just a map for navigation, but also a map of more abstract imaginaries and histories originating in colonial times, where maps was the outset of Europe and an intrinsic part of the conquest of territories. In this sense, a map always also reflects hierarchies of power and control that can be inverted or exposed (for instance by turning the map upside down, letting the south be a point of departure). Counter-mapping technological territories would, following this logic, involve what the French research and design group Bureau d´Études has called "maps of contemporary political, social and economic systems that allow people to inform, reposition and empower themselves."[9] They are maps that reveal underlying structures of social, political or economic dependencies to expose what ought to be of common interest, or the hidden grounds on which a commons rests. Félix Guattari and Gilles Deleuze' notion of 'deterritorialization' can be useful, here, as a way to conceptualise the practices that expose and mutate the social, material, financial, political, or other organisation of relations and dependencies.[10] The aim is ultimately not only to destroy this 'territory' of relations and dependencies, but ultimately a 'reterritorialization' – a reconfiguration of the relations and dependencies. | ||
Conventional world maps are built on set principles of, for instance, North facing up, and Europe at the centre. The map is therefore not just a map for navigation, but also a map of more abstract imaginaries and histories originating in colonial times, where maps was the outset of Europe and an intrinsic part of the conquest of territories. In this sense, a map always also reflects hierarchies of power and control that can be inverted or exposed (for instance by turning the map upside down, letting the south be a point of departure). Counter-mapping technological territories would, following this logic, involve what the French research and design group Bureau d´Études has called " | |||
Utilising the opportunities of info-graphics in mapping can be a powerful tool. At the plane of financial dependencies, one can map, as Matt Turck, the corporate landscape of AI, but one can also draw a different map that reveals how the territory of 'startups' does not compare to a geographical map of land and continents. Strikingly, The United States is double the size of Europe and Asia, whereas there are whole countries and continents that are missing (such as Russia and Africa). This map thereby not only reflects the number of startups, but also how venture capital is dependent on other planes, such as politics and the organisation of capital, or infrastructural gaps. In Africa, for instance, | Utilising the opportunities of info-graphics in mapping can be a powerful tool. At the plane of financial dependencies, one can map, as Matt Turck, the corporate landscape of AI, but one can also draw a different map that reveals how the territory of 'startups' does not compare to a geographical map of land and continents. Strikingly, The United States is double the size of Europe and Asia, whereas there are whole countries and continents that are missing (such as Russia and Africa). This map thereby not only reflects the number of startups, but also how venture capital is dependent on other planes, such as politics and the organisation of capital, or infrastructural gaps. In Africa, for instance, the AI divide is very much also a 'digital divide', as argued by AI researcher Jean-Louis Fendji.[11] | ||
[[File:Numbers of newly funded AI startups.png|none|thumb|640x640px|Numbers of newly funded AI startups per country ( | [[File:Numbers of newly funded AI startups.png|none|thumb|640x640px|Numbers of newly funded AI startups per country (by Visual Capitalist).[12]]] | ||
Counter-mapping the organisation of relations and dependencies is also prevalent in the works of the Barcelona-based artist collective Estampa, which exposes how generative AI depends on different planes: venture capital, energy consumption, a supply chain of minerals, human labour, as well as other infrastructures, such as the internet, which is 'scraped' for images or other media, using e.g. software like [[Clip]]). [[File:Taller Estampa, map of generative AI.png|none|thumb|640x640px| | Counter-mapping the organisation of relations and dependencies is also prevalent in the works of the Barcelona-based artist collective Estampa, which exposes how generative AI depends on different planes: venture capital, energy consumption, a supply chain of minerals, human labour, as well as other infrastructures, such as the internet, which is 'scraped' for images or other media, using e.g. software like [[Clip]]). [[File:Taller Estampa, map of generative AI.png|none|thumb|640x640px|Map of generative AI (by Taller Estampa).[13]]] | ||
==== Epistemic mapping of AI ==== | ==== Epistemic mapping of AI ==== | ||
Maps of AI often also address how AI functions as what | Maps of AI often also address how AI functions as what Celia Lury has called an 'epistemic infrastructure'.[14] That is, AI is an apparatus that builds on knowledge, creates knowledge, but also shapes what knowledge is and we consider to be knowledge. To Lury, the question of 'methods' here becomes central - not as a neutral, 'objective' stance, as one typically regards good methodology in science, but as a cultural and social practice that help articulate the questions we ask and what we consider to be a problem in the first place. When one for, instance, criticises the social, racial or other biases in generative AI (such as all doctors being white males in generative AI image creation), we are not just dealing with bias in the dataset that can be fixed with '[[negative prompts]]' or other technical means. Rather, AI is fundamentally – in its very construction and infrastructure – based in a Eurocentric history of modernity and knowledge production. For instance, as pointed out by Rachel Adams, AI belongs to a genealogy of intelligence, and one also ought to ask, whose intelligence and understanding of knowledge is modelled within the technology – and whose is left out?[15] | ||
There are several attempts to map this territory in the plane of knowledge production, and its many social, material, political or other relations and dependencies. Sharing many of the concerns of Lury and Adams, Vladan Joler and Matteo Pasquinelli's ' | There are several attempts to map this territory in the plane of knowledge production, and its many social, material, political or other relations and dependencies. Sharing many of the concerns of Lury and Adams, Vladan Joler and Matteo Pasquinelli's 'Nooscope' is a good example of this.[16] In their understanding AI belongs to a much longer history of knowledge instruments ('nooscopes', from the Greek ''skopein'' ‘to examine, look’ and ''noos'' ‘knowledge’) that would also include optical instruments, but which in AI is a form of knowledge magnification of patterns and statistical correlations in data. The nooscope map is an abstraction of how AI functions as "Instrument of Knowledge Extractivism". It is therefore not a map of 'intelligence' and logical reasoning, but rather of a "regime of visibility and intelligibility" whose aim is the automation of labour, and of how this aim rests on (as other capitalist extractions of value in modernity) a division of labour – between humans and technology, between for instance historical biases in the selection and labelling of data, and their formalisation in sensors, databases and metadata. The map also refers to how selection, labelling and other laborious tasks in the training of models is done by "ghost workers" thereby referring to a broader geo-politics and body-politics of AI where human labour is often done by subjects of the Global South (although they might oppose being referred to as 'ghosts'). | ||
[[File:Nooscope.png|none|thumb|640x640px|alt=A map of AI as an instrument of knowledge by Vladan Joler and Matteo Pasquinelli (2020) | [[File:Nooscope.png|none|thumb|640x640px|alt=A map of AI as an instrument of knowledge by Vladan Joler and Matteo Pasquinelli (2020)|A map of AI as an instrument of knowledge (by Vladan Joler and Matteo Pasquinelli). ]] | ||
[[Category:Objects of Interest and Necessity]] | [[Category:Objects of Interest and Necessity]] | ||
Line 100: | Line 103: | ||
[5] Gertraud Koch et al., “Layers of Technology in Pluriversal Design: Decolonising Language Technology with the Live Language Initiative,” ''CoDesign'' 20, no. 1 (2024): 77–90, https://doi.org/10.1080/15710882.2024.2341799. | [5] Gertraud Koch et al., “Layers of Technology in Pluriversal Design: Decolonising Language Technology with the Live Language Initiative,” ''CoDesign'' 20, no. 1 (2024): 77–90, https://doi.org/10.1080/15710882.2024.2341799. | ||
[6] Matt Turck, “Full Steam Ahead: The 2024 MAD (Machine Learning, AI & Data) Landscape,” ''MattTurck.com'', March 2024, | [6] Matt Turck, “Full Steam Ahead: The 2024 MAD (Machine Learning, AI & Data) Landscape,” ''MattTurck.com'', March 2024, https://mattturck.com/MAD2024/. | ||
[7] Matt Turck, “A Chart of the Big Data Ecosystem,” ''MattTurck.com'', June 29, 2012, | [7] Matt Turck, “A Chart of the Big Data Ecosystem,” ''MattTurck.com'', June 29, 2012, https://mattturck.com/a-chart-of-the-big-data-ecosystem/. | ||
[8] | [8] Jeremy W. Crampton and John Krygier, “An Introduction to Critical Cartography,” ''ACME: An International Journal for Critical Geographies'' 4, no. 1 (2005): 11–33, https://doi.org/10.14288/acme.v4i1.723. | ||
[9] ''Bureau d'Études'', accessed August 11, 2025, https://bureaudetudes.org/. | |||
[10] Gilles Deleuze and Félix Guattari, ''A Thousand Plateaus: Capitalism and Schizophrenia'', trans. Brian Massumi (Minneapolis: University of Minnesota Press, 1987). | |||
[11] “From Digital Divide to AI Divide – Fellows’ Seminar by Jean-Louis Fendji,” ''STIAS'', April 9, 2024, https://stias.ac.za/2024/04/from-digital-divide-to-ai-divide-fellows-seminar-by-jean-louis-fendji/. | |||
[12] Marcus Lu, “Mapped: The Number of AI Startups by Country,” ''Visual Capitalist'', May 6, 2024, https://www.visualcapitalist.com/mapped-the-number-of-ai-startups-by-country/. | |||
[13] “Cartography of Generative AI,” ''Estampa'', accessed August 11, 2025, <nowiki>https://tallerestampa.com/en/estampa/cartography-of-generative-ai/</nowiki>. | |||
[14] Celia Lury, ''Problem Spaces: How and Why Methodology Matters'' (Cambridge, UK; Medford, MA: Polity, 2021). | |||
[15] Rachel Adams, “Can Artificial Intelligence Be Decolonized?,” ''Area'' 53, no. 1 (2021): 6–13, <nowiki>https://doi.org/10.1080/03080188.2020.1840225</nowiki>. | |||
[16] Vladan Joler and Matteo Pasquinelli, ''The Nooscope Manifested: AI as Instrument of Knowledge Extractivism'', 2020, <nowiki>https://fritz.ai/nooscope/</nowiki>. | |||
++++++++++++++++++++++++++++++++++++++++++++++++++++++ | ++++++++++++++++++++++++++++++++++++++++++++++++++++++ | ||
Line 122: | Line 141: | ||
==== [Images: 'Our' map(s) - perhaps surround by other maps] ==== | ==== [Images: 'Our' map(s) - perhaps surround by other maps] ==== | ||
== Guestbook == | |||
<eplite id="Objects_of_interest_and_necessity_Guestbook" height="600px" width="1000px" /> | |||
[[Category:Objects of Interest and Necessity]] | [[Category:Objects of Interest and Necessity]] |
Latest revision as of 19:00, 11 August 2025
Mapping 'objects of interest and necessity'
If one considers generative AI as an object, there is also a world of ‘para objects’, surrounding AI and shaping its reception and interpretation in the form of maps or diagrams of AI. They are drawn by both amateurs and professionals who need to represent processes that are otherwise sealed off in technical systems, but more generally reflect a need for abstraction – a need for conceptual models of how generative AI functions. However, as Alfred Korzybski famously put it, one should not confuse the map with the territory: the map is not how reality is, but a representation of reality.[1]
Following on from this, mapping the objects of interest in autonomous AI image creation is not to be understood as a map of what it 'really is'. Rather, it is a map of encounters of objects; encounters that can be documented and catalogued, but also positioned in a spatial dimension – representing a 'guided tour', and an experience of what objects are called, how they look, how they connect to other objects, communities or underlying infrastructures (see also Objects of interest and necessity). Perhaps, the map can even be used by others to navigate autonomous generative AI and create their own experiences? But, importantly to note, what is particular about the map of this catalogue of objects of interest and necessity, is that it is an attempt to map autonomous generative AI. It does, in other words, not map what is otherwise concealed in, say, Open AI's DALLE-E or Adobe Firefly. In fact, we know very little of how to navigate such more proprietary systems, and one might speculate if there even exists a complete map of their relations and dependencies.
Perhaps because of this lack of over-view and insight, maps and cartographies are not just shaping the reception and interpretation of generative AI, but can also be regarded as objects of interest and necessity in themselves – intrinsic parts of AI’s existence. Generative AI depends on an abundance of cartography to model, shape, navigate, and also negotiate and criticise its being in the world. There seems to be an inbuilt cultural need to 'map the territory', and the collection of cartographies and maps is therefore also what makes AI a reality – making AI real by externalising its abstraction onto a map, so to speak
A map of 'objects of interest and necessity' (autonomous AI image generation)
To enter the objects of autonomous AI image generation, a map that separates the territories of ‘pixel space’ from ‘latent space’ can be useful as a starting point – that is, a map that separates the objects you see from those that cannot be seen because they exist in a more abstract, computational space.
Latent space
Latent space consists of computational models and is a highly abstract space where it can be difficult to explain the behaviour of the different computational models. Very briefly put, they encode images with noise (using a Variational Autoencoder, VAE), and then learn how to de-code them back into images, also known as diffusion.
In the process of model training, datasets are central. Many of the datasets that are used to train models are made by 'scraping' the internet. Common Crawl, as an example, is a non-profit organisation that has built a repository of 250 billion web pages. Open Images and ImageNet are also commonly used as the backbone of visually training generative AI.
Contrary to common belief, there is not just one dataset used to make a model work, but multiple models and datasets to, for instance, reconstruct missing facial or other bodily details (such as too many fingers on one hand), 'upscale' images of low resolution or 'refine' the details in the image. Importantly, when it comes to autonomous AI image creation, there is typically an organisation and a community behind each dataset and training. LAION (Large-scale-Artificial Intelligence Open Network) is a good example of this. It too is a non-profit community organisation that develops and offers free models and datasets. Stable Diffusion was trained on datasets created by LAION.
For a model to work, it needs to be able to understand the semantic relationship between text and image. Say, how a tree is different from a lamp post, or a photo is different from a 18th century naturalist painting. The software CLIP (by OpenAI) is widely used for this, and also by LAION. CLIP is capable of predicting what images that can be paired with which text in a dataset. The annotation of images is central, here. That is, for the dataset to be useful there needs to be descriptions of what is on the images, what style they are in, their aesthetic qualities, and so on. Whereas ImageNet, for instance, crowdsources the annotation process, LAION uses Common Crawl to find html with <img>
tags, and then use the Alt Text to annotate the images (Alt Text is a descriptive text acts as a substitute for visual items on a page, and is sometimes included in the image data to increase accessibility). This is a highly cost-effective solution, which has enabled its community to produce and make publicly available a range of datasets and models that can be used in generative AI image creation.

Pixel space
In pixel space, you find a range of visible objects that a typical user would normally meet. This includes the interfaces for creating images. In conventional interfaces like DALL-E og Bing Image Creator, users prompt in order to generate images. What is particular for autonomous and decentralised AI image generation is that the interfaces have many more parameters and ways to interact with the models that generate the images. It functions more like an 'expert' interface.
In pixel space one finds many objects of visual culture. Apart from the interface itself, this includes both all the images generated by AI, and all the images used to train the models behind. These images are, as described above, used to create datasets, compiled by crawling the internet and scraping images that all belong to different visual cultures – ranging, e.g., from museum collections of paintings to criminal records with mug shots.
Many users also have specific aesthetic requirements to the images they want to generate. Say, to generate images in a particular manga style or setting. The expert interfaces therefore also contains the possibility to combine different models and even to post-train one's own models, also known as a LoRA (Low-Rank Adaptation). When sharing the images on platforms like Danbooru (on of the first and largest image boards for manga and anime) images are typically well categorised – both descriptively ('tight boots', 'open mouth', 'red earrings', etc.) and according to visual cultural style ('genshin impact', 'honkai', 'kancolle', etc.). Therefore they can also be used to train more models.
A useful annotated and categorised dataset - be it for a foundation model or a LoRA – typically involves specialised knowledge of both the technical requirements of model training (latent space) and the aesthetics and cultural values of visual culture itself. For instance of common visual conventions, such as realism, beauty, horror, and also (in the making of LoRAs) of more specialised conventions – say a visual style that an artist, wants to generate (see e.g. the generated images of Danish Hiphop by Kristoffer Ørum[2]).

An organisational plane
AI generated images as well as other objects of pixel space and latent space (like software, interfaces, datasets, or modesl involved in image generation) are not just products of a technical system, but also exist in a social realm, organised on platforms that are driven by the users themselves or commercial companies.
In mainstream visual culture the organisation is structured as a relation between a user and a corporate service. For example, users use Open AI's DALL-E to generate images, and may also share them afterwards on social media platfoms like Meta's Instagram. In this case, the social organisation is more or less controlled by the corporations who typically allow little interaction between their many users. For instance, DALL-E does not have a feature that allows one to build on or reuse the prompt of other users, or of users to share their experiences and insights with generative AI image creation. Social interaction between users only occurs when they share their images on platforms such as, say, Instagram. Rarely are users involved in the social, legal, technical or other conditions for making and sharing AI generated images.
Conversely, on the platforms for generating and sharing images in more autonomous AI involve users and communities are deeply involved in the conditions for AI image generation. LAION is a good example of this. It is run by a non-commercial organisation or 'team' of around 20 members, led by Christoph Shumann, but their many projects involve a wider community of AI specialists, professionals and researchers. They collaborate on principles of access and openness, and their attempt to 'democratise' AI stands in contrast to the policies of Big Tech AI corporations. In many ways, LAION resembles what the anthropologist Chris Kelty has also labelled a 'recursive publics' – a community that care for and self-maintain the means of its own existence.
However, such openness is not to be taken for granted, as also noted in debates around LAION.[4] There are many platforms in the ecology of autonomous AI ( see also CivitAI and Hugging Face) that easily become valuable resources. The datasets, models, communities, and expertise they offer may therefore also be subject to value extraction. Hugging Face is a prime example of this - a community hub as well as a $4.5 billion company with investments from Amazon, IBM, Google, Intel, and many more; as well as collaborations with Meta and Amazon Web Services. This indicates that in the organisation of autonomous AI there are dependencies on not only communities, but often also on corporate collaboration and venture capital.

A material plane (GPU infrastructure)
Just like the objects of autonomous AI depend on a social organisation (and also one of capital and labour), they also depend on a material infrastructure – and are, so to speak, always suspended between many different planes. First of all on hardware and specifically the GPUs that are needed to generate images as well as the models behind. Like in the social organisation of AI image generation, infrastructures too are organised differently.
The mainstream commercial services are set up as what one might call a 'client-server' relation. The users of DALL-E or similar service access a main server (or a 'stack' of servers). Users have little control of the conditions for generating models and images (say, the way models are reused or their climate impact) as this happens elsewhere, in 'the cloud'.
Where autonomous AI distinguishes itself from mainstream AI is typically the decentralisation organisation of processing power. Firstly, people who generate images or develop LoRAs with Stable Diffusion can use their own GPU. Often a simple laptop will work, but individuals and communities involved with autonomous AI image creation will often have expensive GPUs with high processing capability (built for gaming). Secondly, there is a decentralised network that connects the community's GPUs. That is, using the so-called Stable Horde (or AI Horde), the community can directly access each other's GPUs in a peer-to-peer manner. Granting others access to one's GPU is rewarded with currencies that in turn can be used to skip the line when waiting to access other members' GPUs. This social organisation of a material infrastructure allows the community to generate images almost with the same speed as commercial services.
To be dependent on the distribution of resources, rather than a centralised resource (e.g., a platform in 'the cloud'), points to how dependencies are often deliberately chosen in autonomous AI. One chooses to be dependent on a community because, for instance, one wants to reduce the consumption of hardware, because it is more cost-effective than one's own GPU, because one cannot afford the commercial services, or simply because one prefers this type of organisation of labour (separated from capital) that offers an alternative to Big Tech. That is, simply because one wants to be autonomous.
At this material plane, there are many other dependencies. For instance, energy consumption, the use of expensive minerals for producing hardware, or the exploitation of labour in the production of hardware.

Mapping the many different planes and dependencies of generative AI
What is particular about the map of this catalogue of objects of interest and necessity, is that it purely attempts to map autonomous and decentralised generative AI, serving as a map for a guided tour and experience of autonomous AI. However, both Hugging Face' dependency in venture capital and Stable Diffusion's dependency on hardware and infrastructure point to the fact that there are several planes that are not captured in the above map of this catalogue, but which are equally important. For instance, The EU AI Act or laws on copyright infringement, which Stable Diffusion (like any other AI ecology) will also depend on, point to a plane of governance and regulation. AI, including Stable Diffusion, also connects to the depends on the organisation of human labour, or the extraction of resources.
In describing the plane of objects of interest and necessity, we attempt to describe how Stable Diffusion and autonomous AI image generation build on dependencies to these different planes, but an overview of the many planes of AI and how it 'stacks' can of course also be the centre of a map in itself. One example of this is Kate Crawford's Atlas of AI, a book that displays different maps (and also images) that link AI to 'Earth' and the exploition of energy and minerals, or 'Labour' and the workers who do micro tasks ('clicking' tasks) or the workers in Amazon's warehouses. In continuation, Crawford's book also contains chapters on 'Data', 'Classification', 'Affect', 'State' and 'Power'.


Another abstraction of the layered nature of generative AI is found in Gertraud Koch's a map of all layers that she and her coauthors connects to "technological activity", and which would also pertain to AI.[5] On top of a layer of technology (the 'data models and algorithms') one will find other layers that are interdependent, and which contribute to the political and technological qualities of AI. As such, the map is also meant for navigation – to identify starting points for rethinking its concepts or reimagining alternative futures (in their work, particularly in relation to a potential delinking from a colonial past, and reimagining a pluriversality of technology)

Within the many planes and stacks of AI one can find many different maps that build other types of overviews and conceptual models of AI – perhaps pointing to how maps themselves take part in making AI a reality.
The corporate landscape
The entrepreneur, investor and pod cast host Matt Turck has made the “ultimate annual market map of the data/AI industry”. Since 2012 he has documented the corporate landscape of AI not just to identify key corporate actors, but also developments of trends in business. As he also notes in his blog, the first map from 2012 has merely 139 logos, whereas the 2024 version has 2,011 logos.[6] This reflects the massive investment in AI entrepreneurship, following first 'big data' and now 'generative AI' (and machine learning) - how AI has become a business reality. Comparing the 2012 version with the most recent map from 2024, one can see the corporate landscape of AI changes over time. How, for instance, the division of companies dealing with infrastructure, data analytics, applications, data sources, and open source AI becomes fine grained over the years, forking out into, applications in health, finance and agriculture; or how privacy and security become of increased concern. Clearly, AI reconfigures and intersects with many different realities.

Critical cartography in the mapping of AI
In mapping AI there are also 'counter maps' or 'critical cartography'.[8] Conventional world maps are built on set principles of, for instance, North facing up, and Europe at the centre. The map is therefore not just a map for navigation, but also a map of more abstract imaginaries and histories originating in colonial times, where maps was the outset of Europe and an intrinsic part of the conquest of territories. In this sense, a map always also reflects hierarchies of power and control that can be inverted or exposed (for instance by turning the map upside down, letting the south be a point of departure). Counter-mapping technological territories would, following this logic, involve what the French research and design group Bureau d´Études has called "maps of contemporary political, social and economic systems that allow people to inform, reposition and empower themselves."[9] They are maps that reveal underlying structures of social, political or economic dependencies to expose what ought to be of common interest, or the hidden grounds on which a commons rests. Félix Guattari and Gilles Deleuze' notion of 'deterritorialization' can be useful, here, as a way to conceptualise the practices that expose and mutate the social, material, financial, political, or other organisation of relations and dependencies.[10] The aim is ultimately not only to destroy this 'territory' of relations and dependencies, but ultimately a 'reterritorialization' – a reconfiguration of the relations and dependencies.
Utilising the opportunities of info-graphics in mapping can be a powerful tool. At the plane of financial dependencies, one can map, as Matt Turck, the corporate landscape of AI, but one can also draw a different map that reveals how the territory of 'startups' does not compare to a geographical map of land and continents. Strikingly, The United States is double the size of Europe and Asia, whereas there are whole countries and continents that are missing (such as Russia and Africa). This map thereby not only reflects the number of startups, but also how venture capital is dependent on other planes, such as politics and the organisation of capital, or infrastructural gaps. In Africa, for instance, the AI divide is very much also a 'digital divide', as argued by AI researcher Jean-Louis Fendji.[11]

Counter-mapping the organisation of relations and dependencies is also prevalent in the works of the Barcelona-based artist collective Estampa, which exposes how generative AI depends on different planes: venture capital, energy consumption, a supply chain of minerals, human labour, as well as other infrastructures, such as the internet, which is 'scraped' for images or other media, using e.g. software like Clip).

Epistemic mapping of AI
Maps of AI often also address how AI functions as what Celia Lury has called an 'epistemic infrastructure'.[14] That is, AI is an apparatus that builds on knowledge, creates knowledge, but also shapes what knowledge is and we consider to be knowledge. To Lury, the question of 'methods' here becomes central - not as a neutral, 'objective' stance, as one typically regards good methodology in science, but as a cultural and social practice that help articulate the questions we ask and what we consider to be a problem in the first place. When one for, instance, criticises the social, racial or other biases in generative AI (such as all doctors being white males in generative AI image creation), we are not just dealing with bias in the dataset that can be fixed with 'negative prompts' or other technical means. Rather, AI is fundamentally – in its very construction and infrastructure – based in a Eurocentric history of modernity and knowledge production. For instance, as pointed out by Rachel Adams, AI belongs to a genealogy of intelligence, and one also ought to ask, whose intelligence and understanding of knowledge is modelled within the technology – and whose is left out?[15]
There are several attempts to map this territory in the plane of knowledge production, and its many social, material, political or other relations and dependencies. Sharing many of the concerns of Lury and Adams, Vladan Joler and Matteo Pasquinelli's 'Nooscope' is a good example of this.[16] In their understanding AI belongs to a much longer history of knowledge instruments ('nooscopes', from the Greek skopein ‘to examine, look’ and noos ‘knowledge’) that would also include optical instruments, but which in AI is a form of knowledge magnification of patterns and statistical correlations in data. The nooscope map is an abstraction of how AI functions as "Instrument of Knowledge Extractivism". It is therefore not a map of 'intelligence' and logical reasoning, but rather of a "regime of visibility and intelligibility" whose aim is the automation of labour, and of how this aim rests on (as other capitalist extractions of value in modernity) a division of labour – between humans and technology, between for instance historical biases in the selection and labelling of data, and their formalisation in sensors, databases and metadata. The map also refers to how selection, labelling and other laborious tasks in the training of models is done by "ghost workers" thereby referring to a broader geo-politics and body-politics of AI where human labour is often done by subjects of the Global South (although they might oppose being referred to as 'ghosts').

[1] Alfred Korzybski, “A Non-Aristotelian System and its Necessity for Rigour in Mathematics and Physics,” in Science and Sanity: An Introduction to Non-Aristotelian Systems and General Semantics (Lancaster, PA: The Science Press Printing Company, 1933), 747–61.
[2] Kristoffer Ørum, “Project #253,” accessed August 11, 2025, https://oerum.org/pico/projects/253.
[3] Christopher M. Kelty, Two Bits: The Cultural Significance of Free Software (Durham, NC: Duke University Press, 2008), https://read.dukeupress.edu/books/book/1136/Two-BitsThe-Cultural-Significance-of-Free-Software.
[4] “The Story of LAION: The Dataset Behind Stable Diffusion,” The Batch, June 7, 2023, https://www.deeplearning.ai/the-batch/the-story-of-laion-the-dataset-behind-stable-diffusion/.
[5] Gertraud Koch et al., “Layers of Technology in Pluriversal Design: Decolonising Language Technology with the Live Language Initiative,” CoDesign 20, no. 1 (2024): 77–90, https://doi.org/10.1080/15710882.2024.2341799.
[6] Matt Turck, “Full Steam Ahead: The 2024 MAD (Machine Learning, AI & Data) Landscape,” MattTurck.com, March 2024, https://mattturck.com/MAD2024/.
[7] Matt Turck, “A Chart of the Big Data Ecosystem,” MattTurck.com, June 29, 2012, https://mattturck.com/a-chart-of-the-big-data-ecosystem/.
[8] Jeremy W. Crampton and John Krygier, “An Introduction to Critical Cartography,” ACME: An International Journal for Critical Geographies 4, no. 1 (2005): 11–33, https://doi.org/10.14288/acme.v4i1.723.
[9] Bureau d'Études, accessed August 11, 2025, https://bureaudetudes.org/.
[10] Gilles Deleuze and Félix Guattari, A Thousand Plateaus: Capitalism and Schizophrenia, trans. Brian Massumi (Minneapolis: University of Minnesota Press, 1987).
[11] “From Digital Divide to AI Divide – Fellows’ Seminar by Jean-Louis Fendji,” STIAS, April 9, 2024, https://stias.ac.za/2024/04/from-digital-divide-to-ai-divide-fellows-seminar-by-jean-louis-fendji/.
[12] Marcus Lu, “Mapped: The Number of AI Startups by Country,” Visual Capitalist, May 6, 2024, https://www.visualcapitalist.com/mapped-the-number-of-ai-startups-by-country/.
[13] “Cartography of Generative AI,” Estampa, accessed August 11, 2025, https://tallerestampa.com/en/estampa/cartography-of-generative-ai/.
[14] Celia Lury, Problem Spaces: How and Why Methodology Matters (Cambridge, UK; Medford, MA: Polity, 2021).
[15] Rachel Adams, “Can Artificial Intelligence Be Decolonized?,” Area 53, no. 1 (2021): 6–13, https://doi.org/10.1080/03080188.2020.1840225.
[16] Vladan Joler and Matteo Pasquinelli, The Nooscope Manifested: AI as Instrument of Knowledge Extractivism, 2020, https://fritz.ai/nooscope/.
++++++++++++++++++++++++++++++++++++++++++++++++++++++
[CARD TEXT – possibly needs shortening]
Mapping 'Objects of Interest and Necessity'
There is little knowledge of what AI really looks like. The maps presented here are an attempt to abstract the different objects that one may come across when entering the world of autonomous and decentralised AI image creation. It can serve as a useful guide to experience what the objects of this world are called, how they connect each other, to communities or underlying infrastructures – perhaps also as an outset for one's own exploration.
A distinction between 'pixel space' and 'latent space' can be helpful. That is, what you see from what you do not see.
Latent space refers to the invisible space that exists between the capture of images in datasets and the generation of new images. Images are encoded with 'noise', and the machine then learns how to how to de-code them back into images (aka 'image diffusion'). Contrary to common belief, there is not just one dataset used to make image generation work, but multiple models and datasets to 'upscale' images of low resolution, 'refine' the details in the image, and much more. Behind every model and dataset there is a community and organisation.
Pixel space is where one encounters objects of visual culture. Large-scale datasets are for instance compiled by crawling and scraping repositories of visual culture, such as museum collections. Whereas conventional interfaces for generating images only offer the possibility to 'prompt', interfaces to Stable Diffusion offer advanced parameters, as well as options to train one's own models, aka LoRAs. This demands technical insights into latent space as well as aesthetic/cultural understandings of visual culture (say, of manga, gaming or art).
Both images and LoRAs are organised and shared on dedicated platforms (e.g., Danbooru or CivitAI). The generation of images and use of GPU/hardware can also be distributed to a community of users in a peer-to-peer network (Stable Horde). This points to how models, software, datasets and other objects always also exist suspended between different planes of dependencies - organisational, material, or other.
[Images: 'Our' map(s) - perhaps surround by other maps]
Guestbook