Maps: Difference between revisions
Line 15: | Line 15: | ||
A map of this territory would therefore necessarily contain technical objects (such as models, software, and platforms) that are deeply intertwined with communities that care for, negotiate, and maintain the means of their own existence (what the anthropologist Chris Kelty has also labelled a '[https://www.dukeupress.edu/two-bits recursive publics]'), but which may potentially also (at the same time) be subject to value extraction. [[Hugging Face]] is a prime example of this - a community hub as well as a $4.5 billion company with investments from Amazon, IBM, Google, Intel, and many more; as well as collaborations with Meta and Amazon Web Services. This indicates that there are other dependencies on not only communities, but also on corporate collaboration and venture capital. | A map of this territory would therefore necessarily contain technical objects (such as models, software, and platforms) that are deeply intertwined with communities that care for, negotiate, and maintain the means of their own existence (what the anthropologist Chris Kelty has also labelled a '[https://www.dukeupress.edu/two-bits recursive publics]'), but which may potentially also (at the same time) be subject to value extraction. [[Hugging Face]] is a prime example of this - a community hub as well as a $4.5 billion company with investments from Amazon, IBM, Google, Intel, and many more; as well as collaborations with Meta and Amazon Web Services. This indicates that there are other dependencies on not only communities, but also on corporate collaboration and venture capital. | ||
AI image generation is (and not least) always also dependent on material infrastructures. First of all on hardware and specifically [[GPU]]<nowiki/>s that are needed to both generate images as well as develop and refine the diffusion models (e.g. developing LoRAs). GPUs can be found in personal computers, but very often individuals and communities will invest in expensive GPUs with high processing capability. Subsequently, people who generate images or develop LoRAs with Stable Diffusion can have their own GPU (built into their computer or specifically acquired), but they can also benefit from a distributed network, allowing them to access other people’s GPUs, using the so-called [[Stable Horde]]. This points to how autonomous AI image generation not only depends on, but often also ''chooses'' dependencies. For example, to be dependent on the distributed resources of a community, rather than a centralised resource (e.g., a platform in 'the cloud'). At this material plane, there are also other dependencies that one can choose. For instance, energy. Both generating images and tracing models require a massive energy consumption, and the question of how much and from which source (green or black) may also be a key factor in choosing Stable Horde over one's own GPU, or a platform. The labour and minerals that go into the production of required hardware can be considered another dependency on this material plane. | |||
=== Mapping other planes of AI === | === Mapping other planes of AI === | ||
What is particular about the map of this catalogue of objects of interest and necessity, is that it purely maps autonomous and decentralised generative AI. Both Hugging Face' dependency in venture capital and Stable Diffusion's dependency on hardware and infrastructure point to the fact that there are several planes that are not captured in the map of | What is particular about the map of this catalogue of objects of interest and necessity, is that it purely maps autonomous and decentralised generative AI. Both Hugging Face' dependency in venture capital and Stable Diffusion's dependency on hardware and infrastructure point to the fact that there are several planes that are not captured in the above map of this catalogue, but which are equally important – in that they point to a number of other dependencies in other territories that can be explored and mapped. One might also add, a plane of governance and regulation. For instance, The EU AI Act or laws on copyright infringement, which Stable Diffusion (like any other AI ecology) will also depend on. The anthropologists and AI researcher Gertraud Koch has drawn a map (or diagram) of all the planes that she identifies in generative AI. | ||
[Gertraud's map + further commentary] | [Gertraud's map + further commentary] |
Revision as of 09:45, 3 July 2025
Mapping objects of interest and necessity
If one considers generative AI as an object, there is also a world of ‘para objects’ (surrounding AI and shaping its reception and interpretation) in the form of maps or diagrams of AI. // SMTH ON TH NEED TO DRAW 'A LANDSCAPE' OF AI - etc. examples of metaphors // They are drawn by both amateurs and professionals who need to represent processes that are otherwise sealed off in technical systems, but more generally reflect a need for abstraction – a need for conceptual models of how generative AI functions. However, as Alfred Korzybski famously put it, one should not confuse the map with the territory: the map is not how reality is, but a representation of reality [reference].
Following on from this, mapping the objects of interest in autonomous AI image creation is not to be understood as a map of what it 'really is'. Rather, it is a map of encounters of objects; encounters that can be documented and catalogued, but also positioned in a spatial dimension – representing a 'guided tour', and an experience of what objects are called, how they look, how they connect to other objects, communities or underlying infrastructures (see also Objects of interest and necessity). Perhaps, the map can even be used by others to navigate autonomous generative AI and create their own experiences. But, importantly to note, what is particular about the map of this catalogue of objects of interest and necessity, is that it purely maps autonomous and decentralised generative AI. It is therefore not to be considered a plane that necessarily reflects what happens in, for instance, Open AI og Google's generative AI systems. The map is in other words not a map of what is otherwise concealed in generative AI. In fact, we know very little of how to navigate proprietary systems, and one might speculate if there even exists a map of relations and dependencies.
A map of autonomous and decentralised AI image generation
To enter the world of autonomous AI image generation a map of how the territories of ‘pixel space’ from ‘latent space’ – the objects you see, from those who cannot be seen – is fundamental. In a pixel space, you would find both the images that are generated by a prompt (a textual input), as well as the many images that are used to compile the textually annotated data set used for training the image generation models (the foundation models).
Latent space is more complicated to explain. It is, in a sense, a purely computational space that relies on models that can encode images with noise (using a 'Variational Autoencoder', VAE), and learn how to de-code them back into images, thereby giving the model the ability to generate new images using image diffusion. It therefore also deeply depends on both the prompt and the dataset in pixel space – to generate images, and to build or train models for image generation.

Apart from pixel space and latent space, there is also a territory of objects that can be seen, but you typically (as a user) do not. For instance, in Stable Diffusion you find LAION, a non-profit organization that uses the software Clip to scrape the internet for textually annotated images to generate a free and open-source data set for training models in latent space. You would also find communities who contribute to LAION, or who refine the models of latent space using so-called LoRAs, but also models and datasets to, for instance, reconstruct missing facial or other bodily details (such as too many fingers on one hand) – often with both specialised knowledge of the properties of the foundational diffusion models, and of the visual culture they feed into (for instance manga or gaming). These communities are also organized on different platforms, such as CivitAI or Hugging Face, where communities can exhibit their specialised image creations or share their LoRAs, often with the involvement of different tokens or virtual currencies. What is important to realise when navigating this territory, is that behind every dataset, model, software, and platform lies also a specific community.

A map of this territory would therefore necessarily contain technical objects (such as models, software, and platforms) that are deeply intertwined with communities that care for, negotiate, and maintain the means of their own existence (what the anthropologist Chris Kelty has also labelled a 'recursive publics'), but which may potentially also (at the same time) be subject to value extraction. Hugging Face is a prime example of this - a community hub as well as a $4.5 billion company with investments from Amazon, IBM, Google, Intel, and many more; as well as collaborations with Meta and Amazon Web Services. This indicates that there are other dependencies on not only communities, but also on corporate collaboration and venture capital.
AI image generation is (and not least) always also dependent on material infrastructures. First of all on hardware and specifically GPUs that are needed to both generate images as well as develop and refine the diffusion models (e.g. developing LoRAs). GPUs can be found in personal computers, but very often individuals and communities will invest in expensive GPUs with high processing capability. Subsequently, people who generate images or develop LoRAs with Stable Diffusion can have their own GPU (built into their computer or specifically acquired), but they can also benefit from a distributed network, allowing them to access other people’s GPUs, using the so-called Stable Horde. This points to how autonomous AI image generation not only depends on, but often also chooses dependencies. For example, to be dependent on the distributed resources of a community, rather than a centralised resource (e.g., a platform in 'the cloud'). At this material plane, there are also other dependencies that one can choose. For instance, energy. Both generating images and tracing models require a massive energy consumption, and the question of how much and from which source (green or black) may also be a key factor in choosing Stable Horde over one's own GPU, or a platform. The labour and minerals that go into the production of required hardware can be considered another dependency on this material plane.
Mapping other planes of AI
What is particular about the map of this catalogue of objects of interest and necessity, is that it purely maps autonomous and decentralised generative AI. Both Hugging Face' dependency in venture capital and Stable Diffusion's dependency on hardware and infrastructure point to the fact that there are several planes that are not captured in the above map of this catalogue, but which are equally important – in that they point to a number of other dependencies in other territories that can be explored and mapped. One might also add, a plane of governance and regulation. For instance, The EU AI Act or laws on copyright infringement, which Stable Diffusion (like any other AI ecology) will also depend on. The anthropologists and AI researcher Gertraud Koch has drawn a map (or diagram) of all the planes that she identifies in generative AI.
[Gertraud's map + further commentary]
There are, as such, many other maps that each in their own way build conceptual models of generative AI. In this sense, maps and cartographies are not just shaping the reception and interpretation of genrative AI, but can also be regarded as objects of interest in themselves and intrinsic parts of AI’s existence: generative AI depends on cartography to build, shape, navigate, and also negotiate and criticise its being in the world. The collection of cartopgraphies and maps is therefore also what makes AI a reality.
There are for instance maps of the corporate plane of AI, such as entrepreneur, investor and pod cast host Matt Turck’s “ultimate annual market map of the data/AI industry”. Since 2012 Matt Turck has documented the ecosystem of AI not just to identify key corporate actors, but also developments of trends in business. Comparing the 2012 version with the most recent map from 2024, one can see how the division of companies dealing with infrastructure, data analytics, applications, data sources, and open source becomes fine grained over the years, and forking out into, for instance, applications in health, finance and agriculture; or how privcy and security become of increased concern in the business of infrastructures.

Maps and hierarchies

"counter maps"

Conceptual mapping / maps of maps - epistemic maps
Atlas of AI
Joler and Pasquinelli
Other maps…. And a map of the process of producing a pony image
