Dataset: Difference between revisions

From CTPwiki

No edit summary
Line 1: Line 1:
A dataset is a collection of data used to train AI models. In AI image generation, the dataset consists at a minimum of a collection of image - text pairs. Iconic datasets include LAION aesthetic dataset, ...
In the context of AI image generation, a dataset is a collection of a collection of image - text pairs (and sometimes other attributes such as provenance or an aesthetic score) used to train AI models. Iconic datasets include the LAION aesthetic dataset, Artemis, ImageNet, or Common Objects in Context (COCO). These collections of images, mostly sourced from the internet, reach dizzying scales. ImageNet became famous for its 14 millions images in the first decade of the century. Today LAION-5B consists of  5,85 billion CLIP-filtered image-text pairs.
 
== What is the network that sustains this object? ==
* How does it move from person to person, person to software, to platform, what things are attached to it (visual culture)
* Networks of attachments
* How does it relate / sustain a collective? (human + non-human)
 
== How does it evolve through time? ==
 
== How does it create value? Or decrease / affect value? ==
 
== What is its place/role in techno cultural strategies? ==
 
== How does it relate to autonomous infrastructure? ==


If large models such as Stable Diffusion require large scale datasets, various components such as LoRAs, VAEs, refiners, or upscalers can be trained with much more modest amount of data.


[[Category:Objects of Interest and Necessity]]
[[Category:Objects of Interest and Necessity]]

Revision as of 12:01, 20 August 2025

In the context of AI image generation, a dataset is a collection of a collection of image - text pairs (and sometimes other attributes such as provenance or an aesthetic score) used to train AI models. Iconic datasets include the LAION aesthetic dataset, Artemis, ImageNet, or Common Objects in Context (COCO). These collections of images, mostly sourced from the internet, reach dizzying scales. ImageNet became famous for its 14 millions images in the first decade of the century. Today LAION-5B consists of 5,85 billion CLIP-filtered image-text pairs.

If large models such as Stable Diffusion require large scale datasets, various components such as LoRAs, VAEs, refiners, or upscalers can be trained with much more modest amount of data.