Show

Recent years have seen a surge in the popularity of commercial AI products based on generative, multi-purpose AI systems promising a unified approach to building machine learning (ML) models into technology. However, this ambition of “generality” comes at a steep cost to the environment, given the amount of energy these systems require and the amount of carbon that they emit. In this work, we propose the first systematic comparison of the ongoing inference cost of various categories of ML systems, covering both task-specific (i.e. finetuned models that carry out a single task) and ‘general-purpose’ models, (i.e. those trained for multiple tasks). We measure deployment cost as the amount of energy and carbon required to perform 1,000 inferences on representative benchmark dataset using these models. We find that multi-purpose, generative architectures are orders of magnitude more expensive than task-specific systems for a variety of tasks, even when controlling for the number of model parameters. We conclude with a discussion around the current trend of deploying multi-purpose generative ML systems, and caution that their utility should be more intentionally weighed against increased costs in terms of energy and emissions. All the data from our study can be accessed via an interactive demo to carry out further exploration and analysis

Text-based tasks are, all things considered, more energy-efficient than image-based tasks, with image classification requiring less energy (median of 0.0068 kWh for 1,000 inferences) than image generation (1.35 kWh) and, conversely, text generation (0.042 KwH) requiring more than text classification (0.0023 kWh). For comparison, charging the average smartphone requires 0.012 kWh of energy 4 , which means that the most efficient text generation model uses as much energy as 16% of a full smartphone charge for 1,000 inferences, whereas the least efficient image generation model uses as much energy as 950 smartphone charges (11.49 kWh), or nearly 1 charge per image generation, although there is also a large variation between image generation models, depending on the size of image that they generate.

the most carbon-intensive image generation model (stable-diffusion-xl-base-1.0) generates 1,594 grams of 𝐶𝑂2 for 1,000 inferences, which is roughly the equivalent to 4.1 miles driven by an average gasoline-powered passenger vehicle 5 , whereas the least carbon-intensive text generation model (distilbert-base-uncased) generates as much carbon as 0.0006 miles driven by a similar vehicle, i.e. 6,833 times less. This can add up quickly when image generation models such as Dall·E and MidJourney are deployed in user-facing applications and used by millions of users globally

hentai-free

  • WoofWoof91 [comrade/them]
    ·
    7 months ago

    i wonder how it compares to playing a videogame or something

    i generate ~20-30 npc portraits per week ish, it tops out my graphics card and puts my cpu at about 75% for about 10 minutes per 9 images

    playing almost any reasonably recent aaa game does that for however long i play it for, usually at least two hours per session

    • JohnBrownsBussy2 [he/him]
      ·
      7 months ago

      The power burden is proportional to the inference time x GPU load, so the gaming is going to be worse, simply because you're running it longer.

      As an aside, that sounds like there might be an optimization issue if you're putting a heavy load on both GPU and CPU, since optimally you want to run everything on the GPU. I did some tests on my setup, and even when using 100% of the VRAM it's not impacting the CPU.

      • JohnBrownNote [comrade/them, des/pair]
        ·
        7 months ago

        some poorly made (especially multiplayer) games are cpu bottlenecked even if you have the good shit.

        my pc time is environmentally neutral(ish) in winter because the heat isn't wasted and it keeps me from using a space heater to keep living tolerable.