Building AI Fairness by Reducing Algorithmic Bias
Emily Diana explores algorithmic bias in machine learning and outlines three intervention stages: pre-processing, in-processing, and post-processing to mitigate algorithmic discrimination.
Generative AI’s operational needs consume massive amounts of energy and emit huge amounts of carbon into the atmosphere. To mitigate the environmental impact and lower costs, businesses must adopt sustainable practices like clean energy, hardware efficiency, and model optimization to continue innovation.
Generative artificial intelligence (genAI) can write code, draft marketing campaigns, generate images, and even mimic human reasoning—all without explicit training for these tasks. Yet, this performance comes at a price. Behind every AI-generated insight lies an immense computational engine, powered by data centers using electricity at an unprecedented scale. U.S. utility companies anticipate data center energy consumption demands by 2028 to equal the energy needed to power three New York cities. For business leaders, the challenge is clear: how do we harness AI’s potential without fueling a climate crisis?
A simple formula is at the heart of genAI’s environmental impact: compute ≈ 6 × parameters × tokens. Doubling a model’s parameters or training data doubles its energy needs. Training GPT-3 (with 175B parameters and 300B tokens), for instance, consumed 1,287 MWh of power—enough to fuel 108 U.S. homes for a year—and emitted 552 metric tons of CO₂. Models like Llama with 7B to 70B parameters and 2T tokens have similar power consumption and emissions, requiring 1227 MWh and emitting 539 metric tons of CO₂.
But these numbers only scratch the surface. Data centers don’t just run graphics processing units; they cool, light, and support them with auxiliary systems. The power usage effectiveness metric, averaging 1.58 in the U.S., reveals that nearly 40 percent of a data center’s energy is wasted on non-compute tasks. This inefficiency compounds as companies race to build bigger models and denser data centers.
The environmental cost of genAI hinges on two variables: where the energy comes from and how the model is used. A data center plugged into France’s nuclear-dominated grid emits just 0.02 kg of CO₂ per kWh. Contrast this with coal-heavy grids in parts of China, where the same kWh generates 0.9 kg of CO₂. Even within the U.S., regional disparities matter: The Pacific Northwest’s hydro-rich grid emits 0.15 kg/kWh, while the Midwest’s coal-reliant grid emits 0.55 kg/kWh.
Then comes the lifecycle problem. In LLMs, training builds the model’s intelligence by processing massive data, whereas inference is the real-time use of that intelligence to respond to user queries. Training a model like GPT-3 is just the beginning. Inference—the phase where users query the model—accounts for over 90 percent of its lifetime emissions. While generating an image (~0.03 kg CO₂) or summarizing a document (~0.02 kg CO₂) seems trivial, scale these actions to millions of daily users, and emissions quickly eclipse training costs.
Current carbon emissions calculations often ignore critical factors. Pre-training steps—like hyperparameter tuning, neural architecture searches, and data cleaning—can multiply a model’s carbon footprint by 20x to 100x. For instance, GPT-3’s estimated 552 metric tons of CO₂ excludes the energy spent on failed experiments, iterative tweaks, hyperparameter tuning, architectural searches, and data cleaning.
Moreover, AI’s environmental toll extends beyond a carbon footprint. Data centers consume billions of gallons of water annually for cooling, while rare earth minerals in hardware fuel resource extraction and e-waste. The full picture demands a broader lens.
Generative AI is reshaping industries, but its legacy will hinge on how we manage its environmental trade-offs. For business leaders, the opportunity lies in reframing sustainability as a driver of innovation—not a constraint.