AI With Lower Emissions

How GLYNT.AI’s purpose-built system uses 95% less energy and water than LLMs

GLYNT.AI is the Sustainable Way To Do Sustainability Data

Relative energy, water and emissions for the same tasks

Chat GPT was released just a few years ago and has captured our imaginations. Since then, Large Language Models (LLMs) have continued to advance, with huge algorithms, huge data sets and huge computing power needs. LLMs have triggered a global data center buildout, raising concerns about costs and GHG emissions.

In contrast, GLYNT.AI is different. Our purpose-built ‘Few Shot’ machine learning is highly accurate with just a few samples of data. We orchestrate libraries of tiny, tiny models. With very low computing needs, GLYNT.AI has less than 5% of the emissions of LLMs for the same tasks.

Carbon Emissions for LLMs By Task

(per 1000 inferences, excluding training and infrastructure)

A chart showing carbon emissions for LLMs by Task with GLYNT.AI at the inner bottom corner showing the least emissions of the options

Source: GLYNT.AI calculations; Hugging Face (2)

And We’re Constantly Improving

To compare our emissions to LLMs and standard cloud computing, the GLYNT.AI did a deep dive into the components of our cloud computing bills for the past 10 months. We used a standard methodology (see the citations below) to calculate our emissions. The GLYNT.AI infrastructure team has been optimizing our cloud usage to lower our costs, and this had an added benefit: lower emissions too.

A graph showing a downward trend of 25% reduction in carbon emissions for GLYNT.AI between January 2024 and October 2024

Reporting Data for Our Customers

A summary of GLYNT.AI’s resource footprint is provided below, normalized to the number of rows of data produced. Customers should report GLYNT.AI’s emissions as part of their Scope 3, based total number of rows of data received per year.

GLYNT.AI Environmental Footprint for Sustainability Data Preparation

(per 1000 rows of data prepared by GLYNT.AI)

RESOURCE
Energy
Water
Carbon Emissions
INTENSITY
0.1033
0.0000516
0.02187
UNIT OF MEASURE
kWh
M3
MT CO2e

In Sum

This detailed analysis has shown that GLYNT.AI is a highly efficient sustainability data service with a very low footprint for energy, water and emissions. GLYNT.AI plays a key role in the AI ecosystem; customers can use our AI to produce accurate, verified and enriched data for best performance by enterprise AI, e.g. LLMs. And our results have shown that GLYNT.AI customers are not making an AI vs sustainability tradeoff. GLYNT.AI provides industry-leading accurate, audit-ready sustainability data while being more sustainable than alternatives

The framework developed for this analysis is both detailed and expandable, so that the results can be updated each year. Comments and feedback on this report are appreciated. Please send them to us here.

DETAILS

The GLYNT.AI System and the Reporting Boundaries

GLYNT.AI began our sustainability footprint calculations by analyzing our cloud usage detail from January 1, 2024 through October 31, 2024. GLYNT.AI uses the AWS cloud service. Here is a summary diagram of the GLYNT.AI system:

The GLYNT.AI Sustainability Data Pipeline

A flowchart showing the stages of GLYNT.AI's AI-First Data Preparation System
The focus of this report is the box highlighted in yellow, the use of GLYNT.AI’s Few Shot ML. The other boxes illustrate the features and functionality needed by our customers for accurate, audit-ready sustainability data, e.g. finance-grade water, energy, waste and emissions data. For energy, water and emissions calculations other boxes are part of the GLYNT.AI infrastructure.

GLYNT.AI is AI-First, in that we started by using ML to support human-oriented tasks in a step-by-step workflow. But in early 2024 we flipped our model, and the entire workflow is optimized to deliver accurate data with efficiency. Human tasks are in support of ML optimization. This has increased our capacity significantly, and led to faster services – including onboarding – for customers.

Our cloud usage includes AWS tools for managing our service and the AWS Optical Character Recognition (OCR) engine. We use a rather basic set of features of the OCR, and add in our purpose-built machine learning and other code to obtain reliable accuracy and efficiency. GLYNT.AI can switch to the OCR engines provided by the other hyperscalers, Google and Microsoft, and testing shows that the three OCR engines perform at about the same level in our system. The data reported below includes our usage of AWS’ OCR but not the training of their service.

The boundaries of our study include data center usage for the four categories mentioned, but exclude two other areas of resource use: transfer of data across global telecom networks, and data center hardware. EPRI reports that global data transfers use the same total energy as data centers, approximately 220 MMWh per year. (12) Business users such as GLYNT.AI have no visibility into this task and its associated resource use. We hope to include it in future years.

Also excluded from the study is data center hardware. Data centers have numerous servers, which are replaced every two or three years. Within the server there are chips, which are key to progress in computing and energy efficiency. The embodied emissions in hardware and its replacement cycles are changing fast and there is no visibility for the typical business user. This too is excluded in our current reporting.

Organizing Cloud Usage Data

GLYNT.AI’s cloud usage comes from the three environments we run: development, stage and production. We combined stage and production for this analysis. On average, the development environment is 13% of total usage. Its usage data were removed from further analysis.

Next we categorized our usage into four groups: compute, storage, networking and other infrastructure as each has a different carbon intensity per unit of use. The Cloud Carbon Footprint (CCF) method for categorization of cloud data and bills was used. (3) This method has been used by other similar reporting endeavors, such as by Etsy and Climatiq.io (4) As noted by the CCF documentation, judgements are required, and when in doubt, we assign the usage item to the compute category, as this has the largest carbon content per unit of use. Throughout our work we have used similar conservative judgements as needed.

Third, we examined the two key areas of data center usage in AI: training and inference. AI can be thought of as an algorithm (e.g. a mathematical approach) and the output of the algorithm is trained models. The models are used again and again to meet user requests, which is known as inference. Based on data science analytics, user feedback and so on, the algorithm code base is updated so that the trained models perform better.

Early analysis of the environmental footprint of LLMs focused training, but recent work has noted that with use of LLM exploding, inference usage will dominate the environmental footprint. But the inference usage depends on a forecast of user adoption, the life span of the trained models before retraining, the pace of LLM development and more. Recently, researchers have taken another tack, looking at the energy burden of LLM inference usage by task. This excludes the training portion of energy use. We use the task-based approach here to normalize the data for an apples-to-apples comparison.

The GLYNT.AI implementation of Few Shot ML operates very differently than management of LLMs. GLYNT.AI’s Few Shot ML uses a library of tiny, tiny training models. Each model can be trained in minutes and needs just three training samples. The model returns data 98%+ accuracy rates from the start and so very little retraining is needed. This is in stark contrast to LLMs, which have enormous models trained on vast data sets, with frequent retraining.

Further, GLYNT.AI is built on a segmented database and ML model library so that we do not co-mingle customer data. When onboarding a new customer, GLYNT.AI creates a customer-specific library of ML models, there is no re-use across customers. Incredibly low cost training is a strength of our system and we exploit it for efficiency and additional security and privacy for our customers. Training and inference are handled on the same data center resources, and are nearly blended together from the start. Thus GLYNT.AI’s environmental footprint is for training and inference combined.

A recent paper from Hugging Face examined the energy and carbon footprint of various LLM algorithms. (2) Using a standard data set, the researchers compared the energy efficiency of various LLMs in inference mode. The researchers also organized the tests and results by task, such as text classification, text-from-image extraction and so on. A key objective of our work is to compare GLYNT.AI’s Few Shot ML to Hugging Face results, for the same tasks.

To obtain these results, we did a calculation thruline from choice of AI algorithm to energy and water usage as well as carbon emissions:

  • The AI algorithm uses four types of data center resources: compute, storage, networking and other
  • Switching between AI algorithms may impact any or all of these
  • Each data center resource has a usage metric, e.g. KB of storage or CPU-hour
  • Each usage metric has an electricity intensity, e.g. kWh per CPU-hour
  • Each data center location has a carbon emissions factor, e.g. kg of CO2e per kWh
  • Each electricity metric has an associated water metric, e.g. L of water per kWh

This chaining of values follows that of CCF, Etsy and Climatiq.io. See the Details section for more information. Given the tight linkages, the environmental resource metrics become interchangeable when describing relative usage.

GLYNT.AI’s Very Efficient Training & Inference

To take our analysis further, we looked at how AI is done, through training, inference and supporting infrastructure. Our focus is on the four tasks that are reported in Hugging Face’s analysis of LLMs that are also portion services delivered by GLYNT.AI. While GLYNT.AI is highly accurate (above 98%), LLMs are not very accurate for these tasks. Recent research shows a less than 50% accuracy rate, even when boosted by RAG. (5)

To operationalize this calculation several assumptions are made. First, we normalize by accuracy rate to enable an apples-to-apples comparison. This means that effectively LLMs need nearly twice as much training and inference per task to reach a comparable accuracy rate as GLYNT.AI. Second, we assumed that the LLM’s emissions footprint was split evenly between inference and training. Hugging Face reports a wide range of data snippets on this matter, ranging from 60% inference, 40% training at Google to 90% inference, 10% training at AWS. (5). GLYNT.AI’s Few Shot ML works quite differently, in a nearly blended training-inference mode and so we report only the the combined emissions intensity for inference plus training.

As the table shows GLYNT.AI’s Few Show ML emits less than 1% of the carbon that LLMs do per comparable transaction. If training is excluded from the LLM emissions, GLYNT.AI remains below 1% on a relative basis. If it is assumed that LLMs are perfectly accurate for these tasks, GLYNT.AI is still less than 1% of the LLM’s emissions.

When looking at the entire system, AI training, AI inference and supporting infrastructure, GLYNT.AI is 87% that of LLMs. We believe that this result is largely due to scale. Data centers and researchers from large companies are far more likely to be able to parse their infrastructure usage into that dedicated to AI and that used for other purposes. GLYNT.AI has one product, and all of our infrastructure computing goes into the delivery of our product. We suspect that as GLYNT.AI grows, we will be able to divide our infrastructure usage more precisely, but we also suspect that as a one product company with a highly efficient AI algorithm, we’ll always have a relatively high share of infrastructure usage.

Carbon Emissions (Kg CO2e) per Successfully Completed Transaction

ACTIVITY
AI Inference
AI Training + Inference
All AI + Infrastructure
GLYNT.AI*
0.056
21.87
LLM**
12.02
18.03
25.24
GLYNT.AI/LLM
0.31%
86.65%

* Per 1000 fields delivered to the customer (100% accuracy rate)
**Per 2000 inferences (50% accuracy rate assumed)

Source: GLYNT.AI calculations; Hugging Face (2).

The Hugging Face research explores the role of scale, complexity and other factors. We used their summary measurements by task and selected four tasks that sum to a comparable service as provided by GLYNT.AI: text classification; extractive QA; token classification; and image classification. We also assumed that LLMs have a 50% accuracy rate on these tasks, which means they will have twice as much training and inference per successfully completed transaction.

Hugging Face reports the energy use of inference only for the tasks selected. Inference energy use per transaction depends on a number of factors, including user adoption, lifecycle of the model, amount of retraining and so on. We assumed that the LLM’s footprint was split evenly between inference and training. GLYNT.AI’s Few Shot ML works quite differently, in a nearly blended training-inference mode. We cannot split our data center usage by AI task, and report only the combined inference plus training result.

As seen in the table above, GLYNT.AI’s Few Show ML emits less than 1% of the carbon that LLMs do per comparable transaction. If training is excluded from the LLM emissions, GLYNT.AI remains below 1% on a relative basis. If it is assumed that LLMs are perfectly accurate for these tasks, GLYNT.AI is still less than 1% of the LLM’s emissions.

When looking at the entire system, AI training, AI inference and supporting infrastructure, GLYNT.AI is 87% that of LLMs. We believe that this result is largely due to scale. Data centers and researchers from large companies are far more likely to be able to parse their infrastructure usage into that dedicated to AI and that used for other purposes. GLYNT.AI has one product, and all of our infrastructure computing goes into the delivery of our product. We suspect that as GLYNT.AI grows, we will be able to divide our infrastructure usage more precisely, but we also suspect that as a one product company with a highly efficient AI algorithm, we’ll always have a relatively high share of infrastructure usage.

Additional Comments on the Carbon Intensity of LLMs (e.g. the Hugging Face Graph)

As we reviewed the results of GLYNT.AI’s low carbon intensity, as seen on the Hugging Face graph shown above, we mentioned how the graph goes from simple tasks on the left to complex tasks on the right, illustrating the ecosystem of AI.

One key extension is the inclusion of contextual data that improves LLM performance. Retrieval Augmented Generation (RAG) is one technique applied to LLMs that provides this result (5). RAG feeds accurate, enriched data into the LLM algorithm, and reduces hallucinations. In the world of sustainability, data enriched with codes (join keys) from other business systems is particularly valuable for LLM performance. This includes general ledger codes from accounting, site codes from an ERP system, building attributes and status from leasing and financial systems and so on.

RAG feeds accurate, enriched data into the LLM algorithm, and reduces hallucinations. In the world of sustainability, data enriched with business metadata from other business systems is particularly valuable for LLM performance. This includes general ledger codes from accounting, site codes from an ERP system, building attributes and status from leasing and financial systems and so on.

The result is known as the contextual layer of data, and is also called a knowledge graph. Researchers have found that these relationships improve LLM accuracy even further, Exposing the cross-functional relationships in accurate, enriched sustainability data makes it an even more valuable tool for LLMs, including agentic AI
(6). As recent research from Microsoft explains, knowledge graphs provide a whole of dataset context.

GLYNT.AI has built in data customization and enrichment into our product offering. We place sustainability data in a business-friendly contextual layer, putting sustainability data at the center of business decisions and future enterprise AI.

Carbon Emissions in Context

To weave our way through the calculations, we put the selection of the AI algorithm into a larger data center and energy use context. While selecting an AI algorithm feels important, there are many other points of energy and water use in cloud computing. We built the graph below to articulate the larger picture. The graph also details the data item available at each point in the ecosystem, and its level of granularity. This helps to track the many data fragments reported on cloud computing, data centers and AI.

From Electricity Trading Region to AI Training Models

A flowchart showing stages of carbon emissions in an AI application

Starting at the left, the local utility provides electricity to the data center. The carbon emissions of the utility are determined by their trading region. (7) The data center is an electricity user and a business, such as GLYNT.AI, is a customer of the data center. The customer has an AI system and other uses. For example, perhaps the customer is also providing a full reporting dashboard to customers. The AI system is sub-divided into three components: Training, Inference and Infrastructure. Other researchers report that data center usage needed to support an AI system is typically 30–40% of the total usage. (8) To compare the choice of AI systems, one finds an alternate algorithm and system that is well-suited to performing the same task.

As mentioned earlier, GLYNT.AI’s data center usage is divided into usage by the AI system and usage for other reasons. We assigned our development environment to the Other Use Case and excluded it from this analysis.

GLYNT.AI’s infrastructure averaged 62% of total usage throughout 2024. This could arise for two reasons. First, GLYNT.AI is rapidly scaling, but is not at the same scale of usage as the large LLMs, and until we reach that scale our infrastructure component is naturally larger. Second, GLYNT.AI’s Few Shot ML could be simply just more efficient than LLMs, making the AI component smaller and the infrastructure component relatively larger.

Along the bottom of the graphic above, we note which data is available by section of the overall system. At the left, one sees that the carbon intensity of electricity is known once the data center location is known. The data center itself helps to determine the level of usage per algorithm, as the data center might be a highly efficient hyperscale center or a legacy small data center with much higher energy use per task. A typical business user simply reports the numbers, and is not aware nor has the scale of operations to influence this outcome.

Emissions Factors

Emissions factors are often a weak link in carbon calculations and this is true for the calculations presented here. The issues are:

  • Stale data. Data centers are rapidly changing, the use of AI is growing rapidly, and electricity markets are feeling the pressure. Fresh data on the carbon intensity of electricity by location is needed. But most studies on AI emissions use emissions factors released in 2017, 2020 and 2021. And the reported emissions factor is based on actual data from prior years, so the emissions factor is even more stale!
  • False precision. The emissions factor reported by Google for networking in a Las Vegas based data center is .0004054 kg CO2e per GB. (4) No doubt the emissions factor is carefully produced by Google based on data center averages of some kind, but it conveys scientific precision while in fact it is a coarse approximation.
  • Regional variation in the carbon content of electricity. The carbon intensity of electricity in the continental U.S. varies from a low of 497 pounds of CO2e per MWh in California to a high of 1479 pounds of CO2e per MWh in eastern Wisconsin. (7) TThis is a three-fold span in carbon intensity, but the location of data centers is often left out their emissions analysis. The impact of location dwarfs other factors and this omission is a significant.

For typical data center use, we suggest using the data sets available from Climatiq.io, as these are carefully assembled by location, consistently documented and updated as new data becomes available. (4)

We note that we included the data center location in our assignment of emissions factors.

Water

Data centers put pressure on local water resources, as data centers use water for cooling and use electricity, which in turn uses water for power generation. Rigorous studies of the environmental footprint of data centers, such as that by LBNL, draw their boundaries more broadly than those used in our work. The LBNL researchers find that takes an average 7.6 liters of water to generate 1 kWh of energy in the US, while an average data center uses 1.8 liters of water per kWh consumed. (9) Thus, in total, data centers use 9.4 liters of water per kWh. This is the total impact of data centers on local water supplies and is alarmingly high.

The LBNL researchers note that location matters for data center water use, with a nearly 100X range of water use per kWh across the U.S.. The same conclusion is noted in a recent interview with a data center facility management company, with stated variation from 0.01 L/kWh to 1.5 L/kWh or higher. (10)

GLYNT.AI’s energy use is far below that of LLMs, and thus in the chain-linked calculations of this study, its water use is lower as well. There is no data available on water use by data center location, so we use the LBNL average of 1.8 L per kWh in our analysis.

Next Steps

Ready to see what GLYNT can do for you and your company?
Reach out to our team!

More Like This

4 people sitting around a meeting table in a conference room

Finance-Grade Sustainability Data

Learn about finance-grade sustainability data and how it unlocks business opportunities

Read the Guide

A view of an oil refinery at night

Awesome Sustainability Data

A GLYNT.AI guide to the eight attributes of high-confidence sustainability data

Read the Guide

A view of skyscrapers seen from the ground, looking up towards their tops

The Sustainability Data Buyer’s Guide

Manual data is slowing teams down. See how top enterprises cut prep time, reduce costs, and increase accuracy — with finance-grade, audit-ready automation

Read the Guide

Contact GLYNT.AI