Measuring AI/ML applications
GMT can also measure AI/ML workloads.
Example for ML workload
See our example ML example application to have a usage scenario for a simple Python ML workload to get started.
Example for GenAI Text LLM workload
The simplest way is to use ollama as a manager and encapsulate it inside of the GMT.
See our example ollama LLM example application to have a usage scenario to get started.
Quick LLM query measuring
Since LLM chat queries are so common GMT comes with a quick measurement function for that.
In the root folder you find the run-template.sh
file.
Measure a sample query like this: bash run-template.sh ai "How cool is the GMT?"
It will download ollama, setup the containers and download an example model (gemma3:1b). Once you got this quick measurement running iterate on it by using our our example ollama LLM example application.
Bonus tip: If you apply --quick
to the run-template.sh
call the measurement is quicker for debugging purposes. However results will be not as reliable. Use only for debugging!
Trying out our hosted service
We operate green-coding.ai as a simple demo vertical that uses the underlying Green Metrics Tool Cluster Hosted Service →.
Check it out if you do not feel like installing the GMT and just want to get carbon and energy info on single prompts.