Comparative Evaluation of Energy Efficiency in Large Language Models: Analyzing Improvements Across Incremental Versions in Inference Tasks
This experiment is a group project for the GreenLab course, under the Computer Science Master's programme at VU Amsterdam.
The versions that are going to be tested are incrementally as follows:
The versions that are going to be tested are incrementally as follows:
These versions are all instruct versions of the Gemma model. This will not affect our study because we are not drawing any comparative conclusions on the model performance between the LLM candidates.
The versions that are going to be tested are incrementally as follows:
These versions are all instruct versions of the open-source Mistral model model. Just as for Gemma, this will not affect our study because we are not drawing any comparative conclusions on the model performance between the LLM candidates.
We automated the experiment using the following framework: Experiment-Runner
- Energy Consumption (CPU/GPU):
- PowerJoular :
- CPU Consumption (Joules)
- GPU Consumption (Joules)
- PowerJoular :
- Resource Utilization:
- top : CPU Utilization (%), Memory Utilization (Bytes/%)
- nvidia-smi : GPU Utilization (%), GPU VMemory (Bytes/%)
- Model Performance:
git clone --recursive https://github.com/andrei-calin-dragomir/greenlab-course-project.git
cd ./greenlab-course-project
python3 -m venv venv
source ./venv/bin/activate
pip install -r requirements.txt
cd ./experiment-runner
pip install -r requirements.txt
python3 -m venv venv
cd ./experiment-runner
python experiment-runner/ ../RunnerConfig.py
The workflow of the experiment is defined as:
- BEFORE_EXPERIMENT
- BEFORE_RUN
- START_RUN
- START_MEASUREMENT
- INTERACT
- STOP_MEASUREMENT
- STOP_RUN
- POPULATE_RUN_DATA
- AFTER_EXPERIMENT