COLEMAN (Combinatorial VOlatiLE Multi-Armed BANdit) is a Multi-Armed Bandit (MAB) based approach designed to address the Test Case Prioritization in Continuous Integration (TCPCI) Environments in a cost-effective way.
We designed COLEMAN to be generic regarding the programming language in the system under test, and adaptive to different contexts and testers' guidelines.
Modeling a MAB-based approach for TCPCI problem gives us some advantages in relation to studies found in the literature, as follows:
- It learns how to incorporate the feedback from the application of the test cases thus incorporating diversity in the test suite prioritization;
- It uses a policy to deal with the Exploration vs Exploitation (EvE) dilemma, thus mitigating the problem of beginning without knowledge (learning) and adapting to changes in the execution environment, for instance, the fact that some test cases are added (new test cases) and removed (obsolete test cases) from one cycle to another (volatility of test cases);
- It is model-free. The technique is independent of the development environment and programming language, and does not require any analysis in the code level;
- It is more lightweight, that is, needs only the historical failure data to execute, and has higher performance.
In this way, this repository contains the COLEMAN's implementation. For more information about COLEMAN read Ref1 in References.
Furthermore, this repository contains the adaptation to deal with Highly-Configurable System (HCS) context through two strategies:
- Variant Test Set Strategy (VTS) that relies on the test set specific for each variant; and
- Whole Test Set Strategy (WST) that prioritizes the test set composed by the union of the test cases of all variants.
For more information about WTS and VTS read Ref2 in References.
On the other hand, we extended COLEMAN to consider the context surrouding in each CI Cycle. The extended version we named as CONSTANTINE (CONtextual teSt prioriTizAtion for coNTinuous INEgration).
CONSTANTINE can use any feature given a dataset, for instance:
- Test Case Duration (Duration): The time spent by a test case to execute;
- Number of Test Ran Methods (NumRan): The number of test methods executed during the test, considering that some test methods are not executed due to some previous test method(s) have been failed;
- Number of Test Failed Methods (NumErrors): The number of test methods which failed during the test.
- Test Case Age (TcAge): This feature measures how long the test case exists and is given by a number which is incremented for each new CI Cycle in that the test case is used;
- Test Case Change (ChangeType): Considers whether a test case changed. If a test case is changed from a commit to another, there is a high probability that the alteration was performed because some change in the software needs to be tested. If the test case was changed, we could detect and consider if the test case was renamed, or it added or removed some methods;
- Cyclomatic Complexity (McCabe): This feature considers the complexity of McCabe. High complexity can be related to a more elaborated test case;
- Test Size (SLOC): Typically, size of a test case refers to either the lines of code or the number of
assertions
in a test case. This feature is note correlated with coverage. For instance, if we have two tests t_1 and t_2 and both cover a method, but t_2 have more assertions than t_1, consequently, t_2 have higher chances to detect failures.
In order to use this version
, use any Contextual-MAB available, for instance, LinUCB and SWLinUCB.
- Citation
- Installation
- Datasets
- About the files input
- Using the tool
- Analysis of COLEMAN4HCS Performance
- References
- Contributors
If this tool contributes to a project which leads to a scientific publication, I would appreciate a citation.
@Article{pradolima2020TSE,
author = {Prado Lima, Jackson A. and Vergilio, Silvia R.},
journal = {IEEE Transactions on Software Engineering},
title = {A Multi-Armed Bandit Approach for Test Case Prioritization in Continuous Integration Environments},
year = {2020},
pages = {12},
doi = {10.1109/TSE.2020.2992428},
}
@article{pradolima2021EMSE,
author = {Prado Lima, Jackson A. and Mendon{\c{c}}a, Willian D. F. and Vergilio, Silvia R. and Assun{\c{c}}{\~a}o, Wesley K. G.},
journal = {Empirical Software Engineering},
title = {{Cost-effective learning-based strategies for test case prioritization in Continuous Integration of Highly-Configurable Software}},
year = {2021}
}
To use this tool, follow these steps:
- Clone the repository:
git clone git@github.com:jacksonpradolima/coleman4hcs.git
- Install pyenv for managing Python versions. Refer to the pyenv documentation for installation instructions specific to your operating system.
- Install the required Python version for your project using pyenv. Run the following command in the project's root directory:
pyenv install 3.11.4
- Create a virtual environment for the project using pyenv. Run the following command:
pyenv virtualenv 3.11.4 <env-name>
Replace with a name for your virtual environment.
- Activate the virtual environment:
pyenv activate <env-name>
- Install the required dependencies using pip:
pip install -r requirements.txt
- Copy the .env.example file to .env:
cp .env.example .env
- Edit the .env file and provide values for the following environment variables:
CONFIG_FILE
: The path to your configuration file (e.g., ./config.toml).
The datasets used in the examples (and much more datasets) are available at Harvard Dataverse Repository. You can create your own dataset using out GitLab CI - Torrent tool or our adapted version from TravisTorrent tool. Besides that, you can extract relevant information about each system using our tool named TCPI - Dataset - Utils.
COLEMAN considers two kind of csv files: features-engineered and data-variants. The second file, data-variants.csv, is used by the HCS, and it represents all results from all variants. The information is organized by commit and variant.
-
features-engineered.csv contains the following information:
- Id: unique numeric identifier of the test execution;
- Name: unique numeric identifier of the test case;
- BuildId: a value uniquely identifying the build;
- Duration: approximated runtime of the test case;
- LastRun: previous last execution of the test case as DateTime;
- Verdict: test verdict of this test execution (Failed: 1, Passed: 0).
-
data-variants.csv contains all information that features-engineered.csv has, and in addition the following information:
- Variant: variant name.
In this way, features-engineered.csv organize the information for a single system or variant, and data-variants.csv track the information for all variants used during the software life-cycle (for each commit).
During the COLEMAN's execution, we use data-variants to identify the variants used in a current commit and apply the WTS strategy.
For CONSTANTINE, additional columns can be used and represents a contextual information. In this way, you define what kind of information can be used!
To use COLEMAN, you need to provide the necessary configurations. This includes setting up environment variables and configuration files.
Configure the utility by editing the config.toml
file located in the project's root directory.
The file contains various sections for configuring Execution, Experiment, and Algorithms.
Modify the values as per your project's requirements. Here's an example of the config.toml file:
[execution]
# Execution Configutation
parallel_pool_size = 10
independent_executions = 30
[experiment]
# Experiment Configuration
scheduled_time_ratio = [0.1, 0.5, 0.8]
datasets_dir = "example"
#datasets = ["fakedata"]
datasets = ["square@retrofit"]
# WTS Example
#datasets_dir = "examples/core@dune-common"
#datasets = ["dune@total"]
# VTS Example
#datasets_dir = "examples/core@dune-common"
#datasets = ["dune@debian_10 clang-7-libcpp-17", "dune@debian_11 gcc-10-20", "dune@ubuntu_20_04 clang-10-20"]
experiment_dir = "results/experiments/"
rewards = ["RNFail", "TimeRank"]
policies = ['Random', 'Greedy', 'EpsilonGreedy', 'UCB', 'FRRMAB', 'LinUCB', 'SWLinUCB']
[algorithm.frrmab]
# Algorithm Configuration for FRRMAB
window_sizes = [100]
timerank = { c = 0.5 }
rnfail = { c = 0.3 }
[algorithm.ucb]
# Algorithm Configuration for UCB
timerank = { c = 0.5 }
rnfail = { c = 0.3 }
[algorithm.epsilongreedy]
# Algorithm Configuration for Epsilon-Greedy
timerank = { epsilon = 0.5 }
rnfail = { epsilon = 0.3 }
[algorithm.linucb]
# Algorithm Configuration for LinUCB
timerank = { alpha = 0.5 }
rnfail = { alpha = 0.5 }
[algorithm.swlinucb]
# Algorithm Configuration for SWLinUCB
window_sizes = [100]
timerank = { alpha = 0.5 }
rnfail = { alpha = 0.5 }
[hcs_configuration]
# HCS Configuration
wts_strategy = false
[contextual_information.config]
previous_build = ['Duration', 'NumRan', 'NumErrors']
[contextual_information.feature_group]
feature_group_name = "time_execution"
feature_group_values = ['Duration', 'NumErrors']
where:
- Execution Configuration:
parallel_pool_size
is the number of threads to run COLEMAN in parallel.independent_executions
is the number of independent experiments we desire to run.
- Experiment Configuration:
scheduled_time_ratio
represents the Schedule Time Ratio, that is, time constraints that represents the time available to run the tests. Default: 0.1 (10%), 0.5 (50%), and 0.8 (80%) of the time available.datasets_dir
is the directory that contains your system. For instance, we desire to run the algorithm for the systems that are inside the directory data.datasets
is an array that represents the datasets to analyse. It's the folder name insidedatasets_dir
which contains the required file inputs.experiment_dir
is the directory where we will save the results.rewards
defines the reward functions to be used, available RNFailReward and TimeRankReward (See Ref1 in References).policies
selects the Policies available on COLEMAN, such as Random, Greedy, EpsilonGreedy, UCB, and FRRMAB.
- Algorithm Configuration: each algorithm has its own individual configuration. Next, we present some of them:
- FRRAB:
window_sizes
is an array that contains the sliding window sizesc
is the scaling factor. It's defined for each reward function used.
- UCB:
c
is the scaling factor. It's defined for each reward function used.
- Epsilon-Greedy:
epsilon
is the epsilon value. It's defined for each reward function used.
- FRRAB:
- HCS Configuration:
wts_strategy
represents the usage of Whole Test Set (WTS) Strategy for a system HCS (See Whole Test Set Strategy).
- Contextual Information:
- Config
previous_build
what kind of information we obtain from previous build and not in the current one. For instance, the test case duration (Duration
) will know only after the test execution.
- Feature Group
feature_group_name
represent the name of a feature group. We can create different groups of feature to evaluate the influence of each one.feature_group_values
represent the features selected to be used by the Contextual MAB.
- Config
The following MAB Policies are available on COLEMAN:
Policy | Description | Use Case |
---|---|---|
Random | Selects actions purely based on random choice without any preference. | Used as a baseline to compare with other more sophisticated policies. Doesn't use prior knowledge. |
Greedy | Always selects the action with the highest estimated reward. Exploits current knowledge without exploring other options. | Effective in environments where reward probabilities are static and unchanging. |
EpsilonGreedy | This is a variation of the Greedy policy. With probability ( \epsilon ), a random action is selected (exploration), and with probability ( 1 - \epsilon ), the action with the highest estimated reward is selected (exploitation). | Useful in environments with uncertainty about the optimal action or when the environment changes over time. |
UCB | Selects actions based on both their estimated rewards and the uncertainty around these rewards. Balances exploration and exploitation. | Effective when trials are limited and there's a need to explore and exploit simultaneously. |
SlMAB | Sliding-window Multi-Armed Bandit (SlMAB) approach, which uses a fixed-size sliding window to consider only recent feedback. The window size determines the trade-off between short-term and long-term rewards. | Suitable for non-stationary environments where the optimal action can change over time and recent actions are more indicative. |
FRRMAB | A more sophisticated MAB policy that considers feedback from previous actions using a sliding window to adjust estimated rewards for future decisions. | Beneficial in dynamic environments where actions' reward probabilities change and feedback from previous actions is valuable. |
and the following Contextual MABs for CONSTANTINE:
Policy | Description | Use Case |
---|---|---|
LinUCB | Incorporates linear regression with UCB. Uses contextual information of the arms to estimate the reward. The uncertainty bonus is combined with a linear estimate of the expected reward. | Best for environments with contextual information where the relationship between context and reward can be linearly approximated. |
SWLinUCB | A combination of LinUCB and sliding window approach. It uses the recent contextual information of the arms within a sliding window to better estimate the reward in non-stationary environments. | Ideal for non-stationary environments with contextual information where recent context-reward relationships are more indicative. |
Note: you can execute all the MAB and Contextual MAB policies together if you have the proper data input.
To execute COLEMAN for a non-HCS system, first update the variables in the provided TOML file:
datasets_dir = "examples"
datasets = ["fakedata"]
Subsequently, you can run the program with the following command:
python main.py
For HCS systems, we provide two distinct strategies to determine optimal solutions for variants: WTS (Whole Test Set Strategy) and VTS (Variant Test Set Strategy). You can learn more about these in Ref2 under the References section.
When employing the WTS and VTS strategies, regard datasets_dir
as the directory housing your system.
For the WTS approach, variants of a system are discerned from subfolders within the datasets_dir
directory.
Essentially, datasets_dir
symbolizes the project name.
This differentiation in execution methodology between HCS and non-HCS systems is crucial,
alongside the wts_strategy
variable. For clarity, please inspect our example directory.
The WTS strategy prioritizes the test set composed by the union of the test cases of all variants. To employ this strategy, modify the TOML file as follows:
wts_strategy = True
;
For a practical demonstration, set datasets = ["dune@total"]
(a dataset amalgamating test cases from all variants)
and datasets_dir = "examples/core@dune-common"
.
This provides a concise example using the Dune dataset.
More details on the dataset are available under Datasets.
Contrastingly, the VTS approach evaluates each variant as an isolated system. To harness this strategy, adjust the TOML file accordingly:
wts_strategy = False
;
As example, use datasets = ["dune@debian_10 clang-7-libcpp-17", "dune@debian_11 gcc-10-20", "dune@ubuntu_20_04 clang-10-20"]
and datasets_dir = "examples/core@dune-common"
to run one small example using Dune dataset (See Datasets).
Now, we consider each variant as single system.
As a hands-on example, set datasets = ["dune@debian_10 clang-7-libcpp-17", "dune@debian_11 gcc-10-20", "dune@ubuntu_20_04 clang-10-20"]
and datasets_dir = "examples/core@dune-common"
.
This offers a succinct example using the Dune dataset, treating each variant as a unique system.
Further insights into the dataset are available in the Datasets section.
As part of our ongoing effort to provide the state-of-the-art tool, Coleman4HCS, for TCPCI optimize COLEMAN4HCS, we've created examples to guide any researcher to understand the performance, effectiveness, and adaptability of our tool. The analysis, available in our Jupyter notebook (analysis.ipynb
), leverages various libraries such as DuckDB, Pandas, Seaborn, and Matplotlib to process data and visualize the results.
The notebook has examples including but not limited to test case execution times, prioritization effectiveness, and algorithm efficiency under different configurations and environments.
The notebook employs SQL queries for data manipulation and leverages Python's data analysis and visualization libraries to derive meaningful insights from historical test data. Our methodology ensures a robust analysis framework capable of handling large datasets and producing actionable intelligence.
Data visualizations play a key role in our analysis, offering intuitive understanding of complex data patterns and algorithm performance. The notebook includes various charts and graphs that elucidate the trade-offs between different prioritization strategies and their impact on test cycle times and failure detection rates.
- π [Ref1] A Multi-Armed Bandit Approach for Test Case Prioritization in Continuous Integration Environments published at IEEE Transactions on Software Engineering (TSE)
- π [Ref2] Learning-based prioritization of test cases in continuous integration of highly-configurable software published at Proceedings of the 24th ACM Conference on Systems and Software Product Line (SPLC'20)
- π¨βπ» Jackson Antonio do Prado Lima π§