title |
---|
Configuration File |
This reference page provides detailed information about the configurable options for ClearML and ClearML Agent.
ClearML and ClearML Agent use the same configuration file clearml.conf
.
This reference page is organized by configuration file section:
- agent - Contains ClearML Agent configuration options. If ClearML Agent was not installed, the configuration
file will not have an
agent
section. - api - Contains ClearML and ClearML Agent configuration options for ClearML Server.
- sdk - Contains ClearML and ClearML Agent configuration options for ClearML Python Package and ClearML Server.
- environment - Define environment variables to apply to the OS environment
- files - Define auto-generated files to apply into local file system
See an example configuration file in the ClearML Agent GitHub repository.
:::info The values in the ClearML configuration file can be overridden by environment variables, the configuration vault, and command-line arguments. :::
To add, change, or delete options, edit your configuration file.
To edit your ClearML configuration file:
-
Open the configuration file for editing, depending upon your operating system:
- Linux -
~/clearml.conf
- Mac -
$HOME/clearml.conf
- Windows -
\User\<username>\clearml.conf
- Linux -
-
In the required section (sections listed on this page), add, modify, or remove required options.
-
Save configuration file.
ClearML's configuration file uses HOCON configuration format, which supports environment variable reference.
For example:
sdk {
google.storage {
# # Default project and credentials file
# # Will be used when no bucket configuration is found
project: "clearml"
credentials_json: ${GOOGLE_APPLICATION_CREDENTIALS}
}
}
${GOOGLE_APPLICATION_CREDENTIALS}
will automatically be substituted by the environment variable value.
See Note on Windows for information about using environment variables with Windows in the configuration file.
agent
(dict)
- Dictionary of top-level ClearML Agent options to configure ClearML Agent for Git credentials, package managers, cache management, workers, and Docker for workers.
agent.crash_on_exception
(bool)
- By default, when encountering an exception while running a task, the agent will catch the exception, log it, and
continue running. When set to
true
, the agent crashes when encountering an exception.
agent.cuda_version
(float)
-
The CUDA version to use.
- If specified, this is the CUDA version used.
- If not specified, the CUDA version is automatically detected.
Alternatively, override this option with the environment variable
CUDA_VERSION
.
agent.cudnn_version
(float)
-
The cuDNN version to use.
- If specified, this is the cuDNN version used.
- If not specified, the cuDNN version is automatically detected.
Alternatively, override this option with the environment variable
CUDNN_VERSION
.
agent.disable_ssh_mount
(bool)
- Set to
true
to disable the auto.ssh
mount into the docker. The environment variableCLEARML_AGENT_DISABLE_SSH_MOUNT
overrides this configuration option.
agent.disable_task_docker_override
(bool)
- If set to
true
, agent uses the default docker image and ignores any docker image and arguments specified in the task's container section (if setup shell script is specified in task container section, it is used in either case).
agent.docker_allow_host_environ
(bool)
- Set to
true
to allow passing host environments into docker container with Task's docker container arguments. For example:"-e HOST_NAME=$HOST_NAME"
.
:::warning Use with care! This might introduce security risks by allowing access to keys/secret on the host machine. :::
agent.docker_apt_cache
(string)
- The apt (Linux package tool) cache folder for mapping Ubuntu package caching into Docker.
agent.docker_args_extra_precedes_task
(bool)
- Allow the arguments specified in
agent.extra_docker_arguments
to override task level docker arguments, in the case that the same argument is passed in both. If set toFalse
, a task's docker arguments will override theextra_docker_arguments
.
agent.docker_container_name_format
(string)
:::note Compatibility Required Compatible with Docker versions 0.6.5 and above :::
-
Set a name format for Docker containers created by an agent running in Docker mode
-
The following variables can be used:
task_id
worker_id
rand_string
- random lower-case letters string (up to 32 characters)
-
The resulting name must start with an alphanumeric character, while the rest of the name may contain alphanumeric characters, underscores (
_
), dots (.
) and/or dashes (-
) -
For example:
clearml-id-{task_id}-{rand_string:.8}
agent.docker_force_pull
(bool)
-
Always update the Docker image by forcing a Docker
pull
before running an experimentThe values are:
true
- Always update the Docker image.false
- Do not always update.
agent.docker_install_opencv_libs
(bool)
- Install the required packages for opencv libraries (
libsm6 libxext6 libxrender-dev libglib2.0-0
), for backwards compatibility reasons. Change tofalse
to skip installation and decrease docker spin-up time.
agent.docker_internal_mounts
(dict)
- Set internal mount points inside the Docker. This is especially useful for non-root Docker container images.
For example:
docker_internal_mounts {
sdk_cache: "/clearml_agent_cache"
apt_cache: "/var/cache/apt/archives"
ssh_folder: "/root/.ssh"
pip_cache: "/root/.cache/pip"
poetry_cache: "/root/.cache/pypoetry"
vcs_cache: "/root/.clearml/vcs-cache"
venvs_cache: "/root/.clearml/venvs-cache"
venv_build: "/root/.clearml/venvs-builds"
pip_download: "/root/.clearml/pip-download-cache"
}
agent.docker_pip_cache
(string)
- The pip (Python package tool) cache folder for mapping Python package caching into Docker.
agent.docker_use_activated_venv
(bool)
- In Docker mode, if the container's entrypoint automatically activates a virtual environment, the activated virtual
environment is used and everything is installed in it. Set to
false
to disable, and always create a new venv inheriting fromsystem_site_packages
agent.enable_git_ask_pass
(bool)
:::note
enable_git_ask_pass
is supported only on Linux systems
:::
- If enabled, uses
GIT_ASKPASS
to pass Git user/pass when cloning/fetching repositories - It solves passing user/token to git submodules.
- This is a safer way to ensure multiple users using the same repository will not accidentally leak credentials
agent.enable_task_env
(bool)
- Set the OS environments based on the Task's Environment section before launching the Task process.
agent.extra_docker_arguments
([string])
- Optional arguments to pass to the Docker image when ClearML Agent is running in Docker mode. These are local for this agent, and will not be updated in the experiment's
docker_cmd
section. For example,["--ipc=host", ]
.
agent.extra_docker_shell_script
([string])
- When ClearML Agent is running in Docker mode, this
optional shell script executes inside the Docker on startup, before the experiment starts. For example,
["apt-get install -y bindfs", ]
.
agent.force_git_root_python_path
(bool)
- Force the root folder of the git repository (instead of the working directory) into the
PYHTONPATH
environment variable.false
by default, so only the working directory will be added toPYHTONPATH
agent.force_git_ssh_protocol
(bool)
-
Force Git protocol to use SSH regardless of the Git URL. This assumes the Git user/pass are blank.
The values are:
true
- Forcefalse
- Do not force
agent.force_git_ssh_port
(integer)
- Force a specific SSH port when converting HTTP to SSH links. The domain remains unchanged.
agent.force_git_ssh_user
(string)
- Force a specific SSH username when converting HTTP to SSH links (the default username is 'git')
agent.git_host
(string)
- Limit Git credentials usage to this host. The environment variable
CLEARML_AGENT_GIT_HOST
overrides this configuration option.
agent.git_pass
(string)
-
Git repository password.
- If using Git SSH credentials, do not specify this option.
- If not using Git SSH credentials, use this option to specify a Git password for cloning your repositories.
agent.git_user
(string)
-
Git repository username.
- If using Git SSH credentials, do not specify this option.
- If not using Git SSH credentials, use this option to specify a Git password for cloning your repositories.
agent.hide_docker_command_env_vars
(dict)
-
Hide Docker environment variables containing secrets when printing out the Docker command. When printed, the variable values will be replaced by
********
-
Enable this feature by setting
enabled
totrue
. Doing this will hide the following environment variables values:CLEARML_API_SECRET_KEY
CLEARML_AGENT_GIT_PASS
AWS_SECRET_ACCESS_KEY
AZURE_STORAGE_KEY
-
To mask additional environment variables, add their keys to the
extra_keys
list. For example, to hide the value of a custom environment variable namedMY_SPECIAL_PASSWORD
, setextra_keys: ["MY_SPECIAL_PASSWORD"]
-
By default,
parse_embedded_urls
is set totrue
, so agent will also hide passwords in URLs and handle environment variables containing docker commands
hide_docker_command_env_vars {
enabled: true
extra_keys: ["MY_SPECIAL_PASSWORD"]
parse_embedded_urls: true
}
agent.ignore_requested_python_version
(bool)
-
Indicates whether to ignore any requested python version
-
The values are:
true
- ignore any requested python versionfalse
- if a task was using a specific python version, and the system supports multiple versions, the agent will use the requested python version (default)
agent.protected_docker_extra_args
([string])
- Prevent listed task docker arguments from being used if they are already specified in
agent.extra_docker_arguments
.
agent.python_binary
(string)
- Set the Python version to use when creating the virtual environment, and when launching the experiment. For example,
/usr/bin/python3
or/usr/local/bin/python3.6
.
agent.reload_config
(bool)
- Indicates whether to reload the configuration each time the worker daemon is executed.
agent.translate_ssh
(bool)
- Translate HTTPS communication to SSH
agent.venvs_dir
(string)
- The target folder for virtual environments builds that are created when executing an experiment.
agent.worker_id
(string)
-
When creating a worker, assign the worker an ID.
-
If specified, a unique name for the worker. For example,
clearml-agent-machine1:gpu0
. -
If not specified, the following is used:
<hostname>:<process_id>
.For example,
MyHost:12345
.Alternatively, specify the environment variable
CLEARML_WORKER_ID
to override this worker name.
-
agent.worker_name
(string)
-
Use to replace the hostname when creating a worker if
agent.worker_id
is not specified. For example, ifworker_name
isMyMachine
and theprocess_id
is12345
, then the worker is namedMyMachine.12345
.Alternatively, specify the environment variable
CLEARML_WORKER_NAME
to override this worker name.
agent.default_docker
(dict)
- Dictionary containing the default options for workers running in Docker mode.
These settings define which Docker image and arguments should be used unless explicitly overridden through the UI or an agent.
-
agent.default_docker.image
(str) - Specifies the default Docker image to use. -
agent.default_docker.arguments
([str]) - Specifies the list of options to pass to the Docker container. For example:arguments: ["--ipc=host", ]
. -
agent.default_docker.match_rules
([dict]):::important Enterprise Feature This feature is available under the ClearML Enterprise plan. :::
- Lookup table of rules that determine the default container and arguments when running a worker in Docker mode. The first matched rule will be picked, according to rule order.
- Each dictionary in the list lays out rules, and the container and container arguments to be used if the rules are matched.
:::note Match rule arguments
default_docker.match_rules.arguments
should be formatted as a single string (for example:"-e VALUE=1 --ipc=host"
), unlikeagent.default_docker.arguments
::::::note
match_rules
are ignore if--docker <container>
is passed in the command line. :::- The rules can be:
script.requirements
- Match all script requirementsscript.repostiry
,script.branch
- Match based on Git repository or branch where the script is storedscript.binary
- Match binary executable used to launch the entry pointproject
- Match the Task project's name
- Matching is done via regular expression. For example
"^searchme$"
will match exactly the"searchme"
string, and^examples
will match that starts withexamples
(e.g.,examples
,examples/sub_project
). - Examples:
-
In the example configuration below, the rules match tasks where the Python binary is
python3.6
,tensorflow~=2.6
is required, the script's Git repository is/my_repository/
, the branch ismain
, and the task's project isproject/sub_project
. If all conditions are met, thenvidia/cuda:10.1-cudnn7-runtime-ubuntu18.04
image is used with the argument-e define=value
.agent { default_docker { match_rules [ { image: "nvidia/cuda:10.1-cudnn7-runtime-ubuntu18.04" arguments: "-e define=value" match: { script { # Optional: must match all requirements (not partial) requirements: { # version selection matching PEP-440 pip: { tensorflow: "~=2.6" }, } # Optional: matching based on regular expression, example: "^exact_match$" repository: "/my_repository/" branch: "main" binary: "python3.6" } # Optional: matching based on regular expression, example: "^exact_match$" project: "project/sub_project" } } ] } }
-
In the example configuration below, two
match_rules
are used to specify different Docker images based on the Python binary version. The first rule applies thepython:3.6-bullseye
image with the--ipc=host
argument when the task requirespython3.6
. The second rule applies thepython:3.7-bullseye
image with the same argument when the script requirespython3.7
. If no match is found, the defaultnvidia/cuda:11.0.3-cudnn8-runtime-ubuntu20.04
image is used.``` agent { default_docker: { image: "nvidia/cuda:11.0.3-cudnn8-runtime-ubuntu20.04", match_rules: [ { image: "python:3.6-bullseye" arguments": "--ipc=host" match: { script { binary: "python3.6$" }, } }, { image: "python:3.7-bullseye" arguments: "--ipc=host" match: { script { binary: "python3.7$" }, } }, ] } } ```
-
-
agent.package_manager
(dict)
- Dictionary containing the options for the Python package manager. The currently supported package managers are pip, conda,
and, if the repository contains a
poetry.lock
file, poetry.
agent.package_manager.conda_channels
([string])
- If conda is used, then this is the list of conda channels to use when installing Python packages.
agent.package_manager.conda_full_env_update
(bool)
- Enables update of conda environment (Conda environment does not update by default as it might break)
agent.package_manager.conda_env_as_base_docker
(bool)
- Uses conda environment for execution (like a docker)
agent.package_manager.use_conda_base_env
(bool)
- When set to
True
, installation will be performed into the base Conda environment. Use in Docker mode.
agent.package_manager.extra_index_url
([string])
- A list of URLs for additional artifact repositories when installing Python packages.
agent.package_manager.extra_pip_install_flags
([string])
- A list of additional flags to use when the agent install packages. For example:
["--use-deprecated=legacy-resolver", ]
agent.package_manager.force_upgrade
(bool)
-
Indicates whether to force an upgrade of Python packages.
The values are:
true
- Forcefalse
- Do not force
agent.package_manager.pip_version
(string)
- The
pip
version to use. For example,"<20"
,"==19.3.1"
,""
(empty string will install the latest version).
agent.package_manager.poetry_version
(string)
- The
poetry
version to use. For example,"<2"
,"==1.1.1"
,""
(empty string will install the latest version).
agent.package_manager.poetry_install_extra_args
(list)
- List extra command-line arguments to pass when using
poetry
agent.package_manager.post_optional_packages
(string)
- A list of optional packages that will be installed after the required packages. If the installation of an optional post package fails, the package is ignored, and the virtual environment process continues.
agent.package_manager.post_packages
([string])
- A list of packages that will be installed after the required packages.
agent.package_manager.priority_optional_packages
([string])
- A list of optional priority packages to be installed before the rest of the required packages, but in case a package installation fails, the package will be ignored, and the virtual environment process will continue.
agent.package_manager.priority_packages
([string])
- A list of packages with priority to be installed before the rest of the required packages. For example:
["cython", "numpy", "setuptools", ]
agent.package_manager.pytorch_resolve
(str)
- Set the PyTorch resolving mode. The options are:
pip
(default) - Sets extra index based on cuda and lets pip resolvenone
- No resolving. Install PyTorch like any other packagedirect
- Resolve a direct link to the PyTorch wheel by parsing the pytorch.org pip repository and matching the automatically detected cuda version with the required PyTorch wheel. If the exact cuda version is not found for the required PyTorch wheel, it will try a lower cuda version until a match is found
agent.package_manager.system_site_packages
(bool)
-
Indicates whether Python packages for virtual environments are inherited from the system when building a virtual environment for an experiment.
The values are:
true
- Inheritfalse
- Do not inherit (load Python packages)
agent.package_manager.torch_nightly
(bool)
-
Indicates whether to support installing PyTorch Nightly builds.
The values are:
true
- If a stabletorch
wheel is not found, install the nightly build.false
- Do not install.
:::note Torch Nightly builds are ephemeral and are deleted from time to time. :::
agent.package_manager.type
(string)
-
Indicates the type of Python package manager to use.
The values are:
pip
conda
poetry
-
If
pip
orconda
are used, the agent installs the required packages based on the "installed packages" section of the Task. If the "installed packages" is empty, it will revert to usingrequirements.txt
from the repository's root directory. Ifpoetry
is selected, and the root repository containspoetry.lock
orpyproject.toml
, the "installed packages" section is ignored, andpoetry
is used. Ifpoetry
is selected and no lock file is found, it reverts topip
package manager behaviour.
agent.pip_download_cache
(dict)
- Dictionary containing pip download cache options.
agent.pip_download_cache.enabled
(bool)
-
Indicates whether to use a specific cache folder for Python package downloads.
The values are:
true
- Use a specific folder which is specified in the optionagent.pip_download_cache.path
false
- Do not use a specific folder.
agent.pip_download_cache.path
(string)
- If
agent.pip_download_cache.enabled
istrue
, then this specifies the cache folder.
agent.vcs_cache
(dict)
- Dictionary containing version control system clone cache folder.
agent.vcs_cache.enabled
(bool)
-
Indicates whether the version control system cache is used.
The values are:
true
- Use cachefalse
- Do not use cache
agent.vcs_cache.path
(string)
- The version control system cache clone folder when executing experiments.
agent.venvs_cache
(dict)
- Dictionary containing virtual environment cache options.
agent.venvs_cache.free_space_threshold_gb
(integer)
- Minimum required free space to allow for cache entry.
- Disable minimum by passing 0 or negative value.
agent.venvs_cache.max_entries
(integer)
- Maximum number of cached virtual environments.
agent.venvs_cache.path
(string)
- Folder of the virtual environment cache.
- Uncomment to enable virtual environment caching.
:::note
This option is deprecated. Use venvs_cache
and set venvs_cache.path
instead.
:::
agent.venv_update
(dict)
- Dictionary containing virtual environment update options.
agent.venv_update.enabled
(bool)
-
Indicates whether to use accelerated Python virtual environment building (this is a beta feature).
The values are:
true
- Acceleratefalse
- Do not accelerate (default value)
api
(dict)
Dictionary of configuration options for the ClearML Server API, web, and file servers and credentials.
api.api_server
(string)
- The URL of your ClearML API server. For example,
https://api.MyDomain.com
.
api.web_server
(string)
- The URL of your ClearML web server. For example,
https://app.MyDomain.com
.
api.files_server
(string)
- The URL of your ClearML file server. For example,
https://files.MyDomain.com
.
:::warning
You must use a secure protocol with api.web_server
, api.files_server
, and api.api_server
. Use https
, not http
.
:::
api.http.default_method
(string)
- Set the request method for all API requests and auth login. This can be useful when
GET
requests with payloads are blocked by a server, andPOST
requests can be used instead. The request options are: "GET", "POST", "PUT".
:::warning This configuration option is experimental, and has not been vigorously tested, so it may have unintended consequences. :::
api.credentials
(dict)
- Dictionary of API credentials.
Alternatively, specify the environment variable
CLEARML_API_ACCESS_KEY
/CLEARML_API_SECRET_KEY
to override these keys.
api.credentials.access_key
(string)
- Your ClearML access key.
api.credentials.secret_key
(string)
- Your ClearML credentials.
api.verify_certificate
(bool)
-
Indicates whether to verify the host SSL certificate.
The values are:
true
- Verifyfalse
- Do not verify.
:::warning Set to False only if required. :::
sdk
(dict)
- Dictionary that contains configuration options for the ClearML Python Package and related options, including storage, metrics, network, AWS S3 buckets and credentials, Google Cloud Storage, Azure Storage, log, and development.
sdk.aws
(dict)
- Dictionary with AWS storage options.
sdk.aws.boto3
(dict)
- Dictionary of AWS Storage, Boto3 options. The keys include:
max_multipart_concurrency
(integer) - The maximum number of threads making requests for a transfer.multipart_threshold
(integer) - The transfer size threshold. If size above threshold, Boto3 will automatically use multipart uploads, downloads, and copies (in bytes)multipart_chunksize
(integer) - The size of each part of a multipart transfer (in bytes).pool_connections
(integer) - The maximum number of Boto3 pool connections.
sdk.aws.s3
(dict)
- Dictionary of AWS Storage, AWS S3 options.
sdk.aws.s3.extra_args
(dict)
- Additional ExtraArgs
passed to boto3 when uploading files. This can be set per-bucket under
sdk.aws.s3.credentials
.
sdk.aws.s3.key
(string)
- For AWS S3, the default access key for any bucket that is not specified in the
sdk.aws.s3.credentials
section.
sdk.aws.s3.profile
(string)
- For AWS S3, the default profile name for any bucket that is not specified in the
sdk.aws.s3.credentials
section.
sdk.aws.s3.region
(string)
- For AWS S3, the default region name for any bucket that is not specified in the
sdk.aws.s3.credentials
section.
sdk.aws.s3.secret
(string)
- For AWS S3, the default secret access key for any bucket that is not specified in the
sdk.aws.s3.credentials
section.
sdk.aws.s3.use_credentials_chain
(bool)
- Set to
true
to let Boto3 look for and pick the right credentials, instead of using the explicitly provided default credentials (sdk.aws.s3.secret
andsdk.aws.s3.key
). Boto3 looks for credentials in environment variables, a credential file, and metadata service with an IAM role configured. See Boto3 documentation.
sdk.aws.s3.credentials
([dict])
- List of dictionaries, for AWS S3, each dictionary can contain the credentials for individual S3 buckets or hosts for individual buckets.
sdk.aws.s3.credentials.bucket
(string)
- For AWS S3, if specifying credentials for individual buckets, then this is the bucket name for an individual bucket.
:::note See the AWS documentation for restrictions and limitations on bucket naming. :::
sdk.aws.s3.credentials.host
(string)
- For AWS S3, if specifying credentials for individual buckets by host, then this option is the host URL and optionally the port number.
sdk.aws.s3.credentials.key
(string)
-
For AWS S3:
- If specifying individual bucket, then this is the access key for the bucket.
- If specifying individual buckets by host, then this is access key for all buckets on the host.
sdk.aws.s3.credentials.multipart
(bool)
-
For AWS S3, if specifying credentials for individual buckets by host, then this indicates whether to allow multipart upload of a single object (object as a set of parts).
The values are:
true
- Enabledfalse
- Disabled
sdk.aws.s3.credentials.secret
(bool)
-
For AWS S3:
- If specifying credentials for a specific bucket, then this is the secret key for the bucket.
- If specifying credentials for individual buckets by host, then this is the secret key for all buckets on the host.
sdk.aws.s3.credentials.secure
(string)
-
For AWS S3, if specifying credentials for individual buckets by host, then this indicates whether the host is secure.
The values are:
true
- Securefalse
- Not secure
sdk.aws.s3.credentials.verify
(string/boolean)
- Specify whether to verify SSL certificates. By default, verification is enabled (
true
). You can provide a path or a URL to a CA bundle, or set this option tofalse
to skip SSL certificate verification.
sdk.azure.storage.containers
([dict])
- List of dictionaries, each dictionary contains credentials for an Azure Storage container.
sdk.azure.storage.containers.account_key
(string)
- For Azure Storage, this is the credentials key.
sdk.azure.storage.containers.account_name
(string)
- For Azure Storage, this is the account name.
sdk.azure.storage.containers.container_name
(string)
- For Azure Storage, this is the container name.
sdk.dataset.preview
([dict])
- Set limits for the objects that are logged as dataset previews:
sdk.dataset.preview.media
(dict) - Set limits for media files that are logged as dataset previews. Available options:sdk.dataset.preview.media.max_file_size
(int) - Maximum file size in bytes of a preview object (e.g. image, video, html, etc.). Files exceeding this size will not be reported as previews.sdk.dataset.preview.media.image_count
(int) - The maximum number of image files reported as previewssdk.dataset.preview.media.video_count
(int) - The maximum number of video files reported as previewssdk.dataset.preview.media.audio_count
(int) - The maximum number of image files reported as previewssdk.dataset.preview.media.html_count
(int) - The maximum number of html files reported as previewssdk.dataset.preview.media.json_count
(int) - The maximum number of json files reported as previews
sdk.dataset.preview.tabular
(dict) - Set limits for tabular files that are logged as dataset previews. Available options:sdk.dataset.preview.tabular.row_count
(int) - The maximum number of rows for each tabular file reported as a preview. By default, it will report only the first 10 rows from a filesdk.dataset.preview.tabular.table_count
(int) - The maximum number of tables reported as preview in a dataset
sdk.development
(dict)
- Dictionary of development mode options.
sdk.development.artifacts
(dict)
- Control default behavior when logging task artifacts:
sdk.development.artifacts.default_pandas_dataframe_extension_name
(str)- Set the default
extension_name
for pandasDataFrame
objects - Valid values are:
.csv.gz
,.parquet
,.feather
,.pickle
- This value can be overridden by the
extension_name
argument supplied toTask.upload_artifact()
- Set the default
sdk.development.artifacts.auto_pickle
(bool)- If
true
and the artifact is not of a specific type (pathlib2.Path
,dict
,pandas.DataFrame
,numpy.ndarray
,PIL.Image
, url string,local_file
string), the artifact will be pickled and uploaded as a pickle file artifact (with the.pkl
file extension). - If
false
, the auto-pickle behavior is disabled in the artifact upload - This value can be overridden by the
auto_pickle
argument supplied toTask.upload_artifact()
- If
sdk.development.default_output_uri
(string)
- The default output destination for model checkpoints (snapshots) and artifacts. If the
output_uri
parameter is not provided when callingTask.init()
, then use the destination indefault_output_uri
.
sdk.development.detailed_import_report
(bool)
- If
true
(default isfalse
), provide a detailed report of all python package imports as comments inside the "Installed packages" section.
sdk.development.detect_with_conda_freeze
(bool)
- If
true
(default isfalse
), instead of analyzing the code with Pigar, analyze withconda freeze
sdk.development.detect_with_pip_freeze
(bool)
- If
true
(default isfalse
), instead of analyzing the code with Pigar, analyze withpip freeze
sdk.development.force_analyze_entire_repo
(bool)
- Default auto-generated requirements optimize for smaller requirements.
The values are:
* true
- Analyze the entire repository regardless of the entry point.
* false
- First analyze the entry point script, if it does not contain other local files,
do not analyze the entire repository.
sdk.development.log_os_environments
([string])
-
Log specific environment variables. OS environments are listed in the UI, under an experiment's
CONFIGURATION > HYPERPARAMETERS > Environment section. Multiple selected variables are supported including the suffix*
. For example:"AWS_*"
will log any OS environment variable starting with"AWS_"
. Example:log_os_environments: ["AWS_*", "CUDA_VERSION"]
-
This value can be overwritten with OS environment variable
CLEARML_LOG_ENVIRONMENT=AWS_*,CUDA_VERSION
.
sdk.development.store_uncommitted_code_diff
(bool)
-
For development mode, indicates whether to store the uncommitted
git diff
orhg diff
in the experiment manifest.The values are:
true
- Store thediff
in thescript.requirements.diff
sectionfalse
- Do not store the diff.
sdk.development.suppress_update_message
(bool)
-
If
true
(defaultfalse
), clearml update messages will not be printed to the console. -
This value can be overwritten with OS environment variable
CLEARML_SUPPRESS_UPDATE_MESSAGE=1
sdk.development.support_stopping
(bool)
-
For development mode, indicates whether to allow stopping an experiment if the experiment was aborted externally, its status was changed, or it was reset.
The values are:
true
- Allowfalse
- Do not allow
sdk.development.task_reuse_time_window_in_hours
(float)
- For development mode, the number of hours after which an experiment with the same project name and experiment name is reused.
sdk.development.vcs_repo_detect_async
(bool)
-
For development mode, indicates whether to run version control repository detection asynchronously.
The values are:
true
- Run asynchronouslyfalse
- Do not run asynchronously
sdk.development.worker
(dict)
- Dictionary of development mode options for workers.
sdk.development.worker.log_stdout
(bool)
-
For development mode workers, indicates whether all stdout and stderr messages are logged.
The values are:
true
- Log allfalse
- Do not log all
sdk.development.worker.max_wait_for_first_iteration_to_start_sec
(integer)
- Maximum time (in seconds) for allowing the resource monitoring to switch back to reporting iterations as the x-axis after initially starting to report "seconds from start." If the specified time limit is exceeded, the resource monitoring will continue reporting using "seconds from start" as the x-axis.
sdk.development.worker.ping_period_sec
(integer)
- For development mode workers, the interval in seconds for a worker to ping the server testing connectivity.
sdk.development.worker.report_event_flush_threshold
(integer)
- The number of events that trigger a report
sdk.development.worker.report_global_mem_used
(bool)
-
Compatibility feature to report memory usage for the entire machine
The values are:
true
- Report memory usage for the entire machinefalse
(default) - Report memory usage only on the running process and its sub-processes
sdk.development.worker.report_period_sec
(integer)
- For development mode workers, the interval in seconds for a development mode ClearML worker to report.
sdk.development.worker.report_start_sec
(integer)
- The number of seconds after which the development mode worker starts resource reporting.
sdk.development.worker.wait_for_first_iteration_to_start_sec
(integer)
- Controls how long (in seconds) to wait for iteration reporting to be used as x-axis for resource monitoring. If iteration reporting is unavailable once time is exceeded, "seconds from start" will be used for the x-axis.
sdk.google.storage
(dict)
- Dictionary of Google Cloud Storage credentials.
sdk.google.storage.project
(string)
- For Google Cloud Storage, the name of project.
sdk.google.storage.credentials_json
(string)
- For Google Cloud Storage, the file path for the default Google storage credentials JSON file.
sdk.google.storage.credentials
([dict])
- A list of dictionaries, with specific credentials per bucket and subdirectory
sdk.google.storage.credentials.bucket
(string)
- For Google Cloud Storage, if specifying credentials by the individual bucket, the name of the bucket.
sdk.google.storage.credentials.credentials_json
(string)
- For Google Cloud Storage, if specifying credentials by the individual bucket, the file path for the default Google storage credentials JSON file.
sdk.google.storage.credentials.project
(string)
- For Google Cloud Storage, if specifying credentials by the individual bucket, the name of the project.
sdk.google.storage.credentials.subdir
(string)
- For Google Cloud Storage, if specifying credentials by the individual bucket, a subdirectory within the bucket.
sdk.log
(dict)
- Dictionary of log options.
sdk.log.disable_urllib3_info
(bool)
-
Indicates whether to disable
urllib3
info messages.The values are:
true
- Disablefalse
- Do not disable
sdk.log.null_log_propagate
(bool)
-
As debugging feature, indicates whether to allow null log messages to propagate to the root logger (so they appear as stdout).
The values are:
true
- Allowfalse
- Do not allow
sdk.log.task_log_buffer_capacity
(integer)
- The maximum capacity of the log buffer.
sdk.metrics
(dict)
- Dictionary of metrics options.
sdk.metrics.file_history_size
(string)
- The history size for debug files per metric / variant combination
- Each metric / variant combination,
file_history_size
indicates the number of files stored in the upload destination - Files are recycled so that
file_history_size
is the maximum number of files at any time.
sdk.metrics.matplotlib_untitled_history_size
(integer)
- The maximum history size for
matplotlib
imshow
files per plot title. - File names for the uploaded images are recycled so that the number of images stored in the upload destination for each matplotlib plot title
will not exceed the value of
matplotlib_untitled_history_size
sdk.metrics.plot_max_num_digits
(integer)
- The maximum number of digits after the decimal point in plot reporting. This can reduce the report size.
sdk.metrics.tensorboard_single_series_per_graph
(bool)
:::note This option is deprecated. This plot behavior is now controlled via the UI :::
-
Indicates whether plots appear using TensorBoard behavior where each series is plotted in its own graph (plot-per-graph).
The values are:
true
- Support TensorBoard behaviorfalse
- Do not
sdk.metrics.images
(dict)
- Dictionary of metrics images options.
sdk.metrics.images.format
(string)
- The image file format for generated debug images (such as "JPEG").
sdk.metrics.images.quality
(integer)
- The image quality for generated debug images.
sdk.metrics.images.subsampling
(integer)
- The image subsampling for generated debug images.
sdk.network.file_upload_retries
(int)
- Number of retries before failing to upload a file
sdk.network.iteration
(dict)
- Dictionary of network iteration options.
sdk.network.iteration.max_retries_on_server_error
` (integer)
- For retries when getting frames from the server, if the server returned an error (http code 500), then this is the maximum number of retries.
sdk.network.iteration.retry_backoff_factor_sec
(integer)
- For retries when getting frames from the server, this is backoff factor for consecutive retry attempts. This is used to
determine the number of seconds between retries. The retry backoff factor is calculated as
{backoff factor} * (2 ^ ({number of total retries} - 1))
.
sdk.network.metrics
(dict)
- Dictionary of network metrics options.
sdk.network.metrics.file_upload_starvation_warning_sec
(integer)
- The number of seconds before a warning is issued that file-bearing events are sent for upload, but no uploads occur.
sdk.network.metrics.file_upload_threads
(integer)
- The number of threads allocated to uploading files when transmitting metrics for a specific iteration.
sdk.storage
(dict)
- Dictionary of storage options.
sdk.storage.cache
(dict)
- Dictionary of storage cache options. The keys include:
default_base_dir
(str) - The default base directory for caching. The default is the<system_temp_folder>/clearml_cache
.default_cache_manager_size
(int) - Maximum number of files in the cache (default 100 files).
:::important Enterprise features
The ClearML Enterprise plan also supports the following configuration options under sdk.storage.cache
:
size.max_used_bytes
(str) - Maximum size of the local cache directory. If set to-1
, the directory can use the available disk space. Specified in storage units (for example:1GB
,2TB
,500MB
).size.min_free_bytes
(str) - Minimum amount of free disk space that should be left. Ifsize.max_used_bytes
is set to-1
, this configuration option will limit the cache directory maximum size tofree disk space - size.min_free_bytes
. Specified in storage units (for example:1GB
,2TB
,500MB
).zero_file_size_check
(bool)- If set toTrue
, each cache hit will also check the cached file size, making sure it is not zero (defaultFalse
)secondary
(dict) - Set up a secondary cache (acts as an L2 cache). When a request is made, the primary cache is queried first. If the data is not in the primary cache, the secondary cache is queried. In case of a cache miss, the data will be pulled to the primary cache, and then copied to the secondary cache. Thesdk.storage.cache.secondary
dictionary supports the same option as the primary cache:default_base_dir
(required),size.max_used_bytes
,size.min_free_bytes
, etc. If an option is unspecified, it defaults to the primary cache's value. :::
sdk.storage.direct_access
(dict)
- Dictionary of storage direct access options.
sdk.storage.direct_access.url
(string)
- Specify a list of direct access objects using glob patterns which matches sets of files using wildcards. Direct access objects are not downloaded or cached, and any download request will return a direct reference.
sdk.storage.log.report_download_chunk_size_mb
(int)
- Specify how often in MB the
StorageManager
reports its download progress to the console. By default, it reports every 5MB
sdk.storage.log.report_upload_chunk_size_mb
(int)
- Specify how often in MB the
StorageManager
reports its upload progress to the console. By default, it reports every 5MB
sdk.storage.path_substitution
([dict])
-
List of dictionaries, where each dictionary contains path substitution mapping. This is useful in cases where data was originally logged in one location and later moved, or when different workloads access the data through different mounts.
-
Each dictionary contains a
registered_prefix
and alocal_prefix
.registered_prefix
is the URL prefix logged in ClearML.local_prefix
is the URL prefix to be used at runtime instead of theregistered_prefix
to access the data.For example:
sdk { storage { path_substitution = [ { registered_prefix = "s3://bucket/research" local_prefix = "file:///mnt/shared/bucket/research" }, { registered_prefix = "file:///mnt/shared/folder/" local_prefix = "file:///home/user/shared/folder" } ] } }
environment
(dict)
Dictionary of environment variables and values which are applied to the OS environment as key=value
for each key-value
pair.
Enable by setting agent.apply_environment
OR sdk.apply_environment
to true
.
Example:
environment {
key_a: value_a
key_b: value_b
}
files
(dict)
The files
section allows to define files which will be auto-generated at designated paths with predefined content and
target format.
Enable by setting agent.apply_files
OR sdk.apply_files
to true
.
Define each file's contents in a dictionary. Files content options include:
contents
- Target file's content, typically a string (or any base type int/float/list/dict etc.)format
- Custom format for the contents. Currently supportsbase64
to automatically decode a base64-encoded contents string, otherwise ignoredpath
- Target file's path, may include~
and inplace env varstarget_format
- Format used to encode contents before writing into the target file. Supported values arejson
,yaml
,yml
, andbytes
(in which case the file will be written in binary mode). Default is text mode.mode
- File-system mode (permissions) to apply to the file after its creation. The mode string will be parsed into an integer (for example:"0o777"
for-rwxrwxrwx
)overwrite
- Overwrite the target file in case it exists. Default istrue
.
Example:
files {
myfile1 {
contents: "The quick brown fox jumped over the lazy dog"
path: "/tmp/fox.txt"
mode: "0o777"
}
myjsonfile {
contents: {
some {
nested {
value: [1, 2, 3, 4]
}
}
}
path: "/tmp/test.json"
target_format: json
}
}
# Apply top-level files section from configuration into local file system
sdk {
apply_files: true
}
:::important Enterprise Feature This feature is available under the ClearML Enterprise plan. :::
The ClearML Enterprise Server includes the configuration vault. Users can add configuration sections to the vault and, once the vault is enabled, the configurations will be merged into the ClearML and ClearML Agent configurations upon code execution and/or agent launch.
These configurations override the configurations written in the configuration file.
See configuration vault.