From 75bcc1e38cd716e717c2788d109d40675c2bf278 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 6 Oct 2022 22:07:25 -0700 Subject: [PATCH] Multiprocessing fix (#5106) (#5107) Signed-off-by: Matvei Novikov Signed-off-by: Matvei Novikov Signed-off-by: Matvei Novikov Co-authored-by: Matvei Novikov Signed-off-by: Hainan Xu --- .../punctuation_capitalization_lexical_audio_config.yaml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/examples/nlp/token_classification/conf/punctuation_capitalization_lexical_audio_config.yaml b/examples/nlp/token_classification/conf/punctuation_capitalization_lexical_audio_config.yaml index 7641d2be12cf..254c4343eeb6 100644 --- a/examples/nlp/token_classification/conf/punctuation_capitalization_lexical_audio_config.yaml +++ b/examples/nlp/token_classification/conf/punctuation_capitalization_lexical_audio_config.yaml @@ -108,7 +108,7 @@ model: # Number of jobs for tokenization and labels encoding. If 0, then multiprocessing is not used. If null, # number of jobs is equal to the number of CPU cores. # WARNING: can cause deadlocks with tokenizers, which use multiprocessing (e.g. SentencePiece) - n_jobs: null + n_jobs: 0 # Path to tarred dataset metadata file. Required if tarred dataset is used. Metadata file is a JSON file which # contains total number of batches in the dataset, a list of paths to tar files and paths to label vocabularies. @@ -143,7 +143,7 @@ model: # Number of jobs for tokenization and labels encoding. If 0, then multiprocessing is not used. If null, # number of jobs is equal to the number of CPU cores. # WARNING: can cause deadlocks with tokenizers, which use multiprocessing (e.g. SentencePiece) - n_jobs: null + n_jobs: 0 # For more details see `train_ds` section. tar_metadata_file: null @@ -174,7 +174,7 @@ model: # Number of jobs for tokenization and labels encoding. If 0, then multiprocessing is not used. If null, # number of jobs is equal to the number of CPU cores. # WARNING: can cause deadlocks with tokenizers, which use multiprocessing (e.g. SentencePiece) - n_jobs: null + n_jobs: 0 # For more details see `train_ds` section. tar_metadata_file: null