diff --git a/docs/source/quicktour.mdx b/docs/source/quicktour.mdx index b6ec6c7d8..798883f12 100644 --- a/docs/source/quicktour.mdx +++ b/docs/source/quicktour.mdx @@ -129,7 +129,7 @@ accelerate). - **add_special_tokens** (bool, optional, defaults to True): Whether to add special tokens to the input sequences. If `None`, the default value will be set to `True` for seq2seq models (e.g. T5) and `False` for causal models. -- **model_parallel** (bool, optional, defaults to False): +- **model_parallel** (bool, optional, defaults to None): True/False: force to use or not the `accelerate` library to load a large model across multiple devices. Default: None which corresponds to comparing the number of processes with diff --git a/src/lighteval/models/transformers/transformers_model.py b/src/lighteval/models/transformers/transformers_model.py index 48edb4310..b34cfaa3f 100644 --- a/src/lighteval/models/transformers/transformers_model.py +++ b/src/lighteval/models/transformers/transformers_model.py @@ -112,7 +112,7 @@ class TransformersModelConfig: add_special_tokens (bool, optional, defaults to True): Whether to add special tokens to the input sequences. If `None`, the default value will be set to `True` for seq2seq models (e.g. T5) and `False` for causal models. - model_parallel (bool, optional, defaults to False): + model_parallel (bool, optional, defaults to None): True/False: force to use or not the `accelerate` library to load a large model across multiple devices. Default: None which corresponds to comparing the number of processes with