cerebras.modelzoo.data.nlp.bert.BertCSVDataProcessor.BertCSVDataProcessorConfig#
- class cerebras.modelzoo.data.nlp.bert.BertCSVDataProcessor.BertCSVDataProcessorConfig(*args, **kwargs)[source]#
Bases:
cerebras.modelzoo.config.data_config.DataConfig
Methods
check_for_deprecated_fields
check_literal_discriminator_field
copy
get_disable_nsp
get_orig_class
get_orig_class_args
model_copy
model_post_init
post_init
Attributes
The batch size.
A list of bucket boundaries.
Path to the data files to use.
Whether Next Sentence Prediction (NSP) objective is disabled.
discriminator
discriminator_value
do_lower
Whether to drop last batch of epoch if it's an incomplete batch.
Whether to dynamically scale the loss.
masked_lm_prob
max_position_embeddings
max_predictions_per_seq
max_sequence_length
mixed_precision
model_config
The number of PyTorch processes used in the dataloader.
Whether or not to keep workers persistent between epochs.
The number of batches to prefetch in the dataloader.
Whether or not to shuffle the dataset.
Buffer size to shuffle samples across.
The seed used for deterministic shuffling.
vocab_file
vocab_size
whole_word_masking
data_processor
- data_dir = Ellipsis#
Path to the data files to use.
- batch_size = Ellipsis#
The batch size.
- disable_nsp = False#
Whether Next Sentence Prediction (NSP) objective is disabled.
- dynamic_mlm_scale = False#
Whether to dynamically scale the loss.
- buckets = None#
A list of bucket boundaries. If set to None, then no bucketing will happen, and data will be batched normally. If set to a list, then data will be grouped into len(buckets) + 1 buckets. A sample s will go into bucket i if buckets[i-1] <= element_length_fn(s) < buckets[i] where 0 and inf are the implied lowest and highest boundaries respectively. buckets must be sorted and all elements must be non-zero.
- shuffle = False#
Whether or not to shuffle the dataset.
- shuffle_seed = None#
The seed used for deterministic shuffling.
- shuffle_buffer = None#
Buffer size to shuffle samples across. If None and shuffle is enabled, 10*batch_size is used.
- num_workers = 0#
The number of PyTorch processes used in the dataloader.
- prefetch_factor = 2#
The number of batches to prefetch in the dataloader.
- persistent_workers = False#
Whether or not to keep workers persistent between epochs.
- drop_last = True#
Whether to drop last batch of epoch if it’s an incomplete batch.