diff --git a/allennlp_models/generation/dataset_readers/cnn_dm.py b/allennlp_models/generation/dataset_readers/cnn_dm.py index 54704e87e..2fcb5495a 100644 --- a/allennlp_models/generation/dataset_readers/cnn_dm.py +++ b/allennlp_models/generation/dataset_readers/cnn_dm.py @@ -59,7 +59,7 @@ def __init__( **kwargs, ) -> None: super().__init__( - manual_distributed_sharding=True, manual_multi_process_sharding=True, **kwargs + manual_distributed_sharding=True, manual_multiprocess_sharding=True, **kwargs ) self._source_tokenizer = source_tokenizer or SpacyTokenizer() self._target_tokenizer = target_tokenizer or self._source_tokenizer diff --git a/allennlp_models/generation/dataset_readers/copynet_seq2seq.py b/allennlp_models/generation/dataset_readers/copynet_seq2seq.py index f6117297b..259aace23 100644 --- a/allennlp_models/generation/dataset_readers/copynet_seq2seq.py +++ b/allennlp_models/generation/dataset_readers/copynet_seq2seq.py @@ -96,7 +96,7 @@ def __init__( **kwargs, ) -> None: super().__init__( - manual_distributed_sharding=True, manual_multi_process_sharding=True, **kwargs + manual_distributed_sharding=True, manual_multiprocess_sharding=True, **kwargs ) self._target_namespace = target_namespace self._source_tokenizer = source_tokenizer or SpacyTokenizer() diff --git a/allennlp_models/generation/dataset_readers/seq2seq.py b/allennlp_models/generation/dataset_readers/seq2seq.py index 13a828e44..8e00188b8 100644 --- a/allennlp_models/generation/dataset_readers/seq2seq.py +++ b/allennlp_models/generation/dataset_readers/seq2seq.py @@ -84,7 +84,7 @@ def __init__( **kwargs, ) -> None: super().__init__( - manual_distributed_sharding=True, manual_multi_process_sharding=True, **kwargs + manual_distributed_sharding=True, manual_multiprocess_sharding=True, **kwargs ) self._source_tokenizer = source_tokenizer or SpacyTokenizer() self._target_tokenizer = target_tokenizer or self._source_tokenizer diff --git a/allennlp_models/pair_classification/dataset_readers/snli.py b/allennlp_models/pair_classification/dataset_readers/snli.py index f864df381..af100acc3 100644 --- a/allennlp_models/pair_classification/dataset_readers/snli.py +++ b/allennlp_models/pair_classification/dataset_readers/snli.py @@ -46,7 +46,7 @@ def __init__( **kwargs, ) -> None: super().__init__( - manual_distributed_sharding=True, manual_multi_process_sharding=True, **kwargs + manual_distributed_sharding=True, manual_multiprocess_sharding=True, **kwargs ) self._tokenizer = tokenizer or SpacyTokenizer() if isinstance(self._tokenizer, PretrainedTransformerTokenizer): diff --git a/allennlp_models/rc/dataset_readers/squad.py b/allennlp_models/rc/dataset_readers/squad.py index 809e56091..2981cb295 100644 --- a/allennlp_models/rc/dataset_readers/squad.py +++ b/allennlp_models/rc/dataset_readers/squad.py @@ -91,7 +91,7 @@ def __init__( skip_impossible_questions = kwargs.pop("skip_invalid_examples") super().__init__( - manual_distributed_sharding=True, manual_multi_process_sharding=True, **kwargs + manual_distributed_sharding=True, manual_multiprocess_sharding=True, **kwargs ) self._tokenizer = tokenizer or SpacyTokenizer() self._token_indexers = token_indexers or {"tokens": SingleIdTokenIndexer()} diff --git a/allennlp_models/rc/dataset_readers/transformer_squad.py b/allennlp_models/rc/dataset_readers/transformer_squad.py index bd0ad7cab..cef8c921f 100644 --- a/allennlp_models/rc/dataset_readers/transformer_squad.py +++ b/allennlp_models/rc/dataset_readers/transformer_squad.py @@ -103,7 +103,7 @@ def __init__( skip_impossible_questions = kwargs.pop("skip_invalid_examples") super().__init__( - manual_distributed_sharding=True, manual_multi_process_sharding=True, **kwargs + manual_distributed_sharding=True, manual_multiprocess_sharding=True, **kwargs ) self._tokenizer = PretrainedTransformerTokenizer( transformer_model_name, diff --git a/allennlp_models/structured_prediction/tools/write_srl_predictions_to_conll_format.py b/allennlp_models/structured_prediction/tools/write_srl_predictions_to_conll_format.py index 7828aecdd..660113efb 100644 --- a/allennlp_models/structured_prediction/tools/write_srl_predictions_to_conll_format.py +++ b/allennlp_models/structured_prediction/tools/write_srl_predictions_to_conll_format.py @@ -7,7 +7,6 @@ from typing import List import torch -from allennlp.data.samplers import PyTorchSequentialSampler from allennlp_models.structured_prediction.models.srl import write_to_conll_eval_file @@ -16,7 +15,8 @@ from allennlp.common.tqdm import Tqdm from allennlp.common import Params from allennlp.models.archival import load_archive -from allennlp.data import DatasetReader, DataLoader +from allennlp.data import DatasetReader +from allennlp.data.data_loaders import SimpleDataLoader from allennlp.nn.util import move_to_device @@ -63,10 +63,10 @@ def main(serialization_directory: str, device: int, data: str, prefix: str, doma # Load the evaluation data and index it. print("reading evaluation data from {}".format(evaluation_data_path)) - dataset = dataset_reader.read(evaluation_data_path) + dataset = list(dataset_reader.read(evaluation_data_path)) with torch.autograd.no_grad(): - loader = DataLoader(dataset, sampler=PyTorchSequentialSampler(dataset), batch_size=32) + loader = SimpleDataLoader(dataset, 32) model_predictions: List[List[str]] = [] for batch in Tqdm.tqdm(loader): batch = move_to_device(batch, device)