From 7a7df9470456d99cbae2c3765c04b2bcb6a43aad Mon Sep 17 00:00:00 2001 From: George Ohashi Date: Tue, 8 Aug 2023 15:57:31 +0000 Subject: [PATCH] select largest tokenizer --- src/deepsparse/transformers/pipelines/embedding_extraction.py | 2 +- src/deepsparse/transformers/pipelines/question_answering.py | 2 +- src/deepsparse/transformers/pipelines/text_classification.py | 2 +- src/deepsparse/transformers/pipelines/token_classification.py | 2 +- .../transformers/pipelines/zero_shot_text_classification.py | 2 +- 5 files changed, 5 insertions(+), 5 deletions(-) diff --git a/src/deepsparse/transformers/pipelines/embedding_extraction.py b/src/deepsparse/transformers/pipelines/embedding_extraction.py index 10c216296d..b07ba188ea 100644 --- a/src/deepsparse/transformers/pipelines/embedding_extraction.py +++ b/src/deepsparse/transformers/pipelines/embedding_extraction.py @@ -330,7 +330,7 @@ def route_input_to_bucket( :param pipelines: Different buckets to be used :return: The correct Pipeline object (or Bucket) to route input to """ - tokenizer = pipelines[0].tokenizer + tokenizer = pipelines[-1].tokenizer tokens = tokenizer( input_schema.inputs, add_special_tokens=True, diff --git a/src/deepsparse/transformers/pipelines/question_answering.py b/src/deepsparse/transformers/pipelines/question_answering.py index 5363d276e5..c82b2457b5 100644 --- a/src/deepsparse/transformers/pipelines/question_answering.py +++ b/src/deepsparse/transformers/pipelines/question_answering.py @@ -493,7 +493,7 @@ def route_input_to_bucket( :param pipelines: Different buckets to be used :return: The correct Pipeline object (or Bucket) to route input to """ - tokenizer = pipelines[0].tokenizer + tokenizer = pipelines[-1].tokenizer tokens = tokenizer( " ".join((input_schema.context, input_schema.question)), add_special_tokens=True, diff --git a/src/deepsparse/transformers/pipelines/text_classification.py b/src/deepsparse/transformers/pipelines/text_classification.py index d6dfb58fe7..18625586d8 100644 --- a/src/deepsparse/transformers/pipelines/text_classification.py +++ b/src/deepsparse/transformers/pipelines/text_classification.py @@ -281,7 +281,7 @@ def route_input_to_bucket( :param pipelines: Different buckets to be used :return: The correct Pipeline object (or Bucket) to route input to """ - tokenizer = pipelines[0].tokenizer + tokenizer = pipelines[-1].tokenizer tokens = tokenizer( input_schema.sequences, add_special_tokens=True, diff --git a/src/deepsparse/transformers/pipelines/token_classification.py b/src/deepsparse/transformers/pipelines/token_classification.py index cc14154127..e94fa51dfc 100644 --- a/src/deepsparse/transformers/pipelines/token_classification.py +++ b/src/deepsparse/transformers/pipelines/token_classification.py @@ -368,7 +368,7 @@ def route_input_to_bucket( :param pipelines: Different buckets to be used :return: The correct Pipeline object (or Bucket) to route input to """ - tokenizer = pipelines[0].tokenizer + tokenizer = pipelines[-1].tokenizer tokens = tokenizer( input_schema.inputs, add_special_tokens=True, diff --git a/src/deepsparse/transformers/pipelines/zero_shot_text_classification.py b/src/deepsparse/transformers/pipelines/zero_shot_text_classification.py index c6b0e2c5e5..be24d0cd7d 100644 --- a/src/deepsparse/transformers/pipelines/zero_shot_text_classification.py +++ b/src/deepsparse/transformers/pipelines/zero_shot_text_classification.py @@ -299,7 +299,7 @@ def route_input_to_bucket( :param pipelines: Different buckets to be used :return: The correct Pipeline object (or Bucket) to route input to """ - tokenizer = pipelines[0].tokenizer + tokenizer = pipelines[-1].tokenizer tokens = tokenizer( input_schema.sequences, add_special_tokens=True,